CN102508547A - Computer-vision-based gesture input method construction method and system - Google Patents

Computer-vision-based gesture input method construction method and system Download PDF

Info

Publication number
CN102508547A
CN102508547A CN2011103459148A CN201110345914A CN102508547A CN 102508547 A CN102508547 A CN 102508547A CN 2011103459148 A CN2011103459148 A CN 2011103459148A CN 201110345914 A CN201110345914 A CN 201110345914A CN 102508547 A CN102508547 A CN 102508547A
Authority
CN
China
Prior art keywords
gesture
frame
data
buffer model
input method
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN2011103459148A
Other languages
Chinese (zh)
Inventor
王轩
王金磊
于成龙
赵海楠
张加佳
许欣欣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Graduate School Harbin Institute of Technology
Original Assignee
Shenzhen Graduate School Harbin Institute of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Graduate School Harbin Institute of Technology filed Critical Shenzhen Graduate School Harbin Institute of Technology
Priority to CN2011103459148A priority Critical patent/CN102508547A/en
Publication of CN102508547A publication Critical patent/CN102508547A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention relates to a computer-vision-based gesture input method construction method, which comprises the following steps of: gesture acquisition, video processing, gesture analysis, gesture identification and construction of a gesture input method. Gesture frames are clustered by utilizing an asynchronous buffering model to obtain key gesture frames, and only the key gesture frames are subjected to the subsequent gesture analysis and gesture identification; therefore, compared with the conventional system for analyzing and identifying each frame of gestures, a system provided by the invention can give a more real-time response. In addition, the combined application of the gesture identification and the input method is creative in the history of human-computer interaction.

Description

Gesture input method construction method and system based on computer vision
Technical field
The present invention relates to a kind of gesture input method construction method and system, more particularly to a kind of gesture input method construction method and system based on Video clustering processing.
Background technology
With the development of technology, in some special dimensions, gesture input technology has also obtained significant progress, and the premise of gesture input technology is the collection of hand signal, and then the hand signal of collection is analyzed and processed.Because the detection of gesture target refers to detect gesture target from image stream in the case where people is with complicated background condition, gesture identification is then carried out again.Gesture identification is that its high-level implication is explained according to human hand posture and change procedure, because gesture has following four feature, so it is characterized in one of chief technology that gesture input method is realized to extract with geometric invariance:Hand is elastomeric objects, so difference is very big between same gesture;Hand has bulk redundancy information, because gesture identification key is identification finger characteristic, so palm is characterized in redundancy;The position of hand is that, in three dimensions, therefore, it is difficult to position, and the image that computer is obtained is the three-dimensional projection to two dimension, so projecting direction is crucial;The surface Non-smooth surface of hand, so easily producing shade.
Existing Gesture Recognition is that all data frames for capturing camera all carry out gesture analysis and gesture identification, and all data frames include gesture preparatory stage, critical stage and Restoration stage, therefore, gesture recognition system response poor real.Existing input method is that the keyboard of the keyboard, mouse or mobile device by user's control PC, screen are inputted.
The content of the invention
Present invention solves the technical problem that being:A kind of gesture input method construction method and system based on computer video are built, overcomes prior art gesture recognition system response speed slow, relies only on keyboard and mouse touch-screen to use the technical problem of input method.
The technical scheme is that:A kind of gesture input method construction method based on computer vision is provided, step is as follows
Gather gesture:Gesture vision signal is caught, gesture data picture frame is obtained;
Video processing:Using the asynchronous clustering processing model based on interval gesture data frame buffering, the asynchronous clustering processing model includes buffer model, the gesture data picture frame of collection is sequentially stored in buffer model, clustering processing is carried out to the gesture data picture frame in buffer model, gesture keyframe sequence is obtained;
Gesture analysis:After gesture key frame images binaryzation and smothing filtering denoising, carry out gesture contours extract, crucial gesture outline data is obtained, according to gesture key frame images binary map and the gesture contours extract gesture feature parameter of proposition, the gesture feature parameter composition gesture feature vector;
Gesture identification:Gesture identification is carried out according to gesture feature vector, corresponding with identification gesture key letter virtual key or Keyboard Control virtual key is obtained;
Build gesture input method:According to obtained key letter virtual key and the existing input method of Keyboard Control virtual key call operation system.
The present invention further technical scheme be:In video-processing steps, the total data of preparatory stage of each class including gesture, key frame stage and Restoration stage that clustering processing is obtained is carried out to the gesture data picture frame in buffer model, the cluster centre of each class is the gesture key frame of each gesture.
The present invention further technical scheme be:In video-processing steps, threshold parameter is set, the frame by comparison interval data frame is poor, if frame difference removes the gesture data frame not less than threshold parameter in buffer model;If frame difference exceedes threshold parameter, it is defined as drawing class border, and choose the gesture data frame of such Median Position as such gesture key frame.
The present invention further technical scheme be:In video-processing steps, the gesture data picture frame of collection is sequentially stored in buffer model, when buffer model completely starts to carry out clustering processing to gesture data frame.
The present invention further technical scheme be:In video-processing steps, while gesture data frame clustering processing, constantly enter buffer model from the new gesture data frame of seizure, when the gesture data frame clustering processing speed in buffer model is less than gesture data frame capture velocity, when buffer model data are full, wait gesture data frame to be clustered processing, and point out user the size for increasing configuration file;When buffer model gesture data frame clustering processing speed is more than gesture data frame capture velocity, is there is the buffer model data empty time, waiting gesture data frame to be captured that gesture data frame is put into buffer model, and point out user to reduce the size in configuration file.
The present invention further technical scheme be:The gesture feature parameter includes gesture area feature, Hu invariant moment features and Fourier description.
The present invention further technical scheme be:In gesture analysis step, gesture contours extract is carried out using Laplce's Boundary extracting algorithm.
The technical scheme is that:Build a kind of gesture input method constructing system based on computer vision, including gather the gesture collecting unit of gesture vision signal, the video processing unit handled the gesture vision signal of collection, the gesture analysis unit for carrying out gesture analysis and obtaining gesture feature vector, the gesture identification unit for carrying out according to gesture feature vector gesture identification, structure gesture input method construction unit, the gesture collecting unit catches gesture vision signal, obtains gesture data picture frame;The video processing unit uses the asynchronous clustering processing model based on interval gesture data frame buffering, the asynchronous clustering processing model includes buffer model, the gesture data picture frame of collection is sequentially stored in buffer model, clustering processing is carried out to the gesture data picture frame in buffer model, gesture keyframe sequence is obtained;After the gesture analysis unit is to gesture key frame images binaryzation and smothing filtering denoising, carry out gesture contours extract, obtain crucial gesture outline data, according to gesture key frame images binary map and the gesture contours extract gesture feature parameter of proposition, the gesture feature parameter composition gesture feature vector;The gesture identification unit carries out gesture identification according to gesture feature vector, obtains corresponding with identification gesture key letter virtual key or Keyboard Control virtual key;The construction unit is according to obtained key letter virtual key and the existing input method of Keyboard Control virtual key call operation system.
The present invention further technical scheme be:The video processing unit includes carrying out the gesture data picture frame in buffer model the clustering processing module of clustering processing, the clustering processing module reads gesture frame data using the clustering algorithm based on time series from buffer model, and gesture key frame is obtained by being spaced gesture frame frame difference.
The present invention further technical scheme be:The gesture identification unit carries out gesture identification according to gesture feature vector, obtains and the identification corresponding 26 English letter of keyboard virtual keys of gesture and 4 Keyboard Control virtual keys.
The solution have the advantages that:Build a kind of gesture input method construction method based on Video clustering processing and system.Gesture frame is clustered using asynchronous buffer model, crucial gesture frame is obtained, only by the follow-up gesture analysis of crucial gesture frame progress and gesture identification, relative to conventional analysis and the system of each frame gesture of identification, system response real-time is effectively increased.In addition, the present invention is by the connected applications of gesture identification and input method, it is the innovation in man-machine interaction history.
Brief description of the drawings
Fig. 1 is flow chart of the invention.
Fig. 2 is structural representation of the invention.
Fig. 3 is video processing architecture schematic diagram of the present invention.
Fig. 4 is gesture analysis structural representation of the present invention.
Fig. 5 is gesture identification of the present invention and builds input method structural representation.
Embodiment
With reference to specific embodiment, technical solution of the present invention is further illustrated.
As shown in figure 1, the embodiment of the present invention is:The technical scheme is that:A kind of gesture input method construction method based on computer vision is provided, step is as follows
Step 100:Gesture is gathered, i.e.,:Gesture vision signal is caught, gesture data picture frame is obtained.
Specific implementation process is as follows:One gesture word is a limited basic gesture sequence, and each basic gesture includes the following three stage:Preparatory stage, i.e., hand is moved on to the motion stage of crucial hand gesture location from some initial rest position;Crucial gesture stage, i.e. gesture motion stage;Restoration stage, i.e. hand return to the motion stage of initial position.In the specific embodiment of the invention, gesture vision signal is caught using camera, gesture data picture frame is obtained by processing.
Step 200:Video processing, i.e.,:Using the asynchronous clustering processing model based on interval gesture data frame buffering, the asynchronous clustering processing model includes buffer model, the gesture data picture frame of collection is sequentially stored in buffer model, clustering processing is carried out to the gesture data picture frame in buffer model, gesture keyframe sequence is obtained.
As shown in Fig. 2 specific implementation process is as follows:In video processing procedure is carried out, preparatory stage and the useless data frame of Restoration stage are removed as much as possible, identification is trained with the data frame in crucial gesture stage, can so greatly improve training and recognition speed.Video flow processing of the present invention is i.e. according to this principle, using the asynchronous clustering processing model based on interval gesture data frame buffering, the caching of images of gestures frame is captured to camera, compare and cluster to realize, data in poly- each class out are exactly the total data of the preparatory stage of a basic gesture, key frame stage and Restoration stage, therefore each basic hand-type action of gesture word can be represented with each class cluster centre.
Specifically, Video processing of the present invention includes following process:
First, the use of asynchronous clustering processing model.
As shown in figure 3, specific implementation process is as follows:
Buffer model:Essence is a data acquisition buffer model.System according to configuration file, the memory headroom of system distribution gesture data frame sign as buffer model, camera by real-time capture to image data frame be sequentially stored in buffer model, when buffer model completely starts to carry out clustering processing to gesture data frame.
Asynchronous process:While gesture data frame clustering processing, constantly catch new data frame from camera and enter buffer model.When buffer model gesture data frame clustering processing speed is less than data frame capture velocity, in fact it could happen that the full situation of buffer model data, now obstruction catches data frame process, waits gesture data frame to be clustered processing, and point out user the size for increasing configuration file;When buffer model gesture data frame clustering processing speed is more than data frame capture velocity, it is likely to occur the empty situation of buffer model gesture data frame, now obstruction is spaced frame clustering processing process, gesture data frame is put into buffering area etc. gesture data frame process to be captured, and points out user the size for reducing configuration file.
Gesture interval frame cluster based on time series:Clustering algorithm based on time series reads data from buffer model, comparison interval data frame, according to threshold parameter in comparative result and configuration file, if frame difference is not less than threshold parameter, then remove identical data frame in buffer model, if frame difference exceedes threshold parameter, be defined as drawing class border, and then the data frame for choosing such Median Position removes whole class as such critical data frame from buffer model.
2nd, the poor computational methods of images of gestures interval data frame
The system calculates frame poor according to images of gestures data frame feature using the computational methods based on color histogram.Circular is as follows:
(1) the related frame difference method of histogram:CORREL_D parameters in correspondence configuration file
Fd ( H 1 , H 2 ) = Σ i ( H 1 ′ ( i ) · H 2 ′ ( i ) ) Σ i ( H 1 ′ ( i ) 2 ) · Σ i ( H 2 ′ ( i ) 2 ) ,
Wherein,
Figure BDA0000105677990000052
N is the sum of a frame pixel, Fd (H1, H2) represent be two images of gestures compared color histogram difference, H1For the color histogram calculated value for first images of gestures to be compared, H2For the color histogram calculated value for second images of gestures to be compared, i, j, k is intermediate variable.
(2)χ2Examine difference of histograms method:
Fd ( H 1 , H 2 ) = Σ i H 1 ( i ) - H 2 ( i ) H 1 ( i ) + H 2 ( i )
Wherein, Fd (H1, H2) represent be two images of gestures compared color histogram difference, H1For the color histogram calculated value for first images of gestures to be compared, H2For the color histogram calculated value for second images of gestures to be compared, i is intermediate variable.
(3) the minimum frame difference method of histogram:
Fd ( H 1 , H 2 ) = Σ i min ( H 1 ( i ) , H 2 ( i ) )
Wherein, min (H1(i), H2(i)) represent to take H1(i), H2(i) minimum value between, Fd (H1, H2) represent be two images of gestures compared color histogram difference, H1For the color histogram calculated value for first images of gestures to be compared, H2For the color histogram calculated value for second images of gestures to be compared, i is intermediate variable.
(4) histogram interframe Bhattacharyya distance method:BHATTACHARYYA_D in correspondence configuration file, the method is only applicable to normalization histogram.
Wherein, min (H1(i), H2(i)) represent to take H1(i), H2(i) minimum value between, Fd (H1, H2) represent be two images of gestures compared color histogram difference, H1For the color histogram calculated value for first images of gestures to be compared, H2For the color histogram calculated value for second images of gestures to be compared, i is intermediate variable.
System using the logical AND of 4 kinds of comparative approach results in above color histogram computational methods by, as the comparison parameter of images of gestures interval frame, realizing Clustering Effect.
3rd, the interval frame cluster based on time series
Cluster configuration parameter as follows:
BUFFER_SIZE fields are system buffer model size, are accurate to integer-bit;
CORREL_D fields are that color histogram correlation frame difference method compares threshold value, are accurate to 2 significant digits;
CHISQR_D fields are χ2Examine color histogram frame difference method to compare threshold value, be accurate to 2 significant digits;
INTERSECT_D fields are that the minimum frame difference method of color histogram compares threshold value, are accurate to 2 significant digits;
BHATTACHARYYA_D fields are that color histogram interframe Bhattacharyya distance method compares threshold value, are accurate to 2 significant digits;
INTERVALFRAMES_VALUE fields are that buffer model data frame compares hop count, are accurate to integer-bit.Arthmetic statement is as follows:
(1) initialize.System allocated size for (BUFFER_SIZE*SIZEOF (gesture data frame)) memory headroom as buffer model, camera real-time capture to image data frame to be stored in element in buffer model U, U in chronological order be X0, X1, X2..., XBUFFER_SIZE-1
(2) frame is compared poor.Order calculates X in Ui(i-th of element, 0≤i < BUFFER_SIZE) and Xi+INTERVALFRAMES_VALUE((i+INTERVALFRAMES_VALUE) individual element) data frame difference dI, i+INTERVALFRAMES_VALUE, cluster boundary is defined as if following condition is very, same category is otherwise defined as;
(3)(dI, i+INTERVALFRAMES_VALUE≥CORREL_D)           &&
   (dI, i+INTERVALFRAMES_VALUE≥CHISQR_D)           &&
   (dI, i+INTERVALFRAMES_VALUE≥INTERSECT_D)        &&
   (dI, i+INTERVALFRAMES_VALUE≥BHATTACHARYYA_D)
(4) cluster centre.Class set T (M), SIZEOF (T (M))=k, (TM) are obtained after a certain moment T, U clusterj(0≤j < k), chooses (TM)jElementFor class center (For class (TM)jMedian Position element), and as class (TM)jCritical data frame, it is right
Figure BDA0000105677990000073
Carry out after gesture analysis and the processing of gesture identification process, class (TM) is deleted from UjAll data frames;
(5) asynchronous process.While data clusters are read from U obtains gesture key frame, while writing gesture video stream data into U;
(6) deadlock is handled.If data are sky in U, process blocking is clustered, untill there are data in U;If data are full in model, data process obstruction is caught, and reminds user suitably to adjust BUFFERSIZE parameters in configuration file, to increase U size;
(7) algorithm terminates.System receives seizure gesture video flowing and ceased and desisted order, and stops writing data into U, and element in U is disposed, and terminates.
Step 300:Gesture analysis, i.e.,:After gesture key frame images binaryzation and smothing filtering denoising, carry out gesture contours extract, crucial gesture outline data is obtained, according to gesture key frame images binary map and the gesture contours extract gesture feature parameter of proposition, the gesture feature parameter composition gesture feature vector.
As shown in figure 4, specific implementation process is as follows:The input data of image pre-processing module is gesture key frame, and output data is gesture frame outline data, and data processing step includes images of gestures binaryzation, smothing filtering denoising gesture area and the part of gesture contours extract three.Then according to gesture key frame images binary map and the gesture contours extract gesture feature parameter of proposition, the gesture feature parameter composition gesture feature vector
First, images of gestures binaryzation
The system carries out binary conversion treatment according to the Skin Color Information of human body to images of gestures, obtains the gesture area of people.Specific algorithm is that each pixel of gained image after cluster segmentation is proceeded as follows:
(1) Hue the and Satisfaction values of the pixel are calculated by rgb value;
(2) judge Hue the and Satisfaction values of the pixel whether in the colour of skin of the people is interval;
(3) if the pixel is changed into black if;
(4) pixel is otherwise changed into white;
Wherein:Hue and Satisfaction data types are single precision, without unit.
By binary conversion treatment, gesture area binary map is obtained.
2nd, smothing filtering denoising gesture area
Gesture binary map after binaryzation is primarily present salt-pepper noise, is mainly shown as there is chequered with black and white bright dim spot in image.The system carries out denoising using medium filtering and linear smoothing filtering to gesture area binary map, obtains clearly gesture area.
3rd, gesture contours extract
The system carries out gesture contours extract using Laplce's Boundary extracting algorithm.Laplace operator is a kind of second dervative scalar operator that computing is carried out to two-dimensional function, more sensitive to the noise in image.When being handled, gesture edge can produce a precipitous zero crossing, for noiseless and with brink image, gesture contours extract can be come out by Laplace operator.
4th, gesture feature parameter extraction technology
The system carries out gesture model characteristic parameter extraction to the gesture binary map and gesture profile that are obtained after pretreatment.Totally three groups of the gesture feature parameter to be extracted, respectively gesture area feature, Hu invariant moment features and Fourier description, three groups of characteristic parameters collectively constitute system features vector.
Gesture area feature includes two sections ratio, the hand region area represented with number of pixels, the gesture profile girth that gesture area area splits gesture boundary rectangle with gesture rectangular area ratio, gesture area length and width ratio, gesture binary map center of gravity.
Hu invariant moment features
Moment of the orign or central moment are directly used as the feature of image, it is impossible to ensure feature while having translation, Invariant to rotation and scale.If in fact, only representing the feature of image with central moment, feature only has translation invariance;If using central moment is normalized, feature not only has translation invariance, but also with constant rate, but without rotational invariance.M.K.Hu proposes not bending moment concept, provide the definition of continuous function square and the fundamental property on square, demonstrate the properties such as the translation invariance about square, rotational invariance and constant rate, and show in particular the expression formula of seven with translation invariance, rotational invariance and constant rate not bending moments.Seven not bending moment be made up of the linear combination of second order and third central moment, expression formula such as formula:
φ12002
φ2=(μ2002)2+(2μ11)2
φ3=(μ30-3μ12)2+(3μ2103)2
φ4=(μ3012)2+(μ2103)2
φ5=(μ30-3μ12)(μ3012)[(μ3012)2-3(μ2103)2]+(3μ2103)(μ2103)[3(μ3012)2-(μ2103)2]
φ6=(μ2002)[(μ3012)2-(μ2103)2]+4μ113012)(μ2103)
φ7=(3 μ2103)(μ3012)[(μ3012)2-3(μ2103)2]-(μ30-3μ12)(μ2103)[3(μ3012)2-(μ2103)2]
Normalized central moment
Figure BDA0000105677990000091
WhereinP+q=2,3 ... ..
Wherein:φ1, φ2, φ3, φ4, φ5, φ6, φ7For constant matrix.
Fourier describes subcharacter extraction
It is as follows that Fourier describes subcharacter extraction flow:
1) Fourier transformation obtains the Fourier coefficient of gesture contour curve point sequence;
2) Fourier inversion obtains the expression coefficient of gesture contour curve point sequence in a frequency domain;
3) noise of gesture contour curve point sequence in a frequency domain is removed, in order to smoothed curve;
4) a gesture contour curve is obtained to going the gesture contour curve point sequence frequency domain representation after noise to carry out inverse Fourier transform again, takes 12 coefficients for representing this curve, be used as gesture feature.
Fourier descriptors (Fourier Deseriptor, abbreviation FD) it is commonly used to represent the shape facility of single closed curve, its basic thought is that objective contour curve is modeled as into one-dimensional sequence, one-dimensional Fourier transform is carried out to the sequence, so as to obtain a series of fourier coefficient, the objective contour is described with these coefficients.
The advantage of Fourier descriptors method is mainly reflected in Computing Principle simply, and description is clear, with by coarse and fine characteristic.Computing Principle can simply cause feature extraction more to stablize, and a large amount of control parameters need not be set to be obtained with result during calculating, the uniformity of calculating is good.
The basic thought of Fourier descriptors assumes that body form is a closed curve, and all point sequences are along along boundary curve:{ x (l), y (l):L=0,1 ..., n-1 }, it is expressed as with plural form:
P (l)=x (l)+j*y (l) (l=0,1 ..., n-1,
Figure BDA0000105677990000101
)
So, border can just be represented in the one-dimensional space.
The discrete Fourier coefficient of one-dimensional sequence is represented with set Z, is defined as follows:
Z = z ( k ) = 1 n Σ 1 = 0 n - 1 p ( l ) exp ( - j 2 πlk n ) (k=0,1 ... .n-1)
Z is p Fourier transform, is the expression of point sequence in a frequency domain.Its inverse fourier transform coefficient is represented with set P, is defined as follows:
P = p ( l ) = Σ k = 0 n - 1 z ( k ) exp ( j 2 πlk n ) (l=0,1 ... .n-1)
Utilize the property of fourier coefficient
Figure BDA0000105677990000111
(
Figure BDA0000105677990000112
Z (k) conjugate complex number), in coefficient sets Z, eliminate from
Figure BDA0000105677990000113
Radio-frequency component in the range of to n-K-1.Inverse fourier transform is carried out again, will obtain the curve approximate with virgin curve, but Mutational part will be smoothened in virgin curve, and this curve of approximation is referred to as the K curve of approximation of virgin curve.Fourier coefficient subset { z (k) under this meaning:K≤K } it is referred to as Fourier descriptors (Fourier Descriptor).Because fourier coefficient has the characteristic that energy is concentrated to low frequency, therefore the purpose for distinguishing different shape border can be just reached with less coefficient.
Fourier descriptors are relevant with the yardstick of shape, direction and curve initial point position.In order to recognize have rotation, translation and scale invariance shape, it is necessary to which Fourier descriptors are normalized.According to Fourier transform property, body form border initial point position is translated into a length, r times of object amplification, anglec of rotation duty and translational displacement (x0, y0) after, the Fourier transform coefficient collection of new shape shares Z ' expressions, is defined as follows:
Figure BDA0000105677990000114
Figure BDA0000105677990000115
Wherein k=0,1 ... .n-1, x ' (l)+iy ' (l)=x (l+a)+iy (l+a)
As can be seen from the above equation, when describing shape with fourier coefficient, coefficient amplitude ‖ z (k) ‖, k=0,1 ... .n-1, with rotational invariance and translation invariance (z (0) does not have translation invariance), and it is unrelated with the selection of the origin of curve.When object translation, only change its z (0) component F (x0+iy0) value.Amplitude ‖ z (k) ‖ divided by ‖ z (1) ‖ of each coefficient (except z (0)), then
Figure BDA0000105677990000117
Do not change with change of scale, so
Figure BDA0000105677990000118
K=1,2 ... .., n-1, ‖ ‖ represent modulus.
Figure BDA0000105677990000119
There is rotation simultaneously, translation and scale invariance, and unrelated with the selection of origin of curve position, we are using it as Fourier descriptors, therefore normalized Fourier descriptors d (k) is defined as:
d ( k ) = | | z ( k ) | | | | z ( l ) | | , K=1,2 ... n-1, ‖ ‖ represent modulus
The system removes preceding 12 coefficients of z (0) outside, obtains the characteristic vector of one 12 dimension, and this feature vector has rotation, translation and scale invariance, put down and unrelated with the selection of origin of curve position.
Step 400:Gesture identification, i.e.,:Gesture identification is carried out according to gesture feature vector, corresponding with identification gesture key letter virtual key or Keyboard Control virtual key is obtained.
As shown in figure 5, specific implementation process is as follows:Identification learning process uses BP neural network as disaggregated model, and combines Bayesian Classification Model progress contrast experiment.The gesture that system can be recognized includes 26 English alphabets and 4 control gestures, and control gesture is respectively Spaee (correspondence and keyboard strip space bar), StartIME (Input Method Editor:Input method switches, correspondence keyboard shift key), End (correspondence keyboard End keys) and Backspace (correspondence and keyboard Backspace keys).
Step 500:Gesture input method is built, i.e.,:According to obtained key letter virtual key and the existing input method of Keyboard Control virtual key call operation system.
As shown in figure 5, specific implementation process is as follows:During gesture input, a mark gesture is set, cutting is carried out to video flowing according to mark gesture.Send respective virtual key to reach the input of control system input method to operating system according to the result obtained from system cutting cluster and Classification and Identification.
As shown in Fig. 2 the embodiment of the present invention is:Build a kind of gesture input method constructing system based on computer vision, including gather the gesture collecting unit 1 of gesture vision signal, the video processing unit 2 handled the gesture vision signal of collection, the gesture analysis unit 3 for carrying out gesture analysis and obtaining gesture feature vector, the gesture identification unit 4 for carrying out according to gesture feature vector gesture identification, structure gesture input method construction unit 5, the gesture collecting unit 1 catches gesture vision signal, obtains gesture data picture frame;The video processing unit 2 is using the asynchronous clustering processing model based on interval gesture data frame buffering, the asynchronous clustering processing model includes buffer model, the gesture data picture frame of collection is sequentially stored in buffer model, clustering processing is carried out to the gesture data picture frame in buffer model, gesture keyframe sequence is obtained;After the gesture analysis unit 3 is to gesture key frame images binaryzation and smothing filtering denoising, carry out gesture contours extract, obtain crucial gesture outline data, according to gesture key frame images binary map and the gesture contours extract gesture feature parameter of proposition, the gesture feature parameter composition gesture feature vector;The gesture identification unit 4 carries out gesture identification according to gesture feature vector, obtains corresponding with identification gesture key letter virtual key or Keyboard Control virtual key;The construction unit 5 is according to obtained key letter virtual key and the existing input method of Keyboard Control virtual key call operation system.
Specific implementation process of the present invention is as follows:The gesture collecting unit 1 catches gesture vision signal using camera, and gesture data picture frame is obtained by processing.
The course of work of video processing unit 2 is as follows:
In video processing procedure is carried out, preparatory stage and the useless data frame of Restoration stage are removed as much as possible, identification is trained with the data frame in crucial gesture stage, can so greatly improve training and recognition speed.Video flow processing of the present invention is i.e. according to this principle, using the asynchronous clustering processing model based on interval gesture data frame buffering, the caching of images of gestures frame is captured to camera, compare and cluster to realize, data in poly- each class out are exactly the total data of the preparatory stage of a basic gesture, key frame stage and Restoration stage, therefore each basic hand-type action of gesture word can be represented with each class cluster centre.
Specifically, Video processing of the present invention includes following process:
First, the use of asynchronous clustering processing model.
As shown in figure 3, specific implementation process is as follows:
Buffer model:Essence is a data acquisition buffer model.System according to configuration file, the memory headroom of system distribution gesture data frame sign as buffer model, camera by real-time capture to image data frame be sequentially stored in buffer model, when buffer model completely starts to carry out clustering processing to gesture data frame.
Asynchronous process:While gesture data frame clustering processing, constantly catch new data frame from camera and enter buffer model.When buffer model gesture data frame clustering processing speed is less than data frame capture velocity, in fact it could happen that the full situation of buffer model data, now obstruction catches data frame process, waits gesture data frame to be clustered processing, and point out user the size for increasing configuration file;When buffer model gesture data frame clustering processing speed is more than data frame capture velocity, it is likely to occur the empty situation of buffer model gesture data frame, now obstruction is spaced frame clustering processing process, gesture data frame is put into buffering area etc. gesture data frame process to be captured, and points out user the size for reducing configuration file.
Gesture interval frame cluster based on time series:Clustering algorithm based on time series reads data from buffer model, comparison interval data frame, according to threshold parameter in comparative result and configuration file, if frame difference is not less than threshold parameter, then remove identical data frame in buffer model, if frame difference exceedes threshold parameter, be defined as drawing class border, and then the data frame for choosing such Median Position removes whole class as such critical data frame from buffer model.
2nd, the poor computational methods of images of gestures interval data frame
The system calculates frame poor according to images of gestures data frame feature using the computational methods based on color histogram.Circular is as follows:
(1) the related frame difference method of histogram:CORREL_D parameters in correspondence configuration file
Fd ( H 1 , H 2 ) = Σ i ( H 1 ′ ( i ) · H 2 ′ ( i ) ) Σ i ( H 1 ′ ( i ) 2 ) · Σ i ( H 2 ′ ( i ) 2 ) ,
Wherein,
Figure BDA0000105677990000142
N is the sum of a frame pixel, Fd (H1, H2) represent be two images of gestures compared color histogram difference, H1For the color histogram calculated value for first images of gestures to be compared, H2For the color histogram calculated value for second images of gestures to be compared, i, j, k is intermediate variable.
(2)χ2Examine difference of histograms method:
Fd ( H 1 , H 2 ) = Σ i H 1 ( i ) - H 2 ( i ) H 1 ( i ) + H 2 ( i )
Wherein, Fd (H1, H2) represent be two images of gestures compared color histogram difference, H1For the color histogram calculated value for first images of gestures to be compared, H2For the color histogram calculated value for second images of gestures to be compared, i is intermediate variable.
(3) the minimum frame difference method of histogram:
Fd ( H 1 , H 2 ) = Σ i min ( H 1 ( i ) , H 2 ( i ) )
Wherein, min (H1(i), H2(i)) represent to take H1(i), H2(i) minimum value between, Fd (H1, H2) represent be two images of gestures compared color histogram difference, H1For the color histogram calculated value for first images of gestures to be compared, H2For the color histogram calculated value for second images of gestures to be compared, i is intermediate variable.
(4) histogram interframe Bhattacharyya distance method:BHATTACHARYYA_D in correspondence configuration file, the method is only applicable to normalization histogram.
Wherein, min (H1(i), H2(i)) represent to take H1(i), H2(i) minimum value between, Fd (H1, H2) represent be two images of gestures compared color histogram difference, H1For the color histogram calculated value for first images of gestures to be compared, H2For the color histogram calculated value for second images of gestures to be compared, i is intermediate variable.
System using the logical AND of 4 kinds of comparative approach results in above color histogram computational methods by, as the comparison parameter of images of gestures interval frame, realizing Clustering Effect.
3rd, the interval frame cluster based on time series
Cluster configuration parameter as follows:
BUFFER_SIZE fields are system buffer model size, are accurate to integer-bit;
CORREL_D fields are that color histogram correlation frame difference method compares threshold value, are accurate to 2 significant digits;
CHISQR_D fields are χ2Examine color histogram frame difference method to compare threshold value, be accurate to 2 significant digits;
INTERSECT_D fields are that the minimum frame difference method of color histogram compares threshold value, are accurate to 2 significant digits;
BHATTACHARYYA_D fields are that color histogram interframe Bhattacharyya distance method compares threshold value, are accurate to 2 significant digits;
INTERVALFRAMES_VALUE fields are that buffer model data frame compares hop count, are accurate to integer-bit.
Arthmetic statement is as follows:
(8) initialize.System allocated size for (BUFFER_SIZE*SIZEOF (gesture data frame)) memory headroom as buffer model, camera real-time capture to image data frame to be stored in element in buffer model U, U in chronological order be X0, X1, X2..., XBUFFER_SIZE-1
(9) frame is compared poor.Order calculates X in Ui(i-th of element, 0≤i < BUFFER_SIZE) and Xi+INTERVALFRAMES_VALUE((i+INTERVALFRAMES_VALUE) individual element) data frame difference dI, i+INTERVALFRAMES_VALUE, cluster boundary is defined as if following condition is very, it is no
Then it is defined as same category;
(10)(dI, i+INTERVALFRAMES_VALUE≥CORREL_D)        &&
    (dI, i+INTERVALFRAMES_VALUE≥CHISQR_D)        &&
    (dI, i+INTERVALFRAMES_VALUE≥INTERSECT_D)     &&
    (dI, i+INTERVALFRAMES_VALUE≥BHATTACHARYYA_D)
(11) cluster centre.Class set T (M), SIZEOF (T (M))=k, (TM) are obtained after a certain moment T, U clusterj(0≤j < k), chooses (TM)jElement
Figure BDA0000105677990000161
For class center (
Figure BDA0000105677990000162
For class (TM)jMedian Position element), and as class (TM)jCritical data frame, it is right
Figure BDA0000105677990000163
Carry out after gesture analysis and the processing of gesture identification process, class (TM) is deleted from UjAll data frames;
(12) asynchronous process.While data clusters are read from U obtains gesture key frame, while writing gesture video stream data into U;
(13) deadlock is handled.If data are sky in U, process blocking is clustered, untill there are data in U;If data are full in model, data process obstruction is caught, and reminds user suitably to adjust BUFFERSIZE parameters in configuration file, to increase U size;
(14) algorithm terminates.System receives seizure gesture video flowing and ceased and desisted order, and stops writing data into U, and element in U is disposed, and terminates.
The course of work of gesture analysis unit 3 is as follows:
As shown in figure 4, specific implementation process is as follows:The input data of image pre-processing module is gesture key frame, and output data is gesture frame outline data, and data processing step includes images of gestures binaryzation, smothing filtering denoising gesture area and the part of gesture contours extract three.Then according to gesture key frame images binary map and the gesture contours extract gesture feature parameter of proposition, the gesture feature parameter composition gesture feature vector
First, images of gestures binaryzation
The system carries out binary conversion treatment according to the Skin Color Information of human body to images of gestures, obtains the gesture area of people.Specific algorithm is that each pixel of gained image after cluster segmentation is proceeded as follows:
(5) Hue the and Satisfaction values of the pixel are calculated by rgb value;
(6) judge Hue the and Satisfaction values of the pixel whether in the colour of skin of the people is interval;
(7) if the pixel is changed into black if;
(8) pixel is otherwise changed into white;
Wherein:Hue and Satisfaction data types are single precision, without unit.
By binary conversion treatment, gesture area binary map is obtained.
2nd, smothing filtering denoising gesture area
Gesture binary map after binaryzation is primarily present salt-pepper noise, is mainly shown as there is chequered with black and white bright dim spot in image.The system carries out denoising using medium filtering and linear smoothing filtering to gesture area binary map, obtains clearly gesture area.
3rd, gesture contours extract
The system carries out gesture contours extract using Laplce's Boundary extracting algorithm.Laplace operator is a kind of second dervative scalar operator that computing is carried out to two-dimensional function, more sensitive to the noise in image.When being handled, gesture edge can produce a precipitous zero crossing, for noiseless and with brink image, gesture contours extract can be come out by Laplace operator.
4th, gesture feature parameter extraction technology
The system carries out gesture model characteristic parameter extraction to the gesture binary map and gesture profile that are obtained after pretreatment.Totally three groups of the gesture feature parameter to be extracted, respectively gesture area feature, Hu invariant moment features and Pourier description, three groups of characteristic parameters collectively constitute system features vector.
Gesture area feature includes two sections ratio, the hand region area represented with number of pixels, the gesture profile girth that gesture area area splits gesture boundary rectangle with gesture rectangular area ratio, gesture area length and width ratio, gesture binary map center of gravity.
Hu invariant moment features
Moment of the orign or central moment are directly used as the feature of image, it is impossible to ensure feature while having translation, Invariant to rotation and scale.If in fact, only representing the feature of image with central moment, feature only has translation invariance;If using central moment is normalized, feature not only has translation invariance, but also with constant rate, but without rotational invariance.M.K.Hu proposes not bending moment concept, provide the definition of continuous function square and the fundamental property on square, demonstrate the properties such as the translation invariance about square, rotational invariance and constant rate, and show in particular the expression formula of seven with translation invariance, rotational invariance and constant rate not bending moments.Seven not bending moment be made up of the linear combination of second order and third central moment, expression formula such as formula:
φ12002
φ2=(μ2002)2+(2μ11)2
φ3=(μ30-3μ12)2+(3μ2103)2
φ4=(μ3012)2+(μ2103)2
φ5=(μ30-3μ12)(μ3012)[(μ3012)2-3(μ2103)2]+(3μ2103)(μ2103)[3(μ3012)2-(μ2103)2]
φ6=(μ2002)[(μ3012)2-(μ2103)2]+4μ113012)(μ2103)
φ7=(3 μ2103)(μ3012)[(μ3012)2-3(μ2103)2]-(μ30-3μ12)(μ2103)[3(μ3012)2-(μ2103)2]
Normalized central moment
Figure BDA0000105677990000181
Wherein
Figure BDA0000105677990000182
P+q=2,3 ... ..
φ1, φ2, φ3, φ4, φ5, φ6, φ7For constant matrix, uijElement variable in representing matrix.
Fourier describes subcharacter extraction
It is as follows that Fourier describes subcharacter extraction flow:
Fourier transformation obtains the Fourier coefficient of gesture contour curve point sequence;
Fourier inversion obtains the expression coefficient of gesture contour curve point sequence in a frequency domain;
Remove the noise of gesture contour curve point sequence in a frequency domain, in order to smoothed curve;
Again to going the gesture contour curve point sequence frequency domain representation after noise to carry out inverse Fourier transform, a gesture contour curve is obtained, 12 coefficients for representing this curve is taken, is used as gesture feature.
Fourier descriptors (Fourier Descriptor, abbreviation FD) it is commonly used to represent the shape facility of single closed curve, its basic thought is that objective contour curve is modeled as into one-dimensional sequence, one-dimensional Fourier transform is carried out to the sequence, so as to obtain a series of fourier coefficient, the objective contour is described with these coefficients.
The advantage of Fourier descriptors method is mainly reflected in Computing Principle simply, and description is clear, with by coarse and fine characteristic.Computing Principle can simply cause feature extraction more to stablize, and a large amount of control parameters need not be set to be obtained with result during calculating, the uniformity of calculating is good.
The basic thought of Fourier descriptors assumes that body form is a closed curve, and all point sequences are along along boundary curve:{ x (l), y (l):L=0,1 ..., n-1 }, it is expressed as with plural form:
P (l)=x (l)+j*y (l) (l=0,1 ..., n-1,
Figure BDA0000105677990000191
)
So, border can just be represented in the one-dimensional space.
The discrete Fourier coefficient of one-dimensional sequence is represented with set Z, is defined as follows:
Z = z ( k ) = 1 n Σ 1 = 0 n - 1 p ( l ) exp ( - j 2 πlk n ) (k=0,1 ... .n-1)
Z is p Fourier transform, is the expression of point sequence in a frequency domain.Its inverse fourier transform coefficient is represented with set P, is defined as follows:
P = p ( l ) = Σ k = 0 n - 1 z ( k ) exp ( j 2 πlk n ) (l=0,1 ... .n-1)
Utilize the property of fourier coefficient
Figure BDA0000105677990000194
(
Figure BDA0000105677990000195
Z (k) conjugate complex number), in coefficient sets Z, eliminate from
Figure BDA0000105677990000196
Radio-frequency component in the range of to n-K-1.Inverse fourier transform is carried out again, will obtain the curve approximate with virgin curve, but Mutational part will be smoothened in virgin curve, and this curve of approximation is referred to as the K curve of approximation of virgin curve.Fourier coefficient subset { z (k) under this meaning:K≤K } it is referred to as Fourier descriptors (Fourier Descriptor).Because fourier coefficient has the characteristic that energy is concentrated to low frequency, therefore the purpose for distinguishing different shape border can be just reached with less coefficient.
Fourier descriptors are relevant with the yardstick of shape, direction and curve initial point position.In order to recognize have rotation, translation and scale invariance shape, it is necessary to which Fourier descriptors are normalized.According to Fourier transform property, body form border initial point position is translated into a length, r times of object amplification, anglec of rotation duty and translational displacement (x0, y0) after, the Fourier transform coefficient collection of new shape shares Z ' expressions, is defined as follows:
Figure BDA0000105677990000201
Figure BDA0000105677990000202
Figure BDA0000105677990000203
Wherein k=0,1 ... .n-1, x ' (l)+iy ' (l)=x (l+a)+iy (l+a)
As can be seen from the above equation, when describing shape with fourier coefficient, coefficient amplitude ‖ z (k) ‖, k=0,1 ... .n-1, with rotational invariance and translation invariance (z (0) does not have translation invariance), and it is unrelated with the selection of the origin of curve.When object translation, only change its z (0) component F (x0+iy0) value.Amplitude ‖ z (k) ‖ divided by ‖ z (1) ‖ of each coefficient (except z (0)), then
Figure BDA0000105677990000204
Do not change with change of scale, so
Figure BDA0000105677990000205
K=1,2 ... .., n-1, ‖ ‖ represent modulus.There is rotation simultaneously, translation and scale invariance, and unrelated with the selection of origin of curve position, we are using it as Fourier descriptors, therefore normalized Fourier descriptors d (k) is defined as:
d ( k ) = | | z ( k ) | | | | z ( l ) | | , K=1,2 ... n-1, ‖ ‖ represent modulus
The system removes preceding 12 coefficients of z (0) outside, obtains the characteristic vector of one 12 dimension, and this feature vector has rotation, translation and scale invariance, put down and unrelated with the selection of origin of curve position.
The course of work of recognition unit 4 is as follows:
As shown in figure 5, specific implementation process is as follows:Identification learning process uses BP neural network as disaggregated model, and combines Bayesian Classification Model progress contrast experiment.The gesture that system can be recognized includes 26 English alphabets and 4 control gestures, and control gesture is respectively Space (correspondence and keyboard strip space bar), StartIME (Input Method Editor:Input method switches, correspondence keyboard shift key), End (correspondence keyboard End keys) and Backspace (correspondence and keyboard Backspace keys).
The course of work of construction unit 5 is as follows:
As shown in figure 5, specific implementation process is as follows:During gesture input, a mark gesture is set, cutting is carried out to video flowing according to mark gesture.Send respective virtual key to reach the input of control system input method to operating system according to the result obtained from system cutting cluster and Classification and Identification.
The solution have the advantages that:Build a kind of gesture input method construction method based on Video clustering processing and system.Gesture frame is clustered using asynchronous buffer model, crucial gesture frame is obtained, only by the follow-up gesture analysis of crucial gesture frame progress and gesture identification, relative to conventional analysis and the system of each frame gesture of identification, system response real-time is effectively increased.In addition, the present invention is by the connected applications of gesture identification and input method, it is the innovation in man-machine interaction history.
Above content is to combine specific preferred embodiment further description made for the present invention, it is impossible to assert that the specific implementation of the present invention is confined to these explanations.For general technical staff of the technical field of the invention, without departing from the inventive concept of the premise, some simple deduction or replace can also be made, protection scope of the present invention should be all considered as belonging to.

Claims (10)

1. a kind of gesture input method construction method based on computer vision, step is as follows
Gather gesture:Gesture vision signal is caught, gesture data picture frame is obtained;
Video processing:Using the asynchronous clustering processing model based on interval gesture data frame buffering, the asynchronous clustering processing model includes buffer model, the gesture data picture frame of collection is sequentially stored in buffer model, clustering processing is carried out to the gesture data picture frame in buffer model, gesture keyframe sequence is obtained;
Gesture analysis:After gesture key frame images binaryzation and smothing filtering denoising, carry out gesture contours extract, crucial gesture outline data is obtained, according to gesture key frame binary map and the gesture contours extract gesture feature parameter of proposition, the gesture feature parameter composition gesture feature vector;
Gesture identification:Gesture identification is carried out according to gesture feature vector, corresponding with identification gesture key letter virtual key or Keyboard Control virtual key is obtained;
Build gesture input method:According to obtained key letter virtual key and the existing input method of Keyboard Control virtual key call operation system.
2. the gesture input method construction method based on computer vision according to claim 1, it is characterized in that, in video-processing steps, clustering processing is carried out to the gesture data picture frame in buffer model, obtained each class includes the total data of the preparatory stage, key frame stage and Restoration stage of gesture, and the cluster centre of each class is the gesture key frame of each gesture.
3. the gesture input method construction method based on computer vision according to claim 2, it is characterised in that in video-processing steps, threshold parameter is set, frame by comparison interval data frame is poor, if frame difference removes the gesture data frame not less than threshold parameter in buffer model;If frame difference exceedes threshold parameter, it is defined as drawing class border, and choose the gesture data frame of such Median Position as such gesture key frame.
4. the gesture input method construction method based on computer vision according to claim 1, it is characterized in that, in video-processing steps, the gesture data picture frame of collection is sequentially stored in buffer model, when buffer model completely starts to carry out clustering processing to gesture data frame.
5. the gesture input method construction method based on computer vision according to claim 4, it is characterized in that, in video-processing steps, while gesture data frame clustering processing, constantly enter buffer model from the new gesture data frame of seizure, when the gesture data frame clustering processing speed in buffer model is less than gesture data frame capture velocity, when buffer model data are full, wait gesture data frame to be clustered processing, and point out user the size for increasing configuration file;When buffer model gesture data frame clustering processing speed is more than gesture data frame capture velocity, is there is the buffer model data empty time, waiting gesture data frame to be captured that gesture data frame is put into buffer model, and point out user to reduce the size in configuration file.
6. the gesture input method construction method based on computer vision according to claim 1, it is characterised in that the gesture feature parameter includes gesture area feature, Hu invariant moment features and Fourier description.
7. the gesture input method construction method based on computer vision according to claim 1, it is characterised in that in gesture analysis step, gesture contours extract is carried out using Laplce's Boundary extracting algorithm.
8. a kind of gesture input method constructing system based on computer vision, it is characterized in that, including gather the gesture collecting unit of gesture vision signal, the video processing unit handled the gesture vision signal of collection, the gesture analysis unit for carrying out gesture analysis and obtaining gesture feature vector, the gesture identification unit for carrying out according to gesture feature vector gesture identification, structure gesture input method construction unit, the gesture collecting unit catches gesture vision signal, obtains gesture data picture frame;The video processing unit uses the asynchronous clustering processing model based on interval gesture data frame buffering, the asynchronous clustering processing model includes buffer model, the gesture data picture frame of collection is sequentially stored in buffer model, clustering processing is carried out to the gesture data picture frame in buffer model, gesture keyframe sequence is obtained;After the gesture analysis unit is to gesture key frame images binaryzation and smothing filtering denoising, carry out gesture contours extract, obtain crucial gesture outline data, according to gesture key frame images binary map and the gesture contours extract gesture feature parameter of proposition, the gesture feature parameter composition gesture feature vector;The gesture identification unit carries out gesture identification according to gesture feature vector, obtains corresponding with identification gesture key letter virtual key or Keyboard Control virtual key;The construction unit is according to obtained key letter virtual key and the existing input method of Keyboard Control virtual key call operation system.
9. the gesture input method constructing system based on computer vision according to claim 8, it is characterized in that, the video processing unit includes carrying out the gesture data picture frame in buffer model the clustering processing module of clustering processing, the clustering processing module reads gesture frame data using the clustering algorithm based on time series from buffer model, and gesture key frame is obtained by being spaced gesture frame frame difference.
10. the gesture input method constructing system based on computer vision according to claim 8, it is characterized in that, the gesture identification unit carries out gesture identification according to gesture feature vector, obtains and the identification corresponding 26 English letter of keyboard virtual keys of gesture and 4 Keyboard Control virtual keys.
CN2011103459148A 2011-11-04 2011-11-04 Computer-vision-based gesture input method construction method and system Pending CN102508547A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2011103459148A CN102508547A (en) 2011-11-04 2011-11-04 Computer-vision-based gesture input method construction method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2011103459148A CN102508547A (en) 2011-11-04 2011-11-04 Computer-vision-based gesture input method construction method and system

Publications (1)

Publication Number Publication Date
CN102508547A true CN102508547A (en) 2012-06-20

Family

ID=46220647

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2011103459148A Pending CN102508547A (en) 2011-11-04 2011-11-04 Computer-vision-based gesture input method construction method and system

Country Status (1)

Country Link
CN (1) CN102508547A (en)

Cited By (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102968642A (en) * 2012-11-07 2013-03-13 百度在线网络技术(北京)有限公司 Trainable gesture recognition method and device based on gesture track eigenvalue
CN103218601A (en) * 2013-04-03 2013-07-24 华为技术有限公司 Method and device for detecting gesture
WO2014121566A1 (en) * 2013-02-07 2014-08-14 Han Zheng Data collection method and device for action recognition, and action recognition system
WO2015184760A1 (en) * 2014-06-03 2015-12-10 深圳Tcl新技术有限公司 Air gesture input method and apparatus
CN105373215A (en) * 2014-08-25 2016-03-02 中国人民解放军理工大学 Gesture coding and decoding based dynamic wireless gesture identification method
CN105809144A (en) * 2016-03-24 2016-07-27 重庆邮电大学 Gesture recognition system and method adopting action segmentation
WO2017000764A1 (en) * 2015-06-30 2017-01-05 芋头科技(杭州)有限公司 Gesture detection and recognition method and system
CN106503650A (en) * 2016-10-21 2017-03-15 上海未来伙伴机器人有限公司 A kind of recognition methodss of images of gestures and system
CN106886751A (en) * 2017-01-09 2017-06-23 深圳数字电视国家工程实验室股份有限公司 A kind of gesture identification method and system
WO2018018600A1 (en) * 2016-07-29 2018-02-01 深圳市赛亿科技开发有限公司 Gesture recognition device
WO2018145316A1 (en) * 2017-02-13 2018-08-16 深圳市华第时代科技有限公司 Mouse gesture recognition method and apparatus
CN109460151A (en) * 2018-11-13 2019-03-12 江西师范大学 Hand shape input method
CN109558000A (en) * 2017-09-26 2019-04-02 京东方科技集团股份有限公司 A kind of man-machine interaction method and electronic equipment
CN109635776A (en) * 2018-12-23 2019-04-16 广东腾晟信息科技有限公司 Pass through the method for procedure identification human action
CN109919057A (en) * 2019-02-26 2019-06-21 北京理工大学 A kind of multi-modal fusion gesture identification method based on efficient convolutional neural networks
CN110286777A (en) * 2019-06-26 2019-09-27 西南民族大学 A kind of full-automatic keyboard operation device of electromechanical based on machine vision
CN111291749A (en) * 2020-01-20 2020-06-16 深圳市优必选科技股份有限公司 Gesture recognition method and device and robot
US11402923B2 (en) 2020-06-30 2022-08-02 Boe Technology Group Co., Ltd. Input method, apparatus based on visual recognition, and electronic device
CN115582637A (en) * 2022-11-22 2023-01-10 长春森酉科技有限公司 Automatic detection system for laser cutting missing process

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101807198A (en) * 2010-01-08 2010-08-18 中国科学院软件研究所 Video abstraction generating method based on sketch

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101807198A (en) * 2010-01-08 2010-08-18 中国科学院软件研究所 Video abstraction generating method based on sketch

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
刘寅: "复杂背景下基于傅立叶描述子的手势识别", 《计算机仿真》, vol. 22, no. 12, 31 December 2005 (2005-12-31), pages 158 - 161 *
卢晓军: "维修仿真中虚拟人动作建模与行为控制技术研究", 《中国博士学位论文全文数据库 信息科技辑》, 15 November 2007 (2007-11-15) *
王茂吉: "基于视觉的静态手势识别系统", 《中国优秀硕士学位论文全文数据库 信息科技辑》, 15 April 2007 (2007-04-15) *
申永军: "基于聚类的间隔帧分割算法", 《微计算机信息》, vol. 26, no. 3, 31 March 2010 (2010-03-31) *

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102968642A (en) * 2012-11-07 2013-03-13 百度在线网络技术(北京)有限公司 Trainable gesture recognition method and device based on gesture track eigenvalue
CN102968642B (en) * 2012-11-07 2018-06-08 百度在线网络技术(北京)有限公司 A kind of trainable gesture identification method and device based on gesture path characteristic value
WO2014121566A1 (en) * 2013-02-07 2014-08-14 Han Zheng Data collection method and device for action recognition, and action recognition system
CN103218601A (en) * 2013-04-03 2013-07-24 华为技术有限公司 Method and device for detecting gesture
CN103218601B (en) * 2013-04-03 2016-08-03 华为技术有限公司 The method and device of detection gesture
WO2015184760A1 (en) * 2014-06-03 2015-12-10 深圳Tcl新技术有限公司 Air gesture input method and apparatus
CN105320248A (en) * 2014-06-03 2016-02-10 深圳Tcl新技术有限公司 Mid-air gesture input method and device
CN105320248B (en) * 2014-06-03 2018-12-07 深圳Tcl新技术有限公司 Aerial gesture input method and device
CN105373215B (en) * 2014-08-25 2018-01-30 中国人民解放军理工大学 Dynamic radio gesture identification method with decoding is encoded based on gesture
CN105373215A (en) * 2014-08-25 2016-03-02 中国人民解放军理工大学 Gesture coding and decoding based dynamic wireless gesture identification method
WO2017000764A1 (en) * 2015-06-30 2017-01-05 芋头科技(杭州)有限公司 Gesture detection and recognition method and system
CN106325485A (en) * 2015-06-30 2017-01-11 芋头科技(杭州)有限公司 Gesture detection and identification method and system
CN106325485B (en) * 2015-06-30 2019-09-10 芋头科技(杭州)有限公司 A kind of gestures detection recognition methods and system
US10318800B2 (en) 2015-06-30 2019-06-11 Yutou Technology (Hangzhou) Co., Ltd. Gesture detection and recognition method and system
CN105809144B (en) * 2016-03-24 2019-03-08 重庆邮电大学 A kind of gesture recognition system and method using movement cutting
CN105809144A (en) * 2016-03-24 2016-07-27 重庆邮电大学 Gesture recognition system and method adopting action segmentation
WO2018018600A1 (en) * 2016-07-29 2018-02-01 深圳市赛亿科技开发有限公司 Gesture recognition device
CN106503650A (en) * 2016-10-21 2017-03-15 上海未来伙伴机器人有限公司 A kind of recognition methodss of images of gestures and system
CN106503650B (en) * 2016-10-21 2019-09-24 上海未来伙伴机器人有限公司 A kind of recognition methods and system of images of gestures
CN106886751A (en) * 2017-01-09 2017-06-23 深圳数字电视国家工程实验室股份有限公司 A kind of gesture identification method and system
WO2018145316A1 (en) * 2017-02-13 2018-08-16 深圳市华第时代科技有限公司 Mouse gesture recognition method and apparatus
CN109558000A (en) * 2017-09-26 2019-04-02 京东方科技集团股份有限公司 A kind of man-machine interaction method and electronic equipment
WO2019062682A1 (en) * 2017-09-26 2019-04-04 京东方科技集团股份有限公司 Gesture recognition method and electronic device
US10866649B2 (en) 2017-09-26 2020-12-15 Boe Technology Group Co., Ltd. Gesture identification method and electronic device
CN109460151B (en) * 2018-11-13 2020-04-10 江西师范大学 Hand shape input method
CN109460151A (en) * 2018-11-13 2019-03-12 江西师范大学 Hand shape input method
CN109635776A (en) * 2018-12-23 2019-04-16 广东腾晟信息科技有限公司 Pass through the method for procedure identification human action
CN109919057A (en) * 2019-02-26 2019-06-21 北京理工大学 A kind of multi-modal fusion gesture identification method based on efficient convolutional neural networks
CN110286777A (en) * 2019-06-26 2019-09-27 西南民族大学 A kind of full-automatic keyboard operation device of electromechanical based on machine vision
CN111291749A (en) * 2020-01-20 2020-06-16 深圳市优必选科技股份有限公司 Gesture recognition method and device and robot
CN111291749B (en) * 2020-01-20 2024-04-23 深圳市优必选科技股份有限公司 Gesture recognition method and device and robot
US11402923B2 (en) 2020-06-30 2022-08-02 Boe Technology Group Co., Ltd. Input method, apparatus based on visual recognition, and electronic device
CN115582637A (en) * 2022-11-22 2023-01-10 长春森酉科技有限公司 Automatic detection system for laser cutting missing process

Similar Documents

Publication Publication Date Title
CN102508547A (en) Computer-vision-based gesture input method construction method and system
Adouani et al. Comparison of Haar-like, HOG and LBP approaches for face detection in video sequences
Rekha et al. Shape, texture and local movement hand gesture features for indian sign language recognition
CN105975934B (en) Dynamic gesture recognition method and system for augmented reality auxiliary maintenance
Premaratne et al. Hand gesture tracking and recognition system using Lucas–Kanade algorithms for control of consumer electronics
Bilal et al. A hybrid method using haar-like and skin-color algorithm for hand posture detection, recognition and tracking
Mohandes et al. Prototype Arabic Sign language recognition using multi-sensor data fusion of two leap motion controllers
Jambhale et al. Gesture recognition using DTW & piecewise DTW
CN109558855B (en) A kind of space gesture recognition methods combined based on palm contour feature with stencil matching method
Wu et al. Depth-based hand gesture recognition
Thongtawee et al. A novel feature extraction for American sign language recognition using webcam
Dewangan Importance & applications of digital image processing
Vishwakarma et al. Simple and intelligent system to recognize the expression of speech-disabled person
Aly et al. Arabic sign language recognition using spatio-temporal local binary patterns and support vector machine
Sahu et al. Application of feature extraction technique: A review
Azad et al. Real-time human face detection in noisy images based on skin color fusion model and eye detection
Vafadar et al. Human hand gesture recognition using motion orientation histogram for interaction of handicapped persons with computer
Le et al. Smart Elevator Cotrol System Based on Human Hand Gesture Recognition
Patel et al. Vision Based Real-time Recognition of Hand Gestures for Indian Sign Language using Histogram of Oriented Gradients Features.
Malik Handwritten Marathi compound character segmentation using minutiae detection algorithm
Kalangi et al. Deployment of Haar Cascade algorithm to detect real-time faces
Shitole et al. Dynamic hand gesture recognition using PCA, Pruning and ANN
Liu et al. Oracle-bone-inscription image segmentation based on simple fully convolutional networks
Assaleh et al. Recognition of handwritten Arabic alphabet via hand motion tracking
Munir et al. Hand Gesture Recognition: A Review

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C12 Rejection of a patent application after its publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20120620