CN103136986B - Sign Language Recognition Method and system - Google Patents

Sign Language Recognition Method and system Download PDF

Info

Publication number
CN103136986B
CN103136986B CN201210031595.8A CN201210031595A CN103136986B CN 103136986 B CN103136986 B CN 103136986B CN 201210031595 A CN201210031595 A CN 201210031595A CN 103136986 B CN103136986 B CN 103136986B
Authority
CN
China
Prior art keywords
steering order
attitude
natural language
image
sequence
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201210031595.8A
Other languages
Chinese (zh)
Other versions
CN103136986A (en
Inventor
雷敏娟
周雷
师丹玮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Taishan Sports Technology Co.,Ltd.
Original Assignee
SHENZHEN TOL TECHNOLOGY Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by SHENZHEN TOL TECHNOLOGY Co Ltd filed Critical SHENZHEN TOL TECHNOLOGY Co Ltd
Priority to CN201210031595.8A priority Critical patent/CN103136986B/en
Publication of CN103136986A publication Critical patent/CN103136986A/en
Application granted granted Critical
Publication of CN103136986B publication Critical patent/CN103136986B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)
  • Position Input By Displaying (AREA)

Abstract

The present invention relates to a kind of sign Language Recognition Method, comprise the following steps: gather the image comprising marked region; The attitude in identification marking region; Generate the steering order that described attitude is corresponding; Convert described steering order to natural language information.In addition, a kind of sign Language Recognition is additionally provided.Above-mentioned sign Language Recognition Method and system can improve the accuracy rate of identification.<pb pnum="1" />

Description

Sign Language Recognition Method and system
[technical field]
The present invention relates to artificial intelligence field, particularly a kind of sign Language Recognition Method and system.
[background technology]
In daily life, deaf-mute, owing to losing the listening and speaking ability of normal person, with other people communication, adopts sign language to exchange usually.The action that deaf-mute makes specification by limbs represents corresponding semanteme.But because most of normal person does not understand the implication of sign language, therefore deaf-mute can only be exchanged with other deaf-mutes by sign language usually, and still there is obstacle with exchanging of normal person.
In conventional art, exchange with normal person to enable deaf-mute, usually deaf-mute is made to put on the data glove with multiple sensor, and movement track and the orientation of deaf-mute's limbs is gathered by data glove, and generate the text message with semanteme according to the movement track got and orientation, thus sign language is converted to the more intelligible natural language information of normal person.
But in above-mentioned sign Language Recognition Method, data glove small volume, causes the negligible amounts of the sensor in data glove.Therefore, data glove is when gathering movement track and the orientation of limbs, and the data obtained are not accurate enough, make the error rate of the identification error of sign Language Recognition Method in conventional art larger.
[summary of the invention]
Based on this, be necessary to provide a kind of sign Language Recognition Method that can improve accuracy rate.
A kind of sign Language Recognition Method, comprises the following steps:
Gather the image comprising marked region;
The attitude in identification marking region;
Generate the steering order that described attitude is corresponding;
Convert described steering order to natural language information.
Preferably, the described step described steering order being converted to natural language information comprises:
Described steering order is arranged in steering order sequence, generates natural language information according to described steering order sequence.
Preferably, the described step according to described steering order sequence generation natural language information comprises:
Calculate the eigenwert of described steering order sequence, the mapping relations according to described default eigenwert and natural language information generate natural language information.
Preferably, also comprise before the step of the eigenwert of the described steering order sequence of described calculating:
The steering order repeated in described steering order sequence is removed.
Preferably, also comprise before the step of the eigenwert of the described steering order sequence of described calculating:
The steering order being greater than threshold value in described steering order sequence with adjacent steering order difference is removed.
Preferably, the described step described steering order being arranged in steering order sequence is specially:
Described steering order is arranged in queue according to the order generated, when detecting that there is the identical steering order of continuous predetermined number at queue end, arranges complete, and generating steering order sequence according to described queue.
Preferably, described described steering order is converted to the step of natural language information after also comprise:
Described natural language information is shown by the mode of text and/or audio frequency.
In addition, there is a need to provide a kind of sign Language Recognition that can improve accuracy rate.
A kind of sign Language Recognition, comprises with lower module:.
Image capture module, for gathering the image comprising marked region;
Gesture recognition module, for the attitude in identification marking region;
Directive generation module, for generating steering order corresponding to described attitude;
Instruction transformation module, for converting described steering order to natural language information.
Preferably, described instruction transformation module, also for described steering order is arranged in steering order sequence, generates natural language information according to described steering order sequence.
Preferably, described instruction transformation module is also for calculating the eigenwert of described steering order sequence, and the mapping relations according to described default eigenwert and natural language information generate natural language information.
Preferably, the steering order of described instruction transformation module also for repeating in described steering order sequence is removed.
Preferably, described instruction transformation module is also for removing the steering order being greater than threshold value in described steering order sequence with adjacent steering order difference.
Preferably, described instruction transformation module also for by described steering order according to the order arrangement queue generated, when detecting that there is the identical steering order of continuous predetermined number at queue end, arranging complete, and generating steering order sequence according to described queue.
Preferably, also comprise information display module, for showing described natural language information by the mode of text and/or audio frequency.
Above-mentioned sign Language Recognition Method and system, go out according to the image recognition comprising marked region collected the attitude that marked region produces, and generate steering order corresponding to attitude, then this steering order is converted to the natural language information of normal person's easy understand.Because the image by obtaining limb action judges movement locus and the attitude of limb action, therefore, all there is record to the process of whole limb action, thus avoid the situation missing certain action or attitude in default of sensor, thus improve accuracy rate when identifying sign language.
[accompanying drawing explanation]
Fig. 1 is the schematic flow sheet controlling to browse the method for webpage in the present invention;
Fig. 2 is the schematic flow sheet of step S20 in an embodiment;
Fig. 3 is the structural representation of interactive device in an embodiment;
Fig. 4 is the schematic diagram building coordinate system in an embodiment;
Fig. 5 is the structural representation of the interactive device in another embodiment;
Fig. 6 is the structural representation of the interactive device in another embodiment;
Fig. 7 is the schematic flow sheet of step S20 in another embodiment;
Fig. 8 is the schematic diagram building coordinate system in another embodiment;
Fig. 9 is the schematic flow sheet of step S30 in an embodiment;
Figure 10 is the schematic flow sheet of step S30 in another embodiment;
Figure 11 is the structural representation of sign Language Recognition system in an embodiment;
Figure 12 is the structural representation of gesture recognition module in an embodiment;
Figure 13 is the structural representation of gesture recognition module in another embodiment;
Figure 14 is the structural representation of directive generation module in an embodiment;
Figure 15 is the structural representation of directive generation module in another embodiment;
Figure 16 is the structural representation of sign Language Recognition in another embodiment.
[embodiment]
Below in conjunction with specific embodiment and accompanying drawing, technical scheme is described in detail.
In one embodiment, as shown in Figure 1, a kind of sign Language Recognition Method, comprises the following steps:
Step S10, gathers the image comprising marked region.
In the present embodiment, marked region is a region in the image gathered, and this region can be formed by interactive device.
Concrete, in one embodiment, interactive device can be hand-held device, part or all of hand-held device can be set as the color of specifying or shape, gather the image of hand-held device, the part of this designated color in the hand-held device in image or shape forms marked region.In addition, interactive device can also be the hand-held device of tape label, namely on hand-held device, attach the mark (as reflectorized material) of designated color or shape, gather the image of hand-held device, on the hand-held device in image, the mark of incidental designated color or shape forms marked region.
In another embodiment, interactive device can also be human body (such as face, palm, arm etc.), gathers the image of human body, and the human body in image forms marked region.In addition, interactive device can also be the human body of tape label, namely on human body, attach the mark (as reflectorized material) of designated color or shape, when gathering the image of human body, the mark of this designated color in image or shape forms marked region.
Step S20, the attitude in identification marking region.
Concrete, the image collected is processed, extracts the marked region in image, then produce the attitude of marked region according to the pixel coordinate of the pixel in marked region in the image coordinate system built.So-called attitude, refers to the posture state that marked region is formed in the picture.Further, in two dimensional image, attitude is marked region in two dimensional image and the angle between predeterminated position, i.e. attitude angle; In 3-D view, the vector that attitude forms for the multiple attitude angle between the marked region in two dimensional image and predeterminated position, i.e. attitude vectors." marked region produce attitude " said in the present invention, " attitude of marked region ", " attitude " all refer to described attitude, namely the attitude angle of different embodiment and attitude vectors.
Step S30, generates the steering order that attitude is corresponding.
In the present embodiment, preset the mapping relations between the attitude of marked region and steering order, and these mapping relations are stored in a database.After identifying the attitude of marked region, the steering order corresponding with attitude can be searched according to the attitude identified from database.
Step S40, converts steering order to natural language information.
The language message of natural language information and normal person's easy understand, as Chinese, English, Latin etc.The mapping table of steering order and natural language information can be pre-set, and store in a database, then by inquiring about steering order in a database to obtain the natural language information corresponding with it.
Such as, the steering order preset and the mapping table of natural language information can be as shown in table 1:
Table 1
Steering order Natural language information
command_cir Circle
command_heart Like
command_ok Alright
...... ......
Using the finger of human body as interactive device, when the finger of user defines a circle, then corresponding generating represents circular steering order command_cir, and then inquiry obtains corresponding natural language information for " circle " in a database.When the finger of user surround one heart-shaped time, then correspondingly generate steering order command_heart, then inquiry obtaining corresponding natural language information for " love " in a database.Be connected between the forefinger and thumb of user, when OK shape is made in the expansion of other three fingers, then corresponding generation steering order command_ok, then inquiry can obtain corresponding natural language information for " good " in a database.
In one embodiment, steering order can be arranged in steering order sequence, generate natural language information according to steering order sequence.
First by generating multiple steering order every the sampling interval T execution step S10, the step S20 that preset, step S30.Again the steering order of generation is arranged in steering order sequence according to the order generated, and generates natural language information according to steering order sequence.
Such as, when the finger of employing user and arm are as interactive device, caught the finger of user and the attitude of arm every 0.1 second, and generate steering order for identifying the attitude in this moment.After generating multiple steering order within a certain period of time, steering order then illustrates the finger of user and the running orbit of arm according to the sequence of the order composition generated.
Further, can calculate the eigenwert of steering order sequence, the mapping relations according to the eigenwert preset and natural language information generate natural language information.
The attitude part for representing attitude information and the coordinate part for representing this attitude coordinate information is in the picture comprised in steering order.When using finger and arm as interactive device, the attitude part of steering order then can represent the shape that the attitude of finger and arm is formed or abstract vector graphics, as the ring-type that the five fingers expansion shape, finger surround.The coordinate part of steering order then illustrates the position of shape on image that finger or arm surround.When image is two dimensional image, coordinate is two-dimensional coordinate.When image is 3-D view, coordinate is three-dimensional coordinate.
Eigenwert is for representing the feature that the steering order sequence with certain similarity is common.Posture feature value can be adopted to represent the variation characteristic of the attitude part in steering order sequence, adopt coordinate characteristic value to represent the variation characteristic of the coordinate part in steering order sequence.Then by the eigenwert of posture feature value and coordinate characteristic value composition control instruction sequence.
Eigenwert and natural language information have default mapping relations.Such as, can in advance eigenwert 2 (1F@1L) corresponding natural language " be thanks ".This eigenwert 2 (1F@1L) is based on the change of the attitude part of steering order sequence, represent that stretching (being represented by the 1F in eigenwert) by thumb becomes thumb and bend (being represented by the 1L in eigenwert), and bent twice (being multiplied by 2 with bracket to represent).In advance corresponding for eigenwert 5F_y-down natural language " can be pressed ".This eigenwert 5F_y-down, based on the change of the coordinate part in steering order sequence, represents that palm (being represented by the 5F in eigenwert) presses (being represented by the y-down in eigenwert) from top to bottom.Further, first the steering order repeated in steering order sequence is removed, and then calculate the eigenwert of steering order sequence.Remove the steering order repeated and can reduce calculated amount.
Further, first the steering order being greater than threshold value in steering order sequence with adjacent steering order difference is removed, and then calculate the eigenwert of steering order sequence.When the sampling interval preset is less, if when the attitude part of certain steering order in steering order sequence or coordinate part and adjacent steering order difference are greater than threshold value, then judge the steering order of this steering order as mistake.False command will be filtered out steering order sequence, and can not be used to the eigenwert calculating steering order sequence.
Such as, when user adopts finger as input media, if the steering order sequence got is: [3F (0,0), 3F (1,0), 3F (2,0), 2F (3,0), 3F (4,0), 3F (5,0)].Wherein attitude part 3F represents 3 attitudes pointed, and 2F represents 2 attitudes pointed, and coordinate part (0,0)-(5,0) represents that finger gesture is in the coordinate in the image of catching.This steering order sequence is for representing the track that three finger horizontal translations of user are formed, and each steering order is the sampled point of this track.Wherein, may be because user is in moving process, there is the situation pointed and close up, steering order 2F (3,0) larger in the upper gap of the number pointed (the attitude part of steering order) with adjacent 3F (2,0), 3F (4,0), therefore be judged as the steering order of mistake, and be moved out of steering order sequence.When certain steering order and adjacent steering order have bigger difference, normally because capture the nonstandard sign language gesture of user, or be blocked when mobile due to interactive device, can not its shape obtained completely and cause.Such as, in deaf and dumb sign language, some gesture needs two hands alternately mobile, in movement, the situation of blocking may be there is, therefore remove the steering order being greater than threshold value (shape different or coordinate distance be greater than threshold value) from adjacent steering order gap, what eigenwert can be made to calculate is more accurate.
When calculating the eigenwert of steering order sequence, first calculate the steering order in steering order sequence posture feature value, then according to posture feature value by steering order sequence segment, the steering order in the steering order subsequence after segmentation has identical posture feature value.The monotonicity of the coordinate then described according to the coordinate part of the steering order in steering order subsequence, in image, the variance of the distance of a certain reference coordinate calculates the coordinate characteristic value of steering order subsequence.Then the posture feature value of all steering order subsequences and coordinate characteristic value are integrated into the eigenwert of steering order sequence.
Such as, to point as interactive device, user makes following action:
Stretch out one and point translation from left to right, stretch the five fingers after moving on to certain position and do quadrant motion, finally fix, and bending forefinger and thumb, forefinger and thumb tip are collided and forms OK shape gesture.
The steering order sequence got is: [1F (0,0), 1F (1,0), 1F (2,0), 1F (3,0), 5F (2.5,1.5), 5F (1.5,2.5), 5F (0,3), OK (0,3), OK (0,3), OK (0,3)].Wherein 1F is the posture feature value of 1 finger, and 5F is the posture feature value of 5 fingers, and OK represents the posture feature value of OK shape gesture, and bracket inner digital represents attitude coordinate figure in the picture.Remove the steering order of repetition and after being greater than the steering order of threshold value with adjacent steering order difference, then according to the difference of posture feature value, steering order sequence be divided into three sub-steering order sequences:
Subsequence 1:[1F (0,0), 1F (1,0), 1F (2,0), 1F (3,0)].
Subsequence 2:[5F (2.5,1.5), 5F (1.5,2.5), 5F (0,3)].
Subsequence 3:[OK (0,3)].
The coordinate characteristic value of subsequence 1 is that x-right, x-right represent that the monotonicity of the coordinate of subsequence 1 is that horizontal ordinate increases progressively (x-axis), ordinate constant (y-axis).The coordinate characteristic value of subsequence 2 is that q-cir, q-cir represent that the track of the coordinate formation of subsequence 2 is less than threshold value to the variance of reference point coordinate (0,0), and namely track is 1/4th circles to the center of circle (0,0).The coordinate characteristic value of subsequence 3 is hold, namely represents the last attitude that fixes.
Then the posture feature value of subsequence 1, subsequence 2 and subsequence 3 and coordinate characteristic value are integrated, the eigenwert obtaining steering order sequence is: 1F_x-right@5F_q-cir@OK_hold.Wherein _ represent separator, for distinguishing posture feature value and coordinate characteristic value with@.
After obtaining the eigenwert of steering order sequence, then obtain the natural language corresponding with the eigenwert of steering order sequence according to the eigenwert preset with the mapping relations of natural language.As above in example, the natural language that 1F_x-right@5F_q-cir@OK_hold is corresponding is " perfection ", what then user made stretch out one points translation from left to right, stretch the five fingers after moving on to certain position and do quadrant motion, finally fix, and bending forefinger and thumb, namely the sign language forming OK shape gesture of forefinger and thumb tip being collided has been converted into natural language " perfection ".
It should be noted that, sign language is herein not limited to the deaf and dumb sign language of standard, also can be user-defined sign language.The corresponding relation of sign language and natural language depends on default eigenwert and the mapping relations of natural language.
Further, when steering order being arranged in steering order sequence, steering order can be arranged in queue according to the order generated, when detecting that there is the identical steering order of continuous predetermined number at queue end, arrange complete, and generate steering order sequence according to queue.
Pre-set buffer area, whenever this steering order is then stored in buffer area by generation steering order in order, form queue.Meanwhile, detect this buffer area, if when there is an identical steering order of continuous predetermined number at queue end, the steering order OK (0,3) as above in example, queue arrangement is complete, then takes out this queue formation control instruction sequence from buffer area.
Intercepting steering order sequence by the identical steering order of continuous predetermined number, making when distinguishing multiple sign language gesture (pause between similar English word) convenient, thus wrong identification when avoiding many gestures.
In one embodiment, after converting steering order to natural language information, natural language information is shown by the mode of text and/or audio frequency.
By display screen, natural language information is shown with the form of word, also by audio frequency apparatus, natural language information can be played back.
Such as, after steering order is changed into natural language information, natural language information is encoded, and send out this coding to remote terminal (mobile phone, computer, TV etc.), first this coding and decoding is become natural language information after remote terminal reception to this coding, then show this natural language information in a text form by display screen or play this natural language information in the form of speech by audio frequency apparatus.
Show natural language information by the mode of text and/or audio frequency, other people can be facilitated to understand natural language information.Natural language information coding is sent to remote terminal by telephone network or internet or television network, and then show this natural language information by remote terminal by decoding, make deaf-mute can realize distance communicating with the normal person being ignorant of sign language, thus facilitate the communication of deaf-mute and normal person.
As shown in Figure 2, in one embodiment, the image comprising marked region collected is two dimensional image, and the detailed process of above-mentioned steps S20 comprises:
Step S202, extracts the pixel with pre-set color Model Matching in image, carries out connected domain detection to the pixel obtained, extract the marked region in the connected domain detecting and obtain.
Concrete, comprise the image of marked region by camera acquisition, the image obtained is two-dimensional visible light image.Preferably, also can add infrared fileter before the camera lens of video camera, for elimination other wave band light except infrared band, then the image gathered is two-dimensional infrared image.Due in visible images, the object in scene can form interference to the identification of marked region, and infrared image is because having filtered out visible ray information, disturbs less, and therefore two-dimensional infrared image is more conducive to extracting marked region.
In the present embodiment, set up color model in advance.The color of such as marked region is red, then set up red model in advance, in this model, the rgb value component of pixel can between 200 to 255, and G, B component can close to zero; Obtain the pixel meeting the rgb value of this red model in the image gathered and be red pixel.In addition, when forming marked region by human body in the image gathered, then the pixel of mating with default complexion model in the image of collection can be obtained.Connected domain detection is carried out to the pixel obtained, obtains multiple connected domain, if connected domain is the set of individual continuous print pixel composition.
In the present embodiment, because the size and shape of marked region should be roughly changeless, when carrying out connected domain to the pixel obtained and detecting, girth and/or the area of all connected domains in the pixel of acquisition can be calculated.Concrete, the girth of connected domain can be the number of connected domain boundary pixel, and the area of connected domain can be the number of the whole pixels in connected domain.Further, the girth of the girth of the connected domain of acquisition and/or area and default marked region and/or area can be contrasted, obtain to meet and preset the girth of marked region and/or the connected domain of area is marked region.Preferably, also can using girth square with the ratio of area as judgment criterion, this ratio of connected domain meets this ratio of default marked region, then this connected domain is marked region.
Step S204, obtains the pixel coordinate in marked region, produces marked region attitude according to this pixel coordinate.
Concrete, in one embodiment, as shown in Figure 3, interactive device comprises portion of the handle and is attached to the mark of portion of the handle, and wherein, mark can be the reflectorized material of elongate in shape, preferably, can be oval or rectangular shape.In other embodiments, interactive device also can be human body, and as face, palm, arm etc., then the marked region in the image collected is the region of human body.
In the present embodiment, marked region is a continuum, the process then producing the attitude of marked region according to pixel coordinate is: the covariance matrix calculating pixel coordinate, obtain covariance matrix eigenvalue of maximum characteristic of correspondence vector, produce the attitude of marked region according to proper vector, the attitude of this marked region is an attitude angle.
Concrete, as shown in Figure 4, build two dimensional image coordinate system, for the some A (u1 of two on this coordinate system, and B (u2 v1), v2), its attitude angle formed is then the arc tangent of slope, i.e. arctan ((v2-v1)/(u2-u1)).Concrete, in the present embodiment, calculate the covariance matrix of the pixel coordinate in the marked region extracted, obtain covariance matrix eigenvalue of maximum characteristic of correspondence vector, the direction of this proper vector is the direction of marked region major axis place straight line.As shown in Figure 4, marked region major axis place rectilinear direction is the direction of A, B 2 place straight lines, if proper vector is [dir_u, dir_v] t, wherein, the projection of direction on u axle of dir_u descriptive markup region major axis, its absolute value is proportional to the projection of vector on u change in coordinate axis direction (i.e. u2-u1) pointing to B from A; The projection of direction on v axle of dir_v descriptive markup region major axis, its absolute value is proportional to the projection of vector on v change in coordinate axis direction (i.e. v2-v1) pointing to B from A.If dir_u or dir_v is less than 0, be then modified to [-dir_u ,-dir_v] t, then the attitude angle of marked region is: arctan (dir_v/dir_u).
In another embodiment, marked region comprises the first continuum and the second continuum, the detailed process then producing the attitude of marked region according to described pixel coordinate is: calculate the center of gravity of the first continuum and the center of gravity of the second continuum according to pixel coordinate, produces the attitude of marked region according to the pixel coordinate of the pixel coordinate of the center of gravity of the first continuum and the center of gravity of the second continuum.Concrete, in one embodiment, interactive device comprises portion of the handle and is attached to two marks of portion of the handle.As shown in Figure 5, be labeled as two, be attached to portion of the handle front end respectively, the shape of mark can be oval or rectangle.Preferably, mark can for being positioned at two round dots of handgrip part front end.As shown in Figure 6, mark can be arranged on the two ends of portion of the handle.In other embodiments, also can mark be arranged on human body, such as, be arranged on face, palm or arm.It should be noted that two set marks can be inconsistent in the feature such as size, shape, color.
In the present embodiment, the marked region of extraction comprises two continuums, is respectively the first continuum and the second continuum.Further, the center of gravity of these two continuums is calculated according to pixel coordinate.Concrete, calculate the mean value of the whole pixel coordinates in continuum, the pixel coordinate obtained is the center of gravity of continuum.As shown in Figure 4, the center of gravity of two continuums calculated is respectively A (u1, v1) and B (u2, v2), then the attitude angle of marked region is the arc tangent of slope, i.e. arctan ((v2-v1)/(u2-u1)).
In another embodiment, the image gathered can be 3-D view.Concrete, traditional stereo visual system (video camera known by two locus and Correlation method for data processing equipment form), structured-light system (right video camera, a light source and a Correlation method for data processing equipment forms) or TOF (time of flight, flight time) depth camera collection 3-D view (i.e. three dimensional depth image) can be utilized.
In the present embodiment, as shown in Figure 7, the detailed process of step S20 comprises:
Step S210, to Image Segmentation Using, extracts the connected domain in this image, calculates the property value of connected domain, the property value of connected domain and the marked region property value preset is contrasted, and this marked region is the connected domain meeting the marked region property value that this is preset.
Concrete, when two adjacent pixel depth differences are less than the threshold value preset in three dimensional depth image, such as 5 centimetres, then think that two pixels are communicated with, connected domain detection is carried out to whole image, a series of connected domains comprising mark connected domain can be obtained.
In the present embodiment, the property value of connected domain comprises the size and dimension of connected domain.Concrete, calculate the size/shape of connected domain, contrast with the size/shape of the mark on interactive device, the connected domain obtaining the size/shape meeting mark is the connected domain of marked region (marked region).For rectangle marked, being marked in the image of collection namely on interactive device is rectangle, the length of pre-set mark and width, calculate length and the width of physical region corresponding to connected domain, length and the width of this length and width and mark are more close, then connected domain is more similar to marked region.
Further, the process calculating the length of physical region corresponding to connected domain and width is as follows: the covariance matrix calculating the three-dimensional coordinate of connected domain pixel, the length of the physical region adopting following formulae discovery connected domain corresponding and width: wherein, k is the coefficient preset, such as, be set to 4, and when λ is covariance matrix eigenvalue of maximum, then l is the length of connected domain, and when λ is the second largest eigenwert of covariance matrix, then l is the width of connected domain.
Further, also can preset the length breadth ratio of rectangle marked, such as length breadth ratio is 2, then the length breadth ratio of the physical region that connected domain is corresponding is more close to the length breadth ratio of the rectangle marked of default settings, then connected domain is more similar to marked region, concrete, the length breadth ratio of the physical region adopting following formulae discovery connected domain corresponding: wherein, r is the length breadth ratio of connected domain, λ 0for the eigenvalue of maximum of covariance matrix, λ 1for the Second Largest Eigenvalue of covariance matrix.
Step S220, obtains the pixel coordinate in marked region, produces the attitude of marked region according to this pixel coordinate.
Concrete, in the present embodiment, the attitude of marked region is attitude vectors.As shown in Figure 8, build 3-D view coordinate system, this coordinate is right-handed coordinate system.In the coordinate system, if space vector OP, P are projected as p at plane X OY, be then [α, θ] with the attitude vectors of polar coordinate representation vector OP t, α is angle XOp, and namely X-axis is to Op angle, and span is 0 to 360 degree, and θ is angle pOP, i.e. the angle of OP and XOY plane, and span is that-90 degree are to 90 degree.If 2 on the space ray in this coordinate system is A (x1, y1, z1) and B (x2, y2, z2), then this attitude vectors of 2 [α, θ] tavailable following formula is uniquely determined:
cos ( &alpha; ) = x 2 - x 1 ( x 2 - x 1 ) 2 + ( y 2 - y 1 ) 2
sin ( &alpha; ) = y 2 - y 1 ( x 2 - x 1 ) 2 + ( y 2 - y 1 ) 2 - - - ( 1 )
&theta; = arctan ( z 2 - z 1 ( x 2 - x 1 ) 2 + ( y 2 - y 1 ) 2 ) - - - ( 2 )
In the present embodiment, after extracting marked region, calculate the covariance matrix of the pixel coordinate in marked region, obtain covariance matrix eigenvalue of maximum characteristic of correspondence vector, and this proper vector is converted to attitude vectors.Concrete, if the attitude vectors obtained is [dir x, dir y, dir z] t, wherein, dir xrepresent 2 distances in the direction of the x axis, dir yrepresent 2 distances in the y-axis direction, dir zrepresent 2 distances in the z-axis direction.Can think that the ray that this attitude vectors describes has two points, i.e. (0,0,0) and (dir x, dir y, dir z), namely ray triggers from initial point, points to (dir x, dir y, dir z), then attitude angle need meet above-mentioned formula (1) and (2), makes the x1=0 in above-mentioned formula (1) and (2), y1=0, z1=0, x2=dir x, y2=dir y, z2=dir z, attitude vectors [α, θ] can be obtained t.
In one embodiment, marked region is a continuum, the process then producing the attitude of marked region according to pixel coordinate is: the covariance matrix calculating pixel coordinate, obtain covariance matrix eigenvalue of maximum characteristic of correspondence vector, produce the attitude of marked region according to proper vector.As mentioned above, the attitude of this marked region is an attitude vectors.
In another embodiment, marked region comprises the first continuum and the second continuum, the detailed process then producing the attitude of marked region according to described pixel coordinate is: calculate the center of gravity of the first continuum and the center of gravity of the second continuum according to pixel coordinate, calculates the attitude of marked region according to the pixel coordinate of the pixel coordinate of the center of gravity of the first continuum and the center of gravity of the second continuum.As shown in Figure 8, in the present embodiment, the pixel coordinate in marked region is three-dimensional coordinate, concrete, and can produce the attitude of marked region according to the pixel coordinate of the center of gravity of two continuums calculated, this attitude is an attitude vectors.
In one embodiment, also can comprise before the step of the attitude in identification marking region: judge that the image gathered is the step of two dimensional image or 3-D view.Concrete, if the image gathered is two dimensional image, then performs above-mentioned steps S202 to step S204, if the image gathered is 3-D view, then perform above-mentioned steps S210 to S220.
As shown in Figure 9, in one embodiment, the detailed process of above-mentioned steps S30 comprises:
Step S302, obtains the attitude of this marked region in current frame image.
As mentioned above, the attitude obtained in step S302 can be the attitude (i.e. attitude angle) of the marked region in the two dimensional image of present frame, also can be the attitude (i.e. attitude vectors) of the marked region in the dark image of three-dimensional of present frame.In the present embodiment, the mapping relations between attitude and steering order are preset.This attitude also can be described as absolute pose.
Step S304, the attitude according to presetting generates the steering order corresponding with this attitude with the mapping relations between steering order.
Such as, steering order is left mouse button instruction and right button instruction.For two dimensional image, the span of attitude angle is that-180 degree are to 180 degree.Can preset the attitude angle in current frame image in the scope of (a, b), then trigger left button instruction, the attitude angle in current frame image in the scope of (c, d), then triggers right button instruction.Wherein, a, b, c, d angle all for presetting, meets a < b, c < d, and the common factor of set [a, b] and set [c, d] is empty.
In addition, in 3-D view, the attitude identified comprises two attitude angle, and one of them attitude angle can be used to obtain steering order, and two attitude angle also can be used to obtain steering order.Use the Method And Principle of one of them attitude angle and two dimensional image similar, then repeat no more at this.When using two attitude angle, if can arrange two attitude angle all within the scope of the instruction triggers preset time, just trigging control instruction.
As shown in Figure 10, in another embodiment, the image comprising marked region of collection is image sequence, and the detailed process of above-mentioned steps S30 comprises:
Step S310, the relative attitude between the attitude obtaining this marked region in the attitude of this marked region in current frame image and previous frame image.
In the present embodiment, can the image sequence that is made up of multiple image comprising marked region of Real-time Collection.As mentioned above, the attitude obtained in step S310 can be the attitude angle of the marked region in current frame image and previous frame image, also can be the attitude vectors of the marked region in current frame image and previous frame image.Relative attitude between attitude in attitude in current frame image and previous frame image is both differences.
Step S320, the relative attitude according to presetting generates the steering order corresponding with this relative attitude with the mapping relations between steering order.
Such as, for two dimensional image, relative attitude is relative attitude angle, the attitude angle that can preset current frame image is greater than 30 degree than the attitude angle increase of previous frame, namely when relative attitude angle is greater than 30 degree, then the instruction of the roller roll counter-clockwise of mouse is triggered, when the attitude angle of current frame image is greater than 40 degree than the attitude angle minimizing of previous frame, namely relative attitude angle is less than-40 when spending, then the instruction that the roller triggering mouse rolls clockwise.The principle of 3-D view is similar with it, then repeats no more at this.
In 3-D view, the attitude identified comprises two attitude angle, and one of them attitude angle can be used to obtain steering order, and two attitude angle also can be used to obtain steering order.Use the Method And Principle of one of them attitude angle and two dimensional image similar, then repeat no more at this.When using two attitude angle, if can arrange the change of two attitude angle when all meeting pre-conditioned, such as first attitude angle change is greater than the first threshold preset, and second attitude angle change is greater than the Second Threshold preset, then trigging control instruction.
In one embodiment, as shown in figure 11, the system of webpage is browsed in a kind of control, comprises image capture module 10, gesture recognition module 20, directive generation module 30 and instruction transformation module 40, wherein:
Image capture module 10 comprises the image of marked region for gathering.
In the present embodiment, marked region is a region in the image gathered, and this region can be formed by interactive device.Concrete, in one embodiment, interactive device can be hand-held device, part or all of hand-held device can be set as the color of specifying or shape, gather the image of hand-held device, the part of this designated color in the hand-held device in image or shape forms marked region.In addition, interactive device can also be the hand-held device of tape label, namely on hand-held device, attach the mark (as reflectorized material) of designated color or shape, gather the image of hand-held device, on the hand-held device in image, the mark of incidental designated color or shape forms marked region.
In another embodiment, interactive device can also be human body (such as face, palm, arm etc.), gathers the image of human body, and the human body in image forms marked region.In addition, interactive device can also be the human body of tape label, namely on human body, attach the mark (as reflectorized material) of designated color or shape, when gathering the image of human body, the mark of this designated color in image or shape forms marked region.
Gesture recognition module 20 is for the attitude in identification marking region.
Concrete, the image collected is processed, extracts the marked region in image, then obtain the attitude of marked region according to the pixel coordinate of the pixel in marked region in the image coordinate system built.So-called attitude, refers to the posture state that marked region is formed in the picture.Further, in two dimensional image, attitude is marked region in two dimensional image and the angle between predeterminated position, i.e. attitude angle; In 3-D view, the vector that attitude forms for the multiple attitude angle between the marked region in two dimensional image and predeterminated position, i.e. attitude vectors." marked region produce attitude " said in the present invention, " attitude of marked region " all refers to described attitude, namely the attitude angle of different embodiment and attitude vectors.
Directive generation module 30 is for generating steering order corresponding to attitude.
In the present embodiment, preset the mapping relations between the attitude of marked region and steering order, and these mapping relations are stored in database (not shown).After identifying the attitude of marked region, the attitude that directive generation module 30 can be used for identifying according to gesture recognition module 20 searches the steering order corresponding with attitude from database.
Instruction transformation module 40, for converting steering order to natural language information.
The language message of natural language information and normal person's easy understand, as Chinese, English, Latin etc.The mapping table of steering order and natural language information can be pre-set, and store in a database, then by inquiring about steering order in a database to obtain the natural language information corresponding with it.
Such as, the steering order preset and the mapping table of natural language information can be as shown in table 2:
Table 2
Steering order Natural language information
command_cir Circle
command_heart Like
command_ok Alright
...... ......
Using the finger of human body as interactive device, when the finger of user defines a circle, then corresponding generating represents circular steering order command_cir, and then inquiry obtains corresponding natural language information for " circle " in a database.When the finger of user surround one heart-shaped time, then correspondingly generate steering order command_heart, then inquiry obtaining corresponding natural language information for " love " in a database.Be connected between the forefinger and thumb of user, when OK shape is made in the expansion of other three fingers, then corresponding generation steering order command_ok, then inquiry can obtain corresponding natural language information for " good " in a database.
In one embodiment, instruction transformation module 40 also can be used for steering order to be arranged in steering order sequence, generates natural language information according to steering order sequence.
Predeterminable sampling interval T, then generates multiple steering order every sampling interval T, then the steering order of generation is arranged in steering order sequence according to the order generated, and then can generate natural language information according to steering order sequence.
Such as, the finger of user and arm can be adopted as interactive device, then caught the finger of user and the attitude of arm every 0.1 second, and generate steering order for identifying the attitude in this moment.After generating multiple steering order within a certain period of time, steering order then illustrates the finger of user and the running orbit of arm according to the sequence of the order composition generated.
Further, instruction transformation module 40 also can be used for the eigenwert calculating steering order sequence, and the mapping relations according to the eigenwert preset and natural language information generate natural language information.
Can comprise two parts information in steering order, a part is the information attitude part for representing attitude, and another part then represents the coordinate part of this attitude coordinate information in the picture.When using finger and arm as interactive device, the attitude part of steering order then can represent the shape that the attitude of finger and arm is formed or abstract vector graphics, as the ring-type that the five fingers expansion shape, finger surround.The coordinate part of steering order then illustrates the position of shape on image that finger or arm surround.When image is two dimensional image, coordinate is two-dimensional coordinate.When image is 3-D view, coordinate is three-dimensional coordinate.
Eigenwert is for representing the feature that the steering order sequence with certain similarity is common.Posture feature value can be adopted to represent the variation characteristic of the attitude part in steering order sequence, adopt coordinate characteristic value to represent the variation characteristic of the coordinate part in steering order sequence.Then by the eigenwert of posture feature value and coordinate characteristic value composition control instruction sequence.
Eigenwert and natural language information have default mapping relations.Such as, can in advance eigenwert 2 (1F@1L) corresponding natural language " be thanks ".This eigenwert 2 (1F@1L) is based on the change of the attitude part of steering order sequence, represent that stretching (being represented by the 1F in eigenwert) by thumb becomes thumb and bend (being represented by the 1L in eigenwert), and bent twice (being multiplied by 2 with bracket to represent).In advance corresponding for eigenwert 5F_y-down natural language " can be pressed ".This eigenwert 5F_y-down, based on the change of the coordinate part in steering order sequence, represents that palm (being represented by the 5F in eigenwert) presses (being represented by the y-down in eigenwert) from top to bottom.
Further, instruction transformation module 40 also for first the steering order repeated in steering order sequence being removed, and then calculates the eigenwert of steering order sequence.Remove the steering order repeated and can reduce calculated amount.
Further, instruction transformation module 40 also for first the steering order being greater than threshold value in steering order sequence with adjacent steering order difference being removed, and then calculates the eigenwert of steering order sequence.When the sampling interval preset is less, if when the attitude part of certain steering order in steering order sequence or coordinate part and adjacent steering order difference are greater than threshold value, then judge the steering order of this steering order as mistake.False command will be filtered out steering order sequence, and can not be used to the eigenwert calculating steering order sequence.
Such as, when user adopts finger as input media, if the steering order sequence got is: [3F (0,0), 3F (1,0), 3F (2,0), 2F (3,0), 3F (4,0), 3F (5,0)].Wherein attitude part 3F represents 3 attitudes pointed, and 2F represents 2 attitudes pointed, and coordinate part (0,0)-(5,0) represents that finger gesture is in the coordinate in the image of catching.This steering order sequence is for representing the track that three finger horizontal translations of user are formed, and each steering order is the sampled point of this track.Wherein, may be because user is in moving process, there is the situation pointed and close up, steering order 2F (3,0) larger in the upper gap of the number pointed (the attitude part of steering order) with adjacent 3F (2,0), 3F (4,0), therefore be judged as the steering order of mistake, and be moved out of steering order sequence.
When certain steering order and adjacent steering order have bigger difference, normally because capture the nonstandard sign language gesture of user, or be blocked when mobile due to interactive device, can not its shape obtained completely and cause.Such as, in deaf and dumb sign language, some gesture needs two hands alternately mobile, in movement, the situation of blocking may be there is, therefore remove the steering order being greater than threshold value (shape different or coordinate distance be greater than threshold value) from adjacent steering order gap, what eigenwert can be made to calculate is more accurate.
When calculating the eigenwert of steering order sequence, instruction transformation module 40 is also for calculating the posture feature value of the steering order in steering order sequence, then according to posture feature value by steering order sequence segment, the steering order in the steering order subsequence after segmentation has identical posture feature value.The monotonicity of the coordinate then described according to the coordinate part of the steering order in steering order subsequence, in image, the variance of the distance of a certain reference coordinate calculates the coordinate characteristic value of steering order subsequence.Then the posture feature value of all steering order subsequences and coordinate characteristic value are integrated into the eigenwert of steering order sequence.
Such as, to point as interactive device, user makes following action:
Stretch out one and point translation from left to right, stretch the five fingers after moving on to certain position and do quadrant motion, finally fix, and bending forefinger and thumb, forefinger and thumb tip are collided and forms OK shape gesture.
The steering order sequence got is: [1F (0,0), 1F (1,0), 1F (2,0), 1F (3,0), 5F (2.5,1.5), 5F (1.5,2.5), 5F (0,3), OK (0,3), OK (0,3), OK (0,3)].Wherein 1F is the posture feature value of 1 finger, and 5F is the posture feature value of 5 fingers, and OK represents the posture feature value of OK shape gesture, and bracket inner digital represents attitude coordinate figure in the picture.Remove the steering order of repetition and after being greater than the steering order of threshold value with adjacent steering order difference, according to the difference of posture feature value, steering order sequence can be divided into three sub-steering order sequences:
Subsequence 1:[1F (0,0), 1F (1,0), 1F (2,0), 1F (3,0)].
Subsequence 2:[5F (2.5,1.5), 5F (1.5,2.5), 5F (0,3)].
Subsequence 3:[OK (0,3)].
The coordinate characteristic value of subsequence 1 is that x-right, x-right represent that the monotonicity of the coordinate of subsequence 1 is that horizontal ordinate increases progressively (x-axis), ordinate constant (y-axis).The coordinate characteristic value of subsequence 2 is that q-cir, q-cir represent that the track of the coordinate formation of subsequence 2 is less than threshold value to the variance of reference point coordinate (0,0), and namely track is 1/4th circles to the center of circle (0,0).The coordinate characteristic value of subsequence 3 is hold, namely represents the last attitude that fixes.
Then the posture feature value of subsequence 1, subsequence 2 and subsequence 3 and coordinate characteristic value are integrated by instruction transformation module 40, and the eigenwert obtaining steering order sequence is: 1F_x-right@5F_q-cir@OK_hold.Wherein _ represent separator, for distinguishing posture feature value and coordinate characteristic value with@.
After obtaining the eigenwert of steering order sequence, then obtain the natural language corresponding with the eigenwert of steering order sequence according to the eigenwert preset with the mapping relations of natural language.As above in example, the natural language that 1F_x-right@5F_q-cir@OK_hold is corresponding is " perfection ", what then user made stretch out one points translation from left to right, stretch the five fingers after moving on to certain position and do quadrant motion, finally fix, and bending forefinger and thumb, namely the sign language forming OK shape gesture of forefinger and thumb tip being collided has been converted into natural language " perfection ".
It should be noted that, sign language is herein not limited to the deaf and dumb sign language of standard, also can be user-defined sign language.The corresponding relation of sign language and natural language depends on default eigenwert and the mapping relations of natural language.
Further, when steering order being arranged in steering order sequence, instruction transformation module 40 also can be used for steering order to be arranged in queue according to the order generated, when detecting that there is the identical steering order of continuous predetermined number at queue end, arrange complete, and generate steering order sequence according to queue.
Pre-set buffer area, whenever this steering order is then stored in buffer area by generation steering order in order, form queue.Meanwhile, detect this buffer area, if when there is an identical steering order of continuous predetermined number at queue end, the steering order OK (0,3) as above in example, queue arrangement is complete, then takes out this queue formation control instruction sequence from buffer area.
Intercepting steering order sequence by the identical steering order of continuous predetermined number, making when distinguishing multiple sign language gesture (pause between similar English word) convenient, thus wrong identification when avoiding many gestures.
In one embodiment, as shown in figure 16, sign Language Recognition also comprises information display module 50, for after converting steering order to natural language information, shows natural language information by the mode of text and/or audio frequency.
Natural language information also for being shown with the form of word by natural language information by display screen, also can be played back by audio frequency apparatus by information display module 50.
Such as, after steering order is changed into natural language information, natural language information is encoded, and send out this coding to remote terminal (mobile phone, computer, TV etc.), first this coding and decoding is become natural language information after remote terminal reception to this coding, then show this natural language information in a text form by display screen or play this natural language information in the form of speech by audio frequency apparatus.
Show natural language information by the mode of text and/or audio frequency, other people can be facilitated to understand natural language information.Natural language information coding is sent to remote terminal by telephone network or internet or television network, and then show this natural language information by remote terminal by decoding, make deaf-mute can realize distance communicating with the normal person being ignorant of sign language, thus facilitate the communication of deaf-mute and normal person.
As shown in figure 12, in one embodiment, the image that image capture module 10 collects is two dimensional image, and gesture recognition module 20 comprises the first image processing module 202 and the first attitude generation module 204, wherein:
First image processing module 202, for extracting the pixel with pre-set color Model Matching in image, carries out connected domain detection to the pixel obtained, and extracts the marked region in the connected domain detecting and obtain.
Concrete, image capture module 10 can be video camera, and its image collected can be two-dimensional visible light image.Preferably, also can add infrared fileter before the camera lens of video camera, for elimination other wave band light except infrared band, then the image that image capture module 10 gathers is two-dimensional infrared image.Due in visible images, the object in scene can form interference to the identification of marked region, and infrared image is because having filtered out visible ray information, disturbs less, and therefore two-dimensional infrared image is more conducive to extracting marked region.
Concrete, the first image processing module 202 is for setting up color model in advance.The color of such as marked region is red, then set up red model in advance, in this model, the rgb value component of pixel can between 200 to 255, and G, B component can close to zero; First image processing module 202 is red pixel for the pixel of the rgb value meeting this red model in getting frame image.In addition, when forming marked region by human body in the image gathered, the first image processing module 202 is for obtaining the pixel of mating with default complexion model in image.First image processing module 202 also for carrying out connected domain detection to the pixel obtained, obtains multiple connected domain, if connected domain is the set of individual continuous print pixel composition.
In the present embodiment, because the size and shape of marked region should be roughly changeless, the first image processing module 202, when carrying out connected domain to the pixel obtained and detecting, can calculate girth and/or the area of all connected domains in the pixel of acquisition.Concrete, the girth of connected domain can be the number of connected domain boundary pixel, and the area of connected domain can be the number of the whole pixels in connected domain.Further, the first image processing module 202 can be used for the girth of the girth of the connected domain of acquisition and/or area and default marked region and/or area to contrast, and obtains to meet to preset the girth of marked region and/or the connected domain of area is marked region.Preferably, the first image processing module 202 also can be used for using girth square with the ratio of area as judgment criterion, this ratio of connected domain meets this ratio of default marked region, then this connected domain is marked region.
First attitude generation module 204, for obtaining the pixel coordinate in marked region, produces the attitude of marked region according to this pixel coordinate.
In the present embodiment, the attitude that marked region produces is attitude angle.In one embodiment, marked region is a continuum, then the first attitude generation module 204 is for calculating the covariance matrix of pixel coordinate, obtain covariance matrix eigenvalue of maximum characteristic of correspondence vector, produce the attitude of marked region according to proper vector, the attitude of this marked region is an attitude angle.
In another embodiment, marked region comprises the first continuum and the second continuum, then the first attitude generation module 204 is for the center of gravity of the center of gravity and the second continuum that calculate the first continuum according to pixel coordinate, calculates the attitude of marked region according to the pixel coordinate of the pixel coordinate of the center of gravity of the first continuum and the center of gravity of the second continuum.Concrete, calculate the mean value of the whole pixel coordinates in continuum, the pixel coordinate obtained is the center of gravity of continuum.
In another embodiment, the image that image capture module 10 collects is 3-D view.Concrete, image capture module 10 can adopt traditional stereo visual system (video camera known by two control positions and related software form), structured-light system (right video camera, a light source and a related software forms) or TOF (time of flight, flight time) depth camera to realize collection 3-D view (i.e. three dimensional depth image).
In the present embodiment, as shown in figure 13, gesture recognition module 20 comprises the second image processing module 210 and the second attitude generation module 220, wherein:
Second image processing module 210 is for described Image Segmentation Using, extract the connected domain in image, and calculate the property value of connected domain, the property value of connected domain and the marked region property value preset are contrasted, described marked region is the connected domain meeting described default marked region property value.
Concrete, second image processing module 210, for when two adjacent pixel depth differences are less than the threshold value preset in 3-D view, such as 5 centimetres, then thinks that two pixels are communicated with, connected domain detection is carried out to whole image, a series of connected domains comprising mark connected domain can be obtained.
In the present embodiment, the property value of connected domain comprises the size and dimension of connected domain.Concrete, second image processing module 210 is for calculating the size/shape of connected domain, contrast with the size/shape of the mark on interactive device, the connected domain obtaining the size/shape meeting mark is the connected domain of marked region (marked region).For rectangle marked, being marked in the image of collection namely on interactive device is rectangle, the length of pre-set mark and width, second image processing module 210 is for calculating length and the width of physical region corresponding to connected domain, length and the width of this length and width and mark are more close, then connected domain is more similar to marked region.
Further, second image processing module 210 is as follows for the process of the length and width that calculate physical region corresponding to connected domain: the covariance matrix calculating the three-dimensional coordinate of connected domain pixel, the length of the physical region adopting following formulae discovery connected domain corresponding and width: wherein, k is the coefficient preset, such as, be set to 4, and when λ is covariance matrix eigenvalue of maximum, then l is the length of connected domain, and when λ is the second largest eigenwert of covariance matrix, then l is the width of connected domain.
Further, second image processing module 210 also can be used for the length breadth ratio presetting rectangle marked, such as length breadth ratio is 2, then the length breadth ratio of the physical region that connected domain is corresponding is more close to the length breadth ratio of the rectangle marked of default settings, then connected domain is more similar to marked region, concrete, attribute matching module 234 is for the length breadth ratio of the physical region that adopts following formulae discovery connected domain corresponding: wherein, r is the length breadth ratio of connected domain, λ 0for the eigenvalue of maximum of covariance matrix, λ 1for the Second Largest Eigenvalue of covariance matrix.
Second attitude generation module 220, for obtaining the pixel coordinate in marked region, produces the attitude of marked region according to described pixel coordinate.
In the present embodiment, the attitude of marked region is attitude vectors.In one embodiment, marked region is a continuum, then the second attitude generation module 220 is for calculating the covariance matrix of pixel coordinate, obtains covariance matrix eigenvalue of maximum characteristic of correspondence vector, produces the attitude of marked region according to proper vector.As mentioned above, the attitude of this marked region is an attitude vectors.
In another embodiment, marked region comprises the first continuum and the second continuum, then the second attitude generation module 220 is for the center of gravity of the center of gravity and the second continuum that calculate the first continuum according to pixel coordinate, produces the attitude of marked region according to the pixel coordinate of the pixel coordinate of the center of gravity of the first continuum and the center of gravity of the second continuum.In the present embodiment, the pixel coordinate in marked region is three-dimensional coordinate, concrete, and can produce the attitude of marked region according to the pixel coordinate of the center of gravity of two continuums calculated, this attitude is an attitude vectors.
In one embodiment, gesture recognition module 20 also comprises judge module (not shown), for judging that the image gathered is two dimensional image or 3-D view.Concrete, in the present embodiment, when the image that judge module determines collection is two dimensional image, then notifies that the first image processing module 202 extracts the marked region in two dimensional image, and then produce the attitude of this marked region by the first attitude generation module 204.When the image that judge module determines collection is two dimensional image, then notifies that the second image processing module 210 extracts the marked region in 3-D view, and then produce the attitude of this marked region by the second attitude generation module 220.Understandable, in the present embodiment, gesture recognition module 20 comprises judge module (not shown), the first image processing module 202, first attitude generation module 204, second image processing module 210 and the second attitude generation module 220 simultaneously.The present embodiment, both by the attitude in two dimensional image identification marking region, again by the attitude in two dimensional image identification marking region.
As shown in figure 14, in one embodiment, directive generation module 30 comprises the first attitude acquisition module 302 and module 304 is searched in the first instruction, wherein:
First attitude acquisition module 302 for obtaining the attitude of the described marked region in current frame image from gesture recognition module 20.
Concrete, this attitude can be the attitude angle of the marked region in the two dimensional image of present frame, also can be the attitude vectors of the marked region in the three dimensional depth image of present frame.In the present embodiment, the mapping relations between attitude and steering order are preset.This attitude also can be described as absolute pose.
First instruction searches module 304 for generating the steering order corresponding with described attitude according to the attitude preset with the mapping relations between steering order.
In the present embodiment, the image comprising marked region gathered can be image sequence.First attitude acquisition module 302 also for from gesture recognition module 20, obtain the marked region in the attitude of the marked region in current frame image and previous frame image attitude between relative attitude.First instruction searches module 304 also for generating the steering order corresponding with relative attitude according to the relative attitude preset with the mapping relations between steering order.
In another embodiment, the image comprising marked region gathered can be image sequence.As shown in figure 15, directive generation module 30 comprises the second attitude acquisition module 310 and module 320 is searched in the second instruction, wherein:
Second attitude acquisition module is used for the relative attitude between the attitude of the marked region obtained from gesture recognition module 20 in the attitude of the marked region in current frame image and previous frame image.
Second instruction searches module 320 for generating the steering order corresponding with relative attitude according to the relative attitude preset with the mapping relations between steering order.
Above-mentioned sign Language Recognition Method and system, go out according to the image recognition comprising marked region collected the attitude that marked region produces, and generate steering order corresponding to attitude, then this steering order is converted to the natural language information of normal person's easy understand.Because the image by obtaining limb action judges movement locus and the attitude of limb action, therefore, all there is record to the process of whole limb action, thus avoid the situation missing certain action or attitude in default of sensor, thus improve accuracy rate when identifying sign language.
The above embodiment only have expressed several embodiment of the present invention, and it describes comparatively concrete and detailed, but therefore can not be interpreted as the restriction to the scope of the claims of the present invention.It should be pointed out that for the person of ordinary skill of the art, without departing from the inventive concept of the premise, can also make some distortion and improvement, these all belong to protection scope of the present invention.Therefore, the protection domain of patent of the present invention should be as the criterion with claims.

Claims (8)

1. a sign Language Recognition Method, comprises the following steps:
Gather the image comprising marked region;
The attitude in identification marking region;
Generate the steering order that described attitude is corresponding;
Convert described steering order to natural language information;
The described step described steering order being converted to natural language information comprises:
Described steering order is arranged in steering order sequence, generates natural language information according to described steering order sequence;
The described step according to described steering order sequence generation natural language information comprises:
Calculate the eigenwert of described steering order sequence, the mapping relations according to the eigenwert preset and natural language information generate natural language information;
When calculating the eigenwert of steering order sequence, first calculate the posture feature value of the steering order in steering order sequence, then according to posture feature value by steering order sequence segment so that the steering order in steering order subsequence after segmentation has identical posture feature value; The monotonicity of the coordinate then described according to the coordinate part of the steering order in steering order subsequence, in image, the variance of the distance of a certain reference coordinate calculates the coordinate characteristic value of steering order subsequence; Then the posture feature value of all steering order subsequences and coordinate characteristic value are integrated into the eigenwert of steering order sequence;
Also comprise before the step of the eigenwert of the described steering order sequence of described calculating:
The steering order repeated in described steering order sequence is removed.
2. according to the sign Language Recognition Method described in claim 1, it is characterized in that, also comprise before the step of the eigenwert of the described steering order sequence of described calculating:
The steering order being greater than threshold value in described steering order sequence with adjacent steering order difference is removed.
3. according to the sign Language Recognition Method described in claim 1, it is characterized in that, the described step described steering order being arranged in steering order sequence comprises:
Described steering order is arranged in queue according to the order generated, when detecting that there is the identical steering order of continuous predetermined number at queue end, arranges complete, and generating steering order sequence according to described queue.
4. the sign Language Recognition Method according to any one of claims 1 to 3, is characterized in that, described described steering order is converted to the step of natural language information after also comprise:
Described natural language information is shown by the mode of text and/or audio frequency.
5. a sign Language Recognition, is characterized in that, comprises with lower module:
Image capture module, for gathering the image comprising marked region;
Gesture recognition module, for the attitude in identification marking region;
Directive generation module, for generating steering order corresponding to described attitude;
Instruction transformation module, for converting described steering order to natural language information;
Described instruction transformation module, also for described steering order is arranged in steering order sequence, generates natural language information according to described steering order sequence;
Described instruction transformation module is also for calculating the eigenwert of described steering order sequence, and the mapping relations according to the eigenwert preset and natural language information generate natural language information;
When calculating the eigenwert of steering order sequence, described instruction transformation module is also for calculating the posture feature value of the steering order in steering order sequence, then according to posture feature value by steering order sequence segment so that the steering order in steering order subsequence after segmentation has identical posture feature value; The monotonicity of the coordinate then described according to the coordinate part of the steering order in steering order subsequence, in image, the variance of the distance of a certain reference coordinate calculates the coordinate characteristic value of steering order subsequence; Then the posture feature value of all steering order subsequences and coordinate characteristic value are integrated into the eigenwert of steering order sequence;
The steering order of described instruction transformation module also for repeating in described steering order sequence is removed.
6. according to the sign Language Recognition described in claim 5, it is characterized in that, described instruction transformation module is also for removing the steering order being greater than threshold value in described steering order sequence with adjacent steering order difference.
7. according to the sign Language Recognition described in claim 5, it is characterized in that, described instruction transformation module is also for arranging queue by described steering order according to the order generated, when detecting that there is the identical steering order of continuous predetermined number at queue end, arrange complete, and generate steering order sequence according to described queue.
8. the sign Language Recognition according to any one of claim 5 to 7, is characterized in that, also comprises information display module, for showing described natural language information by the mode of text and/or audio frequency.
CN201210031595.8A 2011-12-02 2012-02-13 Sign Language Recognition Method and system Active CN103136986B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210031595.8A CN103136986B (en) 2011-12-02 2012-02-13 Sign Language Recognition Method and system

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN201110396235 2011-12-02
CN201110396235.3 2011-12-02
CN201210031595.8A CN103136986B (en) 2011-12-02 2012-02-13 Sign Language Recognition Method and system

Publications (2)

Publication Number Publication Date
CN103136986A CN103136986A (en) 2013-06-05
CN103136986B true CN103136986B (en) 2015-10-28

Family

ID=48488552

Family Applications (12)

Application Number Title Priority Date Filing Date
CN201110453879.1A Active CN103135756B (en) 2011-12-02 2011-12-29 Generate the method and system of control instruction
CN201110451724.4A Active CN103135754B (en) 2011-12-02 2011-12-29 Adopt interactive device to realize mutual method
CN201110451741.8A Active CN103135755B (en) 2011-12-02 2011-12-29 Interactive system and method
CN201210011346.2A Active CN103135882B (en) 2011-12-02 2012-01-13 Control the method and system that window picture shows
CN201210011308.7A Active CN103135881B (en) 2011-12-02 2012-01-13 Display control method and system
CN201210023419XA Pending CN103139508A (en) 2011-12-02 2012-02-02 Method and system for controlling display of television pictures
CN201210024389.4A Active CN103127717B (en) 2011-12-02 2012-02-03 The method and system of control operation game
CN201210024483.XA Active CN103135883B (en) 2011-12-02 2012-02-03 Control the method and system of window
CN201210025300.6A Active CN103135453B (en) 2011-12-02 2012-02-06 Control method and system of household appliances
CN201210031595.8A Active CN103136986B (en) 2011-12-02 2012-02-13 Sign Language Recognition Method and system
CN201210032934.4A Active CN103135759B (en) 2011-12-02 2012-02-14 Control method for playing multimedia and system
CN201210032932.5A Active CN103135758B (en) 2011-12-02 2012-02-14 Realize the method and system of shortcut function

Family Applications Before (9)

Application Number Title Priority Date Filing Date
CN201110453879.1A Active CN103135756B (en) 2011-12-02 2011-12-29 Generate the method and system of control instruction
CN201110451724.4A Active CN103135754B (en) 2011-12-02 2011-12-29 Adopt interactive device to realize mutual method
CN201110451741.8A Active CN103135755B (en) 2011-12-02 2011-12-29 Interactive system and method
CN201210011346.2A Active CN103135882B (en) 2011-12-02 2012-01-13 Control the method and system that window picture shows
CN201210011308.7A Active CN103135881B (en) 2011-12-02 2012-01-13 Display control method and system
CN201210023419XA Pending CN103139508A (en) 2011-12-02 2012-02-02 Method and system for controlling display of television pictures
CN201210024389.4A Active CN103127717B (en) 2011-12-02 2012-02-03 The method and system of control operation game
CN201210024483.XA Active CN103135883B (en) 2011-12-02 2012-02-03 Control the method and system of window
CN201210025300.6A Active CN103135453B (en) 2011-12-02 2012-02-06 Control method and system of household appliances

Family Applications After (2)

Application Number Title Priority Date Filing Date
CN201210032934.4A Active CN103135759B (en) 2011-12-02 2012-02-14 Control method for playing multimedia and system
CN201210032932.5A Active CN103135758B (en) 2011-12-02 2012-02-14 Realize the method and system of shortcut function

Country Status (1)

Country Link
CN (12) CN103135756B (en)

Families Citing this family (42)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104349197B (en) * 2013-08-09 2019-07-26 联想(北京)有限公司 A kind of data processing method and device
JP5411385B1 (en) * 2013-08-12 2014-02-12 株式会社 ディー・エヌ・エー Server and method for providing game
CN104801042A (en) * 2014-01-23 2015-07-29 鈊象电子股份有限公司 Method for switching game screens based on player's hand waving range
CN103810922B (en) * 2014-01-29 2016-03-23 上海天昊信息技术有限公司 Sign language interpretation system
CN103902045A (en) * 2014-04-09 2014-07-02 深圳市中兴移动通信有限公司 Method and device for operating wallpaper via non-contact postures
CN105094785A (en) * 2014-05-20 2015-11-25 腾讯科技(深圳)有限公司 Method and device for generating color matching file
CN104391573B (en) * 2014-11-10 2017-05-03 北京华如科技股份有限公司 Method and device for recognizing throwing action based on single attitude sensor
CN104460988B (en) * 2014-11-11 2017-12-22 陈琦 A kind of input control method of smart mobile phone virtual reality device
KR101608172B1 (en) 2014-12-22 2016-03-31 주식회사 넥슨코리아 Device and method to control object
CN106139590B (en) * 2015-04-15 2019-12-03 乐线韩国股份有限公司 The method and apparatus of control object
US10543427B2 (en) * 2015-04-29 2020-01-28 Microsoft Technology Licensing, Llc Game controller function remapping via external accessory
CN105204354A (en) * 2015-09-09 2015-12-30 北京百度网讯科技有限公司 Smart home device control method and device
US10234955B2 (en) * 2015-09-28 2019-03-19 Nec Corporation Input recognition apparatus, input recognition method using maker location, and non-transitory computer-readable storage program
CN105892638A (en) * 2015-12-01 2016-08-24 乐视致新电子科技(天津)有限公司 Virtual reality interaction method, device and system
CN106896732B (en) * 2015-12-18 2020-02-04 美的集团股份有限公司 Display method and device of household appliance
CN105592367A (en) * 2015-12-23 2016-05-18 青岛海信电器股份有限公司 Image display parameter adjusting method and system
JP6370820B2 (en) * 2016-02-05 2018-08-08 株式会社バンダイナムコエンターテインメント Image generation system, game device, and program.
CN105760106B (en) * 2016-03-08 2019-01-15 网易(杭州)网络有限公司 A kind of smart home device exchange method and device
CN105930050B (en) * 2016-04-13 2018-01-26 腾讯科技(深圳)有限公司 Behavior determines method and device
CN106682593A (en) * 2016-12-12 2017-05-17 山东师范大学 Method and system for sign language conference based on gesture recognition
WO2018120657A1 (en) * 2016-12-27 2018-07-05 华为技术有限公司 Method and device for sharing virtual reality data
CN108668042B (en) * 2017-03-30 2021-01-15 富士施乐实业发展(中国)有限公司 Compound machine system
CN109558000B (en) 2017-09-26 2021-01-22 京东方科技集团股份有限公司 Man-machine interaction method and electronic equipment
CN107831996B (en) * 2017-10-11 2021-02-19 Oppo广东移动通信有限公司 Face recognition starting method and related product
CN107861682A (en) * 2017-11-03 2018-03-30 网易(杭州)网络有限公司 The control method for movement and device of virtual objects
CN108228251B (en) * 2017-11-23 2021-08-27 腾讯科技(上海)有限公司 Method and device for controlling target object in game application
CN108036479A (en) * 2017-12-01 2018-05-15 广东美的制冷设备有限公司 Control method, system, vision controller and the storage medium of air conditioner
CN110007748B (en) * 2018-01-05 2021-02-19 Oppo广东移动通信有限公司 Terminal control method, processing device, storage medium and terminal
WO2019153971A1 (en) * 2018-02-06 2019-08-15 广东虚拟现实科技有限公司 Visual interaction apparatus and marker
CN108765299B (en) * 2018-04-26 2022-08-16 广州视源电子科技股份有限公司 Three-dimensional graphic marking system and method
CN108693781A (en) * 2018-07-31 2018-10-23 湖南机电职业技术学院 Intelligent home control system
JP7262976B2 (en) * 2018-11-02 2023-04-24 キヤノン株式会社 Information processing device, information processing method and program
TWI681755B (en) * 2018-12-24 2020-01-11 山衛科技股份有限公司 System and method for measuring scoliosis
CN109711349B (en) * 2018-12-28 2022-06-28 百度在线网络技术(北京)有限公司 Method and device for generating control instruction
CN109816650B (en) * 2019-01-24 2022-11-25 强联智创(北京)科技有限公司 Target area identification method and system based on two-dimensional DSA image
CN111665727A (en) * 2019-03-06 2020-09-15 北京京东尚科信息技术有限公司 Method and device for controlling household equipment and household equipment control system
CN111803930A (en) * 2020-07-20 2020-10-23 网易(杭州)网络有限公司 Multi-platform interaction method and device and electronic equipment
CN115623254A (en) * 2021-07-15 2023-01-17 北京字跳网络技术有限公司 Video effect adding method, device, equipment and storage medium
CN113326849B (en) * 2021-07-20 2022-01-11 广东魅视科技股份有限公司 Visual data acquisition method and system
CN113499585A (en) * 2021-08-09 2021-10-15 网易(杭州)网络有限公司 In-game interaction method and device, electronic equipment and storage medium
CN113822186A (en) * 2021-09-10 2021-12-21 阿里巴巴达摩院(杭州)科技有限公司 Sign language translation, customer service, communication method, device and readable medium
CN113822187A (en) * 2021-09-10 2021-12-21 阿里巴巴达摩院(杭州)科技有限公司 Sign language translation, customer service, communication method, device and readable medium

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101527092A (en) * 2009-04-08 2009-09-09 西安理工大学 Computer assisted hand language communication method under special session context
CN101539994A (en) * 2009-04-16 2009-09-23 西安交通大学 Mutually translating system and method of sign language and speech
CN101763515A (en) * 2009-09-23 2010-06-30 中国科学院自动化研究所 Real-time gesture interaction method based on computer vision

Family Cites Families (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5594469A (en) * 1995-02-21 1997-01-14 Mitsubishi Electric Information Technology Center America Inc. Hand gesture machine control system
JPH0918708A (en) * 1995-06-30 1997-01-17 Omron Corp Image processing method, image input device, controller, image output device and image processing system using the method
KR19990011180A (en) * 1997-07-22 1999-02-18 구자홍 How to select menu using image recognition
US9292111B2 (en) * 1998-01-26 2016-03-22 Apple Inc. Gesturing with a multipoint sensing device
US10279254B2 (en) * 2005-10-26 2019-05-07 Sony Interactive Entertainment Inc. Controller having visually trackable object for interfacing with a gaming system
US20070124694A1 (en) * 2003-09-30 2007-05-31 Koninklijke Philips Electronics N.V. Gesture to define location, size, and/or content of content window on a display
CN100573548C (en) * 2004-04-15 2009-12-23 格斯图尔泰克股份有限公司 The method and apparatus of tracking bimanual movements
JP2006068315A (en) * 2004-09-02 2006-03-16 Sega Corp Pause detection program, video game device, pause detection method, and computer-readable recording medium recorded with program
CN100345085C (en) * 2004-12-30 2007-10-24 中国科学院自动化研究所 Method for controlling electronic game scene and role based on poses and voices of player
JP2009514106A (en) * 2005-10-26 2009-04-02 株式会社ソニー・コンピュータエンタテインメント System and method for interfacing with a computer program
KR100783552B1 (en) * 2006-10-11 2007-12-07 삼성전자주식회사 Input control method and device for mobile phone
US8726194B2 (en) * 2007-07-27 2014-05-13 Qualcomm Incorporated Item selection using enhanced control
CN101388138B (en) * 2007-09-12 2011-06-29 原相科技股份有限公司 Interaction image system, interaction apparatus and operation method thereof
CN101398896B (en) * 2007-09-28 2012-10-17 三星电子株式会社 Device and method for extracting color characteristic with strong discernment for image forming apparatus
JP4938617B2 (en) * 2007-10-18 2012-05-23 幸輝郎 村井 Object operating device and method for specifying marker from digital image frame data
CN101483005A (en) * 2008-01-07 2009-07-15 致伸科技股份有限公司 Remote control device for multimedia file playing
JP5174492B2 (en) * 2008-03-05 2013-04-03 公立大学法人首都大学東京 Image recognition apparatus, image recognition method, image recognition program, gesture motion recognition system, gesture motion recognition method, and gesture motion recognition program
CN101551700B (en) * 2008-03-31 2012-03-28 联想(北京)有限公司 Electronic game input device, electronic game machine and electronic game input method
JP5697590B2 (en) * 2008-04-02 2015-04-08 オブロング・インダストリーズ・インコーポレーテッド Gesture-based control using 3D information extracted from extended subject depth
KR100978929B1 (en) * 2008-06-24 2010-08-30 한국전자통신연구원 Registration method of reference gesture data, operation method of mobile terminal and mobile terminal
CN101504728B (en) * 2008-10-10 2013-01-23 深圳泰山在线科技有限公司 Remote control system and method of electronic equipment
CN101729808B (en) * 2008-10-14 2012-03-28 Tcl集团股份有限公司 Remote control method for television and system for remotely controlling television by same
CN101465116B (en) * 2009-01-07 2013-12-11 北京中星微电子有限公司 Display equipment and control method thereof
CN101504586A (en) * 2009-03-25 2009-08-12 中国科学院软件研究所 Instruction method based on stroke tail gesture
CN101673094A (en) * 2009-09-23 2010-03-17 曾昭兴 Control device of home appliance and control method thereof
US20110151974A1 (en) * 2009-12-18 2011-06-23 Microsoft Corporation Gesture style recognition and reward
CN101799717A (en) * 2010-03-05 2010-08-11 天津大学 Man-machine interaction method based on hand action catch
CN101833653A (en) * 2010-04-02 2010-09-15 上海交通大学 Figure identification method in low-resolution video
US20110289455A1 (en) * 2010-05-18 2011-11-24 Microsoft Corporation Gestures And Gesture Recognition For Manipulating A User-Interface
CN201750431U (en) * 2010-07-02 2011-02-16 厦门万安智能股份有限公司 Smart home centralized control device
CN102179048A (en) * 2011-02-28 2011-09-14 武汉市高德电气有限公司 Method for implementing realistic game based on movement decomposition and behavior analysis
CN102226880A (en) * 2011-06-03 2011-10-26 北京新岸线网络技术有限公司 Somatosensory operation method and system based on virtual reality

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101527092A (en) * 2009-04-08 2009-09-09 西安理工大学 Computer assisted hand language communication method under special session context
CN101539994A (en) * 2009-04-16 2009-09-23 西安交通大学 Mutually translating system and method of sign language and speech
CN101763515A (en) * 2009-09-23 2010-06-30 中国科学院自动化研究所 Real-time gesture interaction method based on computer vision

Also Published As

Publication number Publication date
CN103135882B (en) 2016-08-03
CN103139508A (en) 2013-06-05
CN103135754A (en) 2013-06-05
CN103135453A (en) 2013-06-05
CN103135881B (en) 2016-12-14
CN103135755A (en) 2013-06-05
CN103135882A (en) 2013-06-05
CN103127717A (en) 2013-06-05
CN103135759B (en) 2016-03-09
CN103135754B (en) 2016-05-11
CN103135758A (en) 2013-06-05
CN103135758B (en) 2016-09-21
CN103135883B (en) 2016-07-06
CN103135881A (en) 2013-06-05
CN103135756B (en) 2016-05-11
CN103135756A (en) 2013-06-05
CN103135755B (en) 2016-04-06
CN103135759A (en) 2013-06-05
CN103127717B (en) 2016-02-10
CN103135883A (en) 2013-06-05
CN103136986A (en) 2013-06-05
CN103135453B (en) 2015-05-13

Similar Documents

Publication Publication Date Title
CN103136986B (en) Sign Language Recognition Method and system
CN102822862B (en) Calculation element interface
CN103336576B (en) A kind of moving based on eye follows the trail of the method and device carrying out browser operation
CN102638653B (en) Automatic face tracing method on basis of Kinect
CN108509026B (en) Remote maintenance support system and method based on enhanced interaction mode
CN105718878A (en) Egocentric vision in-the-air hand-writing and in-the-air interaction method based on cascade convolution nerve network
CN103150019A (en) Handwriting input system and method
CN103500010B (en) A kind of video fingertip localization method
CN107992792A (en) A kind of aerial handwritten Chinese character recognition system and method based on acceleration transducer
CN102799277A (en) Wink action-based man-machine interaction method and system
CN109241909A (en) A kind of long-range dance movement capture evaluating system based on intelligent terminal
US20160252966A1 (en) Method by which eyeglass-type display device recognizes and inputs movement
EP1148411A3 (en) Information processing apparatus and method for recognising user gesture
CN103995595A (en) Game somatosensory control method based on hand gestures
CN206209206U (en) 3D glasses with fixed sample point and the virtual reality system of Portable multi-person interaction
CN109839827B (en) Gesture recognition intelligent household control system based on full-space position information
CN109375766A (en) A kind of Novel learning method based on gesture control
CN103389793B (en) Man-machine interaction method and system
CN110134241A (en) Dynamic gesture exchange method based on monocular cam
CN108521594A (en) A kind of free viewpoint video playback method based on body-sensing camera gesture identification
CN102289328A (en) Human-computer interaction method and device based on handheld camera
CN103995586B (en) Non- wearing based on virtual touch screen refers to gesture man-machine interaction method
CN103324291A (en) Method for obtaining position of human body interesting area relative to screen window
CN109218833A (en) The method and system that control television image is shown
Wang et al. A multimodal gesture recognition dataset for desktop human-computer interaction

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
C56 Change in the name or address of the patentee
CP01 Change in the name or title of a patent holder

Address after: 518000 Shenzhen, Nanshan District Province, science and technology, South Road, the building of the big square, building 02, room 4,

Patentee after: SHENZHEN TAISHAN SPORTS TECHNOLOGY CORP., LTD.

Address before: 518000 Shenzhen, Nanshan District Province, science and technology, South Road, the building of the big square, building 02, room 4,

Patentee before: Shenzhen Tol Technology Co., Ltd.

CP01 Change in the name or title of a patent holder

Address after: 518000 room 02, 4th floor, Fangda building, Keji South 12th Road, Nanshan District, Shenzhen City, Guangdong Province

Patentee after: Shenzhen Taishan Sports Technology Co.,Ltd.

Address before: 518000 room 02, 4th floor, Fangda building, Keji South 12th Road, Nanshan District, Shenzhen City, Guangdong Province

Patentee before: SHENZHEN TAISHAN SPORTS TECHNOLOGY Corp.,Ltd.

CP01 Change in the name or title of a patent holder