CN107622257A - A kind of neural network training method and three-dimension gesture Attitude estimation method - Google Patents
A kind of neural network training method and three-dimension gesture Attitude estimation method Download PDFInfo
- Publication number
- CN107622257A CN107622257A CN201710954487.0A CN201710954487A CN107622257A CN 107622257 A CN107622257 A CN 107622257A CN 201710954487 A CN201710954487 A CN 201710954487A CN 107622257 A CN107622257 A CN 107622257A
- Authority
- CN
- China
- Prior art keywords
- gesture
- mrow
- neural network
- depth
- network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 238000012549 training Methods 0.000 title claims abstract description 44
- 238000013528 artificial neural network Methods 0.000 title claims abstract description 21
- 238000012545 processing Methods 0.000 claims abstract description 22
- 238000012360 testing method Methods 0.000 claims abstract description 11
- 238000013527 convolutional neural network Methods 0.000 claims abstract description 10
- 238000007637 random forest analysis Methods 0.000 claims abstract description 10
- 238000002372 labelling Methods 0.000 claims description 9
- 238000011176 pooling Methods 0.000 claims description 9
- 238000005070 sampling Methods 0.000 claims description 8
- 239000000203 mixture Substances 0.000 claims description 5
- 210000002569 neuron Anatomy 0.000 claims description 3
- 230000036544 posture Effects 0.000 description 20
- 238000001514 detection method Methods 0.000 description 5
- 230000008859 change Effects 0.000 description 3
- 238000013135 deep learning Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 230000009471 action Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 230000018109 developmental process Effects 0.000 description 2
- 210000004247 hand Anatomy 0.000 description 2
- 238000005286 illumination Methods 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 210000000707 wrist Anatomy 0.000 description 1
Landscapes
- Image Analysis (AREA)
Abstract
The invention discloses a kind of neural network training method and three-dimension gesture Attitude estimation method, including:S1:The data set of multiple gesture depth maps is included by depth camera collection;S2:Random forest learner is trained using step S1 data set;S3:Multiple gesture depth maps in step S1 data set are split using random forest learner, it is partitioned into gesture subgraph, again the gesture subgraph is handled to obtain processing figure, multiple gesture depth maps in the processing figure and step S1 data set are carried out out of order to be divided into training set and test set;S4:The obtained training sets of step S3 and test set are used for training convolutional neural networks, training obtains network model.Three-dimension gesture Attitude estimation method is that the three-dimension gesture posture in individual depth picture is estimated using the network model.The present invention can accurately identify the particular location and posture of palm finger in gesture.
Description
Technical Field
The invention relates to the field of computer vision and deep learning, in particular to a neural network training method and a three-dimensional gesture posture estimation method.
Background
In recent years, with the rapid development of computer vision and deep learning, virtual reality and augmented reality technologies are gradually popularized, and still have immeasurable development prospects. As an important means of human-computer interaction, the gesture recognition technology has been highly concerned by the field of computer vision, and because the human hand has more joints, complicated shape, higher degree of freedom and easy shielding phenomenon, the rapid and accurate recognition of the gesture position and the hand action is always a difficult problem.
Conventional gesture pose estimation methods can be generally divided into two categories: sensor based and image based. The gesture posture estimation technology based on the sensor is that the sensors such as an accelerometer, an angular velocity meter and the like are fixed at specific parts of a palm and a finger of a person; the position and motion state information of a specific part of a human hand are acquired through the worn sensor equipment, and then the states of the palm and the fingers of the human hand are calculated by using a kinematics method, so that the gesture posture estimation effect is achieved; the method has great limitation on gesture detection due to the fact that sensor equipment needs to be worn, and detection errors are generally large under the influence of factors such as the accuracy of the sensor and the change of the wearing position. Another gesture posture estimation method based on images is generally to use edge or region detection-based methods such as edge detection and skin color detection on images including human hands shot by an RGB camera, firstly determine approximate regions of the human hands in the images, and then segment detailed information such as fingers and wrists by means of image segmentation and the like; because a picture containing a hand is shot by a common camera, the picture can only reflect plane information of a scene generally, if occlusion occurs between fingers, action details of the occluded fingers cannot be identified, and therefore a large error exists.
The above background disclosure is only for the purpose of assisting understanding of the concept and technical solution of the present invention and does not necessarily belong to the prior art of the present patent application, and should not be used for evaluating the novelty and inventive step of the present application in the case that there is no clear evidence that the above content is disclosed at the filing date of the present patent application.
Disclosure of Invention
In order to solve the technical problems, the invention provides a neural network training method and a three-dimensional gesture attitude estimation method, which can accurately identify the specific position and attitude of a palm finger in a gesture.
In order to achieve the purpose, the invention adopts the following technical scheme:
the invention discloses a neural network training method, which comprises the following steps:
s1: acquiring, by a depth camera, a data set comprising a plurality of gesture depth maps;
s2: training a random forest learner by using the data set of the step S1;
s3: adopting a random forest learner to segment the gesture depth maps in the data set of the step S1 to obtain gesture sub-maps, processing the gesture sub-maps to obtain a processing map, and disordering the processing map and the gesture depth maps in the data set of the step S1 to divide the processing map and the gesture depth maps into a training set and a testing set;
s4: and (5) using the training set and the test set obtained in the step (S3) to train the convolutional neural network, and training to obtain a network model.
Preferably, the step S3 of processing the gesture sub-graph to obtain a processing graph includes: s32: projecting the gesture subgraph in the direction of three axes X, Y, Z to obtain three single-channel projection graphs; wherein the processing map includes the projection map in step S32.
Preferably, the step S3 of processing the gesture sub-graph to obtain a processing graph further includes: s33: respectively carrying out down-sampling on the three projection drawings to obtain a plurality of down-sampling drawings with different sizes; wherein the processing map includes the projection map in step S32 and the down-sampling map in step S33.
Preferably, step S1 specifically includes:
s11: acquiring a plurality of gesture depth maps of different people by adopting a plurality of depth cameras;
s12: labeling each gesture depth map, and storing a plurality of gesture depth maps and corresponding labeling information in a data set.
Preferably, the labeling of each gesture depth map in step S12 specifically includes: and labeling coordinate information (x, y, d) on the preset position of the finger and the palm in each gesture depth map, wherein x and y are horizontal and vertical coordinates on the gesture depth map, and d is the pixel depth.
Preferably, the predetermined position of the finger comprises all joint points of the finger.
Preferably, step S4 specifically includes:
s41: randomly selecting m pictures and corresponding label information from the training set, and randomly selecting n pictures and corresponding label information from the testing set;
s42: the pictures are rolled and laminated in the network;
s43: pictures are passed through a pooling layer in the network;
s44: the output layer restores the picture;
s45: calculating the error between the network output and the label information, learning the network, and updating the network parameters;
s46: repeatedly iterating the steps S42-S45, and continuously updating the parameters until the parameters are converged; and storing the trained parameters to finally obtain the trained network model.
Preferably, step S45 is specifically: the formula for calculating the error between the network output and the tag information is as follows:
wherein,for predicted tag coordinates, byComposition, J is original label, consisting of (J)1,j2,...,jn) The composition, n is the number of labels,
assuming that the network parameter of the neuron in the network is ω, updating the network parameter according to the following formula:
the invention also discloses a three-dimensional gesture posture estimation method, and the three-dimensional gesture posture in a single depth picture is estimated by adopting the network model obtained by training of the neural network training method.
Compared with the prior art, the invention has the beneficial effects that: according to the neural network training method, the gesture depth map acquired by the depth camera can accurately identify the gesture and position information of the palm and each finger, then the gesture depth map is segmented by the random forest learner, so that the feature information of the gesture in the picture can be beneficially explored, the residual neural network is trained through the set of the picture, and the convolution pooling layer of the neural network can learn the features in the regions with different scales in the picture, so that the trained network model is applied to three-dimensional gesture posture estimation, the shielding influence can be weakened, and the image-based method is not restricted by wearable equipment; by using the residual convolutional neural network, the problem of gradient dispersion in the process of back propagation and parameter updating is avoided, so that the network training effect is better; the three-dimensional gesture posture estimation method combines the use of a depth learning method and a depth camera and is applied to gesture recognition, and the influence of factors such as illumination change, object shielding and the like can be reduced for the gesture recognition.
In a further scheme, by performing horizontal, vertical and depth triaxial projection on the gesture subgraph, a picture with a three-dimensional view angle in the picture can be obtained, and the feature information of the gesture can be conveniently found; furthermore, the gesture subgraphs are downsampled to obtain multiscale pictures with different sizes, so that the pixel characteristics and the region characteristics with different sizes of the pictures can be better developed, and the specific positions and postures of the palms and fingers in the gestures can be more accurately identified by a trained network model; and can accurately recognize the detailed information of the occluded gesture.
Drawings
FIG. 1 is a flow chart diagram of a three-dimensional gesture pose estimation method in accordance with a preferred embodiment of the present invention;
FIG. 2 is a schematic diagram of the marker points of the fingers and palm of the hand of the preferred embodiment of the present invention;
FIG. 3 is a schematic diagram of the steps of the neural network training method according to the preferred embodiment of the present invention.
Detailed Description
The invention will be further described with reference to the accompanying drawings and preferred embodiments.
As shown in fig. 1, the three-dimensional gesture posture estimation method of the preferred embodiment of the present invention includes the following steps:
s1: collecting a data set of a gesture depth map; the method specifically comprises the following steps:
s11: acquiring gesture depth pictures of different people by using a plurality of depth cameras, acquiring a plurality of pictures containing a plurality of gestures with different angles and different postures for each gesture of each person, and sorting the acquired pictures into a picture library;
s12: labeling each picture in the picture library; in order to accurately position the detailed information of the positions and postures of the joints of the gesture, coordinate information (x, y, d) is marked at specific positions of fingers and palms in the embodiment, wherein x and y are horizontal and vertical coordinates of a foreground image (on a gesture depth image), and d is pixel depth which is the embodiment of the gesture depth on the image; as shown in fig. 2, a plurality of key points are set at specific positions of fingers and palms as specific mark points, the hand on each picture is marked as a label of a picture library, and the picture name and the corresponding label are stored in a file form; wherein the key points marked in fig. 2 include all the joint points of the five fingers and the important position point of the palm, the posture of the current hand can be accurately estimated by accurately predicting the position of each joint point.
S2: training a random forest learner by using the data set obtained in the step S1;
s3: preprocessing a gesture depth map in the data set; as shown in fig. 3, the method specifically includes the following steps:
s31: a random forest learner is adopted to segment the gesture depth map in the data set in the step S1 to obtain a gesture sub-map;
s32: projecting the gesture subgraph in the direction of three axes X, Y, Z to obtain three single-channel projection graphs;
s33: respectively carrying out down-sampling on the three projection drawings to obtain a plurality of down-sampling drawings with different sizes;
s34: sorting all the gesture depth maps in the data set in the step S1, the projection map obtained in the step S32 and the downsampling map obtained in the step S33, and dividing the data into a training set and a testing set according to the proportion of 90% and 10%.
S4: using the training set and the test set obtained in the step S34 to train a convolutional neural network, and training to obtain a network model; the method specifically comprises the following steps:
s41: randomly selecting m pictures and corresponding label information from the training set, and randomly selecting n pictures and corresponding label information from the testing set;
s42: the pictures are rolled and laminated in the network; assuming that the original size of the picture is l × l, k square matrices with the same size and different pixel values are selected as convolution kernels, and the size of the convolution kernels can be represented as k × c. Wherein k is the number of convolution kernels, and c is the number of parameters of each dimension of the convolution kernels; and (3) performing convolution operation on each picture and k convolution kernels respectively to obtain k pictures with the same size but not the same pixel point. New dimension lc*lcThe size is shown in the following formula:
lc*lc=(l-c+1)*(l-c+1)
s43: pictures pass through a pooling layer in a network; assuming that the size of the picture before entering the pooling layer is l x l, pooling is a sliding of an area of size p x p over the picture each time in f steps; each sliding time, one pixel in the area is selected to represent all pixels of the area, and the size of each picture is changed into l after passing through the pooling layerp*lpAs shown in the following formula:
lp*lp=((lp-(f-p))/f)*((lp-(f-p))/f)
s44: after the picture is processed by convolution pooling and the like in a network, an output layer restores the predicted picture; assuming that h convolution kernels are output after network training, and when the h convolution kernels reach the inlet of an output layer through the network, assuming that the size of each picture is le*le(le<li),liThe size of the picture entering the output layer is h x le*leDimension l is coupled through the output layero*loIs recovered to li*li。
S45: calculating the difference between the network output and the standard label, learning the network, and updating the network parameters; the Euclidean distance of the calculated error is shown as follows:
wherein,for predicted tag coordinates, byAnd (4) forming. J is the original label, consisting of (J)1,j2,...,jn) The composition, n is the number of labels,therefore, it is
Assuming that the network parameter of the neuron in the network is ω, updating the network parameter according to the following formula:
s46: repeatedly iterating the steps S42-S45, and continuously updating the parameters until the parameters are converged; and storing the trained parameters to finally obtain the trained convolutional neural network model.
S5: and estimating the three-dimensional gesture attitude of the single depth picture by adopting the convolutional neural network model obtained by training in the step S4.
The preferred embodiment of the invention also discloses a neural network training method, which comprises the steps S1 to S4.
According to the three-dimensional gesture posture estimation method in the preferred embodiment of the invention, a depth camera is used for collecting a large number of pictures; segmenting the gesture foreground by using a random forest classifier; manually marking gesture joint point information; training a convolutional neural network using a data set; storing the trained convolutional network, and directly using the convolutional network for three-dimensional gesture attitude estimation of a single depth map; the method applies the use of the depth learning method and the depth camera to gesture recognition, and the recognition of the gesture can reduce the influence of factors such as illumination change, object shielding and the like.
Acquiring a gesture posture image through a depth camera, displaying the depth information of a gesture in the form of a single-channel gray-scale image which represents the distance between an object and the camera by the pixel value, and restoring a gesture posture skeleton in the form of joint points according to the acquired gesture posture depth image; because the convolution pooling layer of the convolutional neural network can learn features in regions of different scales in the picture, occlusion effects can be reduced, and the image-based method is not constrained by the wearable device.
The three-dimensional gesture posture estimation method of the preferred embodiment of the invention overcomes the bottleneck of traditional gesture posture estimation, and realizes the accurate recognition of the specific position and posture of the palm and the fingers in the gesture by using a detection method of a human hand depth map shot by a depth camera and a new convolutional neural network method based on deep learning.
The foregoing is a more detailed description of the invention in connection with specific preferred embodiments and it is not intended that the invention be limited to these specific details. For those skilled in the art to which the invention pertains, several equivalent substitutions or obvious modifications can be made without departing from the spirit of the invention, and all the properties or uses are considered to be within the scope of the invention.
Claims (9)
1. A neural network training method is characterized by comprising the following steps:
s1: acquiring, by a depth camera, a data set comprising a plurality of gesture depth maps;
s2: training a random forest learner by using the data set of the step S1;
s3: adopting a random forest learner to segment the gesture depth maps in the data set of the step S1 to obtain gesture sub-maps, processing the gesture sub-maps to obtain a processing map, and disordering the processing map and the gesture depth maps in the data set of the step S1 to divide the processing map and the gesture depth maps into a training set and a testing set;
s4: and (5) using the training set and the test set obtained in the step (S3) to train the convolutional neural network, and training to obtain a network model.
2. The neural network training method of claim 1, wherein the processing the gesture sub-graph in step S3 to obtain a processing graph comprises:
s32: projecting the gesture subgraph in the direction of three axes X, Y, Z to obtain three single-channel projection graphs;
wherein the processing map includes the projection map in step S32.
3. The neural network training method of claim 2, wherein the processing the gesture sub-graph in step S3 to obtain a processing graph further comprises:
s33: respectively carrying out down-sampling on the three projection drawings to obtain a plurality of down-sampling drawings with different sizes;
wherein the processing map further includes the down-sampling map in step S33.
4. The neural network training method according to claim 1, wherein step S1 specifically includes:
s11: acquiring a plurality of gesture depth maps of different people by adopting a plurality of depth cameras;
s12: labeling each gesture depth map, and storing a plurality of gesture depth maps and corresponding labeling information in a data set.
5. The neural network training method of claim 4, wherein labeling each gesture depth map in step S12 specifically includes: and labeling coordinate information (x, y, d) on the preset position of the finger and the palm in each gesture depth map, wherein x and y are horizontal and vertical coordinates on the gesture depth map, and d is the pixel depth.
6. The neural network training method of claim 5, wherein the predetermined positions of the finger include all joint points of the finger.
7. The neural network training method according to claim 4, wherein the step S4 specifically includes:
s41: randomly selecting m pictures and corresponding label information from the training set, and randomly selecting n pictures and corresponding label information from the testing set;
s42: the pictures are rolled and laminated in the network;
s43: pictures are passed through a pooling layer in the network;
s44: the output layer restores the picture;
s45: calculating the error between the network output and the label information, learning the network, and updating the network parameters;
s46: repeatedly iterating the steps S42-S45, and continuously updating the parameters until the parameters are converged; and storing the trained parameters to finally obtain the trained network model.
8. The neural network training method according to claim 7, wherein the step S45 specifically comprises: the formula for calculating the error between the network output and the tag information is as follows:
<mrow> <mi>E</mi> <mi>r</mi> <mi>r</mi> <mi>o</mi> <mi>r</mi> <mo>=</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <mo>|</mo> <mo>|</mo> <mover> <mi>J</mi> <mo>^</mo> </mover> <mo>-</mo> <mi>J</mi> <mo>|</mo> <msup> <mo>|</mo> <mn>2</mn> </msup> <mo>=</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <mrow> <mo>|</mo> <mrow> <msup> <mrow> <mo>(</mo> <mover> <mi>x</mi> <mo>^</mo> </mover> <mo>-</mo> <mi>x</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msup> <mrow> <mo>(</mo> <mover> <mi>y</mi> <mo>^</mo> </mover> <mo>-</mo> <mi>y</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> <mo>+</mo> <msup> <mrow> <mo>(</mo> <mover> <mi>d</mi> <mo>^</mo> </mover> <mo>-</mo> <mi>d</mi> <mo>)</mo> </mrow> <mn>2</mn> </msup> </mrow> <mo>|</mo> </mrow> </mrow>
wherein,for predicted tag coordinates, byComposition, J is original label, consisting of (J)1,j2,...,jn) Composition, n is the number of labels, ji=(xi,yi,di),
Assuming that the network parameter of the neuron in the network is ω, updating the network parameter according to the following formula:
<mrow> <msup> <mi>&omega;</mi> <mo>&prime;</mo> </msup> <mo>=</mo> <mi>&omega;</mi> <mo>+</mo> <mfrac> <mrow> <mo>&part;</mo> <mi>E</mi> <mi>r</mi> <mi>r</mi> <mi>o</mi> <mi>r</mi> </mrow> <mrow> <mo>&part;</mo> <mi>&omega;</mi> </mrow> </mfrac> <mo>.</mo> </mrow>
9. a three-dimensional gesture posture estimation method is characterized in that a network model obtained by training through the neural network training method according to any one of claims 1 to 8 is adopted to estimate the three-dimensional gesture posture in a single depth picture.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710954487.0A CN107622257A (en) | 2017-10-13 | 2017-10-13 | A kind of neural network training method and three-dimension gesture Attitude estimation method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710954487.0A CN107622257A (en) | 2017-10-13 | 2017-10-13 | A kind of neural network training method and three-dimension gesture Attitude estimation method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107622257A true CN107622257A (en) | 2018-01-23 |
Family
ID=61092146
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710954487.0A Pending CN107622257A (en) | 2017-10-13 | 2017-10-13 | A kind of neural network training method and three-dimension gesture Attitude estimation method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107622257A (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108681402A (en) * | 2018-05-16 | 2018-10-19 | Oppo广东移动通信有限公司 | Identify exchange method, device, storage medium and terminal device |
CN108717524A (en) * | 2018-04-28 | 2018-10-30 | 天津大学 | It is a kind of based on double gesture recognition systems and method for taking the photograph mobile phone and artificial intelligence system |
CN108960036A (en) * | 2018-04-27 | 2018-12-07 | 北京市商汤科技开发有限公司 | 3 D human body attitude prediction method, apparatus, medium and equipment |
CN108960178A (en) * | 2018-07-13 | 2018-12-07 | 清华大学 | A kind of manpower Attitude estimation method and system |
CN109035327A (en) * | 2018-06-25 | 2018-12-18 | 北京大学 | Panorama camera Attitude estimation method based on deep learning |
CN109215080A (en) * | 2018-09-25 | 2019-01-15 | 清华大学 | 6D Attitude estimation network training method and device based on deep learning Iterative matching |
CN109858380A (en) * | 2019-01-04 | 2019-06-07 | 广州大学 | Expansible gesture identification method, device, system, gesture identification terminal and medium |
CN110197156A (en) * | 2019-05-30 | 2019-09-03 | 清华大学 | Manpower movement and the shape similarity metric method and device of single image based on deep learning |
WO2019196099A1 (en) * | 2018-04-09 | 2019-10-17 | 深圳大学 | Method for positioning boundaries of target object in medical image, storage medium, and terminal |
CN110555335A (en) * | 2018-05-30 | 2019-12-10 | 深圳市掌网科技股份有限公司 | Gesture recognition artificial intelligence training system and method thereof |
CN111104820A (en) * | 2018-10-25 | 2020-05-05 | 中车株洲电力机车研究所有限公司 | Gesture recognition method based on deep learning |
CN111325166A (en) * | 2020-02-26 | 2020-06-23 | 南京工业大学 | Sitting posture identification method based on projection reconstruction and multi-input multi-output neural network |
CN111368733A (en) * | 2020-03-04 | 2020-07-03 | 电子科技大学 | Three-dimensional hand posture estimation method based on label distribution learning, storage medium and terminal |
CN111428555A (en) * | 2020-01-17 | 2020-07-17 | 大连理工大学 | Joint-divided hand posture estimation method |
CN112085161A (en) * | 2020-08-20 | 2020-12-15 | 清华大学 | Graph neural network method based on random information transmission |
CN112836597A (en) * | 2021-01-15 | 2021-05-25 | 西北大学 | Multi-hand posture key point estimation method based on cascade parallel convolution neural network |
CN113408443A (en) * | 2021-06-24 | 2021-09-17 | 齐鲁工业大学 | Gesture posture prediction method and system based on multi-view images |
TWI777153B (en) * | 2020-04-21 | 2022-09-11 | 和碩聯合科技股份有限公司 | Image recognition method and device thereof and ai model training method and device thereof |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105005769A (en) * | 2015-07-08 | 2015-10-28 | 山东大学 | Deep information based sign language recognition method |
CN106326860A (en) * | 2016-08-23 | 2017-01-11 | 武汉闪图科技有限公司 | Gesture recognition method based on vision |
CN106815578A (en) * | 2017-01-23 | 2017-06-09 | 重庆邮电大学 | A kind of gesture identification method based on Depth Motion figure Scale invariant features transform |
CN107103613A (en) * | 2017-03-28 | 2017-08-29 | 深圳市未来媒体技术研究院 | A kind of three-dimension gesture Attitude estimation method |
-
2017
- 2017-10-13 CN CN201710954487.0A patent/CN107622257A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105005769A (en) * | 2015-07-08 | 2015-10-28 | 山东大学 | Deep information based sign language recognition method |
CN106326860A (en) * | 2016-08-23 | 2017-01-11 | 武汉闪图科技有限公司 | Gesture recognition method based on vision |
CN106815578A (en) * | 2017-01-23 | 2017-06-09 | 重庆邮电大学 | A kind of gesture identification method based on Depth Motion figure Scale invariant features transform |
CN107103613A (en) * | 2017-03-28 | 2017-08-29 | 深圳市未来媒体技术研究院 | A kind of three-dimension gesture Attitude estimation method |
Non-Patent Citations (1)
Title |
---|
贺芳姿: "基于Kinect深度信息的手势识别研究", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019196099A1 (en) * | 2018-04-09 | 2019-10-17 | 深圳大学 | Method for positioning boundaries of target object in medical image, storage medium, and terminal |
CN108960036A (en) * | 2018-04-27 | 2018-12-07 | 北京市商汤科技开发有限公司 | 3 D human body attitude prediction method, apparatus, medium and equipment |
CN108717524A (en) * | 2018-04-28 | 2018-10-30 | 天津大学 | It is a kind of based on double gesture recognition systems and method for taking the photograph mobile phone and artificial intelligence system |
CN108717524B (en) * | 2018-04-28 | 2022-05-06 | 天津大学 | Gesture recognition system based on double-camera mobile phone and artificial intelligence system |
WO2019218880A1 (en) * | 2018-05-16 | 2019-11-21 | Oppo广东移动通信有限公司 | Interaction recognition method and apparatus, storage medium, and terminal device |
CN108681402A (en) * | 2018-05-16 | 2018-10-19 | Oppo广东移动通信有限公司 | Identify exchange method, device, storage medium and terminal device |
CN110555335B (en) * | 2018-05-30 | 2022-12-27 | 深圳市掌网科技股份有限公司 | Gesture recognition artificial intelligence training system and method thereof |
CN110555335A (en) * | 2018-05-30 | 2019-12-10 | 深圳市掌网科技股份有限公司 | Gesture recognition artificial intelligence training system and method thereof |
CN109035327A (en) * | 2018-06-25 | 2018-12-18 | 北京大学 | Panorama camera Attitude estimation method based on deep learning |
CN109035327B (en) * | 2018-06-25 | 2021-10-29 | 北京大学 | Panoramic camera attitude estimation method based on deep learning |
CN108960178A (en) * | 2018-07-13 | 2018-12-07 | 清华大学 | A kind of manpower Attitude estimation method and system |
CN109215080A (en) * | 2018-09-25 | 2019-01-15 | 清华大学 | 6D Attitude estimation network training method and device based on deep learning Iterative matching |
US11200696B2 (en) | 2018-09-25 | 2021-12-14 | Tsinghua University | Method and apparatus for training 6D pose estimation network based on deep learning iterative matching |
CN111104820A (en) * | 2018-10-25 | 2020-05-05 | 中车株洲电力机车研究所有限公司 | Gesture recognition method based on deep learning |
CN109858380A (en) * | 2019-01-04 | 2019-06-07 | 广州大学 | Expansible gesture identification method, device, system, gesture identification terminal and medium |
CN110197156A (en) * | 2019-05-30 | 2019-09-03 | 清华大学 | Manpower movement and the shape similarity metric method and device of single image based on deep learning |
CN111428555A (en) * | 2020-01-17 | 2020-07-17 | 大连理工大学 | Joint-divided hand posture estimation method |
CN111428555B (en) * | 2020-01-17 | 2022-09-20 | 大连理工大学 | Joint-divided hand posture estimation method |
CN111325166A (en) * | 2020-02-26 | 2020-06-23 | 南京工业大学 | Sitting posture identification method based on projection reconstruction and multi-input multi-output neural network |
CN111325166B (en) * | 2020-02-26 | 2023-07-07 | 南京工业大学 | Sitting posture identification method based on projection reconstruction and MIMO neural network |
CN111368733A (en) * | 2020-03-04 | 2020-07-03 | 电子科技大学 | Three-dimensional hand posture estimation method based on label distribution learning, storage medium and terminal |
TWI777153B (en) * | 2020-04-21 | 2022-09-11 | 和碩聯合科技股份有限公司 | Image recognition method and device thereof and ai model training method and device thereof |
CN112085161A (en) * | 2020-08-20 | 2020-12-15 | 清华大学 | Graph neural network method based on random information transmission |
CN112085161B (en) * | 2020-08-20 | 2022-12-13 | 清华大学 | Graph neural network method based on random information transmission |
CN112836597A (en) * | 2021-01-15 | 2021-05-25 | 西北大学 | Multi-hand posture key point estimation method based on cascade parallel convolution neural network |
CN112836597B (en) * | 2021-01-15 | 2023-10-17 | 西北大学 | Multi-hand gesture key point estimation method based on cascade parallel convolution neural network |
CN113408443A (en) * | 2021-06-24 | 2021-09-17 | 齐鲁工业大学 | Gesture posture prediction method and system based on multi-view images |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107622257A (en) | A kind of neural network training method and three-dimension gesture Attitude estimation method | |
CN107103613B (en) | A kind of three-dimension gesture Attitude estimation method | |
CN108345869B (en) | Driver posture recognition method based on depth image and virtual data | |
JP7178396B2 (en) | Method and computer system for generating data for estimating 3D pose of object included in input image | |
CN108717531B (en) | Human body posture estimation method based on Faster R-CNN | |
CN106055091B (en) | A kind of hand gestures estimation method based on depth information and correcting mode | |
Hasan et al. | RETRACTED ARTICLE: Static hand gesture recognition using neural networks | |
CN112906604B (en) | Behavior recognition method, device and system based on skeleton and RGB frame fusion | |
CN108256504A (en) | A kind of Three-Dimensional Dynamic gesture identification method based on deep learning | |
Prisacariu et al. | 3D hand tracking for human computer interaction | |
CN112560741A (en) | Safety wearing detection method based on human body key points | |
CN111401266B (en) | Method, equipment, computer equipment and readable storage medium for positioning picture corner points | |
CN104794737B (en) | A kind of depth information Auxiliary Particle Filter tracking | |
CN106951840A (en) | A kind of facial feature points detection method | |
WO2021238548A1 (en) | Region recognition method, apparatus and device, and readable storage medium | |
WO2015000286A1 (en) | Three-dimensional interactive learning system and method based on augmented reality | |
CN113034652A (en) | Virtual image driving method, device, equipment and storage medium | |
CN104537705A (en) | Augmented reality based mobile platform three-dimensional biomolecule display system and method | |
CN111046734A (en) | Multi-modal fusion sight line estimation method based on expansion convolution | |
CN110941996A (en) | Target and track augmented reality method and system based on generation of countermeasure network | |
EP3185212B1 (en) | Dynamic particle filter parameterization | |
JP2016014954A (en) | Method for detecting finger shape, program thereof, storage medium of program thereof, and system for detecting finger shape | |
CN114641799A (en) | Object detection device, method and system | |
CN114333046A (en) | Dance action scoring method, device, equipment and storage medium | |
CN111709268A (en) | Human hand posture estimation method and device based on human hand structure guidance in depth image |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20180123 |
|
RJ01 | Rejection of invention patent application after publication |