CN110796018A - Hand motion recognition method based on depth image and color image - Google Patents

Hand motion recognition method based on depth image and color image Download PDF

Info

Publication number
CN110796018A
CN110796018A CN201910945063.7A CN201910945063A CN110796018A CN 110796018 A CN110796018 A CN 110796018A CN 201910945063 A CN201910945063 A CN 201910945063A CN 110796018 A CN110796018 A CN 110796018A
Authority
CN
China
Prior art keywords
detection
color
depth
iou
layer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910945063.7A
Other languages
Chinese (zh)
Other versions
CN110796018B (en
Inventor
刘玉婷
李公法
李蔚
田泉
蒋国璋
陶波
江都
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University of Science and Engineering WUSE
Original Assignee
Wuhan University of Science and Engineering WUSE
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University of Science and Engineering WUSE filed Critical Wuhan University of Science and Engineering WUSE
Priority to CN201910945063.7A priority Critical patent/CN110796018B/en
Publication of CN110796018A publication Critical patent/CN110796018A/en
Application granted granted Critical
Publication of CN110796018B publication Critical patent/CN110796018B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Abstract

The invention discloses a hand motion recognition method based on a depth image and a color image, which comprises the steps of utilizing 36 gestures in an ASL (active library language) gesture library as templates, obtaining gesture data through a Kinect sensor, and establishing a gesture database under a depth background and a color background; and (3) taking a regression-based target detection algorithm SSD as a research basis, and respectively carrying out transfer learning on the selected target detection model by utilizing a self-built gesture database based on color and depth backgrounds under a Tensorflow deep learning framework to obtain two types of network models capable of identifying and detecting hand movement under the depth and color backgrounds. The hand motion recognition detection network framework with the detection results fused under the depth and color backgrounds is utilized, the non-maximum suppression algorithm is improved, and the effectiveness of the proposed network framework on hand motion recognition detection is finally obtained. The invention avoids the problems of missed detection and false detection of the target, improves the gesture recognition rate and can realize one-hand recognition and two-hand recognition.

Description

Hand motion recognition method based on depth image and color image
Technical Field
The invention relates to the technical field of image processing and intelligent interaction, in particular to a hand motion recognition method based on a depth image and a color image.
Background
With the development of related disciplines of machine vision and artificial intelligence, human-computer interaction technology gradually starts to become an important research direction. The randomness and variability of hand morphology makes the current gesture recognition technology unable to replace traditional interactive devices in practical applications to achieve human-computer interaction, which means that the gesture recognition detection interaction technology and some other recognition technologies based on computer vision still need to be explored continuously. The depth data provides more shape information, sharp edges, and is robust to changes in lighting conditions than color data, which provides appearance and texture information and is sensitive to light changes. Therefore, the task of utilizing additional depth information to facilitate color information for joint hand motion recognition detection has become a focus of research.
It is still a key to hand motion recognition detection to propose a method that combines color and depth data in an optimal way.
Disclosure of Invention
In order to solve the technical problems, the invention provides a hand motion recognition method based on a depth image and a color image, which can read the color information and the depth information of a gesture in a complex scene and interact with a display by accurately recognizing the gesture.
The technical scheme adopted by the invention is as follows: a hand motion recognition method based on a depth image and a color image is characterized by comprising the following steps:
step 1: self-building a hand movement database;
36 gestures in the ASL gesture simulation library are read through kinect, and a depth hand motion database and a color hand motion database are established;
step 2: preprocessing hand data, and dividing the preprocessed data into a training set and a test set;
and step 3: training a deep hand motion detection model and a color hand motion detection model;
training a SSD _ Mobilene-based target detection model in a deep environment by using a deep training set, and training a SSD _ Mobilene-based target detection model in a color environment by using a color training set; obtaining two types of network models which can respectively identify and detect the hand movement under the depth and color backgrounds;
and 4, step 4: fusing detection results of two types of hand motion recognition detection network models under color and depth backgrounds by utilizing an SSD _ Mobilene double-flow network framework;
and 5: carrying out non-maximum improvement on the fused network model, introducing a non-maximum suppression algorithm of a proportional penalty coefficient, and testing the fused network framework by using the data concentrated in the test to obtain a hand motion high-precision identification model;
step 6: and identifying the hand motion by using a hand motion high-precision identification model.
The method has the advantages that under a tensoflow deep learning framework and an SSD _ Mobilene network model, the color and depth information of each level is extracted from a pair of convolutional neural networks to obtain the characteristic kernels of different layers, all the characteristic kernels are fused at the detection result, and the method is improved in a non-maximum inhibition value algorithm, so that the problems of missing detection and false detection of the target are avoided, and the gesture recognition rate is improved. And the gesture can be recognized in both a static picture and a dynamic video, so that not only can one-hand recognition be realized, but also two-hand recognition can be realized.
The color map can fully consider information such as color and texture of an extracted object, the depth map can extract depth information of the object, the depth information can avoid influence of factors such as illumination and environment, the two kinds of information are fully combined, and the convolutional neural network training model is used to enable gesture recognition to meet requirements of practical application.
Drawings
FIG. 1 is a flow chart of an embodiment of the present invention;
FIG. 2 is a schematic diagram of a network model for recognizing and detecting hand movements in a depth and color background, respectively, according to an embodiment of the present invention;
fig. 3 is a schematic diagram of a network framework in which detection results of hand motion recognition based on depth images and color images are fused according to an embodiment of the present invention.
Detailed Description
In order to facilitate the understanding and implementation of the present invention for those of ordinary skill in the art, the present invention is further described in detail with reference to the accompanying drawings and examples, it is to be understood that the embodiments described herein are merely illustrative and explanatory of the present invention and are not restrictive thereof.
Referring to fig. 1, the hand motion recognition method based on depth images and color images provided by the invention includes the following steps:
step 1: self-building a hand movement database;
establishing a depth hand motion database and a color hand motion database by using kinect to read 36 gestures in an ASL (American Sign language) gesture library;
step 2: preprocessing hand data, and dividing the preprocessed data into a training set and a test set;
in this embodiment, under the tensrflow deep learning framework, the color image data and the depth image data are respectively preprocessed by artificially labeling the color image and the depth image and performing format conversion.
And step 3: training a deep hand motion detection model and a color hand motion detection model;
training a SSD _ Mobilene-based target detection model in a deep environment by using a deep training set, and training a SSD _ Mobilene-based target detection model in a color environment by using a color training set; obtaining two types of network models which can respectively identify and detect the hand movement under the depth and color backgrounds;
the network model in this embodiment is used to identify and output gestures in the color background and the color video, and the depth background and the depth video, respectively, and the output result includes position information of the gesture in the image and a gesture category confidence.
And 4, step 4: fusing detection results of two types of hand motion recognition detection network models under color and depth backgrounds by utilizing an SSD _ Mobilene double-flow network framework;
referring to fig. 2 and 3, in this embodiment, based on a fusion model of an SSD _ mobilene dual-flow network framework, a color and a corresponding depth image are simultaneously input into two SSD _ mobilene networks having the same structure, and features are simultaneously extracted from six feature maps with different scales based on a color network channel and a depth network channel respectively for detection, and detection results of the color network channel and the depth network channel are output after non-maximum suppression is performed simultaneously;
Figure BDA0002223923610000031
wherein the jth neuron of the l layer of the convolutional network outputs
Figure BDA0002223923610000032
For upper layer output, fc(.) is the activation function of the convolutional layer, M is the set of selected input feature maps, wlWeight of layer I of the convolutional network, blFor the biasing of the network layer l,
Figure BDA0002223923610000033
a jth neuron output representing an ith channel of the ith layer;
when the l-th layer is a pooling layer, the output of the j-th neuron of the layer is
Figure BDA0002223923610000034
Figure BDA0002223923610000035
Wherein f isp(.) is a convolution layerP (.) is a pooling function;
when the l layer is a fully connected layer, the output of the jth neuron of the layer is
Figure BDA0002223923610000036
Figure BDA0002223923610000037
Wherein f isF(.) is the activation function of the fully connected layer, wFWeight of the full connection layer, bFBias for full connection, blIs the bias of the l layer of the network layer;
after extracting the characteristic values for a plurality of times, putting all the characteristic values into a detection layer for fusion, and then carrying out non-maximum suppression operation.
And 5: carrying out non-maximum improvement on the fused network model, introducing a non-maximum suppression algorithm of a proportional penalty coefficient, and testing the fused network framework by using the data concentrated in the test to obtain a hand motion high-precision identification model;
the method improves the non-maximum value inhibition operation, introduces a non-maximum value inhibition algorithm of a proportional penalty coefficient, endows a corresponding penalty coefficient to the prediction box according to the IoU value of the prediction box, reduces the confidence score of the prediction box by turns through the penalty coefficient, and removes the prediction box with lower confidence score through multiple iterations.
In the embodiment of this patent, the non-maximum suppression process is: the method comprises the steps of setting a prediction frame (set during image preprocessing) originally as B, setting a confidence score set corresponding to the prediction frame as S, setting a prediction frame set which is finally reserved for output as F, sequencing the prediction frames of the B from large to small according to the confidence score in the S, taking a first prediction frame as a suppression prediction frame, putting the first prediction frame into the prediction frame set F which is reserved for output, removing the prediction frame from the set B, calculating IoU values of the residual prediction frame and the first prediction frame, removing IoU prediction frames which are larger than a given threshold value T from the set B, and repeating the steps until the number of the prediction frames in the B is equal to 0. Iou is the detection accuracy.
Detection accuracy Iou:
Figure BDA0002223923610000041
wherein d represents a detection frame and p represents a preselection detection frame; IoU is the coincidence degree of the frame d predicted by the network framework and the frame p marked in the original picture, the calculation method is that the intersection of the detection result d and the marked sample p is compared with the union, namely the detection accuracy rate IoU;
introducing a non-maximum suppression algorithm of proportional penalty coefficients, endowing corresponding penalty coefficients to a prediction box according to the IoU value of the prediction box, reducing the confidence score of the prediction box by turns through the penalty coefficients and removing the prediction box with lower confidence score through multiple iterations, when the IoU values of a detection box and a preselected detection box are more than or equal to a given threshold value T, calculating the corresponding penalty coefficients α according to the IoU value, and when IoU is less than the given threshold value, the penalty coefficients are 1, and the proportional penalty coefficients are the prediction boxes B with the highest original confidence scores through the prediction box and the original confidence scoremaxIoU, and the penalty coefficient is calculated in the following specific calculation mode:
α=1 IoU(Bmax,Bj)<T;
α=1-lg(IoU(Bmax,Bj)+1) IoU(Bmax,Bj)≥T;
wherein IoU (B)max,Bj) The detection box is preset with IoU value of the box with the highest confidence score, and after a penalty coefficient is introduced, the result is IoU (B)max,Bj) If the result is less than T, the penalty coefficient is 1, and the confidence coefficient is the same as that of the detection frame corresponding to the initial NMS at the moment; IoU (B)max,Bj) ≧ T, at which time the corresponding detection box confidence of the NMS introducing the penalty factor changes from G (1-lg (IoU (B)max,Bj) +1)), G is the original confidence score of the corresponding detection box;
because the initial non-maximum value suppression algorithm directly removes the prediction frame with the value greater than the given threshold value T from the pre-selection detection frame IoU, the condition of removing the prediction frame is limited by introducing the penalty coefficient, so that the problems of missed detection and false detection of the target are avoided.
Step 6: and identifying the hand motion by using a hand motion high-precision identification model.
The hand motion recognition method based on the depth image and the color image solves the problem of how to combine color data and depth data in an optimal mode, can accurately read single-hand and double-hand gesture information under a complex environment, and realizes man-machine interaction recognition.
It should be understood that parts of the specification not set forth in detail are prior art; the above description of the preferred embodiments is intended to be illustrative, and not to be construed as limiting the scope of the invention, which is defined by the appended claims, and all changes and modifications that fall within the metes and bounds of the claims, or equivalences of such metes and bounds are therefore intended to be embraced by the appended claims.

Claims (6)

1. A hand motion recognition method based on a depth image and a color image is characterized by comprising the following steps:
step 1: self-building a hand movement database;
36 gestures in the ASL gesture simulation library are read through kinect, and a depth hand motion database and a color hand motion database are established;
step 2: preprocessing hand data, and dividing the preprocessed data into a training set and a test set;
and step 3: training a deep hand motion detection model and a color hand motion detection model;
training a SSD _ Mobilene-based target detection model in a deep environment by using a deep training set, and training a SSD _ Mobilene-based target detection model in a color environment by using a color training set; obtaining two types of network models which can respectively identify and detect the hand movement under the depth and color backgrounds;
and 4, step 4: fusing detection results of two types of hand motion recognition detection network models under color and depth backgrounds by utilizing an SSD _ Mobilene double-flow network framework;
and 5: carrying out non-maximum improvement on the fused network model, introducing a non-maximum suppression algorithm of a proportional penalty coefficient, and testing the fused network framework by using the data concentrated in the test to obtain a hand motion high-precision identification model;
step 6: and identifying the hand motion by using a hand motion high-precision identification model.
2. The hand motion recognition method based on the depth image and the color image as claimed in claim 1, wherein: in step 2, under a Tensorflow deep learning framework, the color image data and the depth image data are respectively preprocessed in a mode of manually labeling the color image and the depth image and performing format conversion.
3. The hand motion recognition method based on the depth image and the color image as claimed in claim 1, wherein: in step 3, the network model can be used for recognizing and outputting gestures in a color background, a color video, a depth background and a depth video respectively, and output results comprise position information of the gestures in the image and gesture category confidence.
4. The hand motion recognition method based on the depth image and the color image as claimed in claim 1, wherein: step 4, based on a fusion model of an SSD _ Mobilene dual-flow network framework, inputting colors and corresponding depth images into two SSD _ Mobilene networks with the same structure at the same time, simultaneously and respectively extracting features from six feature maps with different scales based on a color network channel and a depth network channel for detection, and outputting detection results of the color network channel and the depth network channel after non-maximum suppression;
wherein the jth neuron of the l layer of the convolutional network outputs
Figure FDA0002223923600000022
For upper layer output, fc(.) is the activation function of the convolutional layer, M is the set of selected input feature maps, wlWeight of layer I of the convolutional network, blFor the biasing of the network layer l,
Figure FDA0002223923600000024
a jth neuron output representing an ith channel of the ith layer;
when the l-th layer is a pooling layer, the output of the j-th neuron of the layer is
Figure FDA0002223923600000025
Figure FDA0002223923600000026
Wherein f isp(.) is the activation function of the convolutional layer, p (.) is the pooling function;
when the l layer is a fully connected layer, the output of the jth neuron of the layer is
Figure FDA0002223923600000027
Figure FDA0002223923600000028
Wherein f isF(.) is the activation function of the fully connected layer, wFWeight of the full connection layer, bFBias for full connection, blIs the bias of the l layer of the network layer;
after extracting the characteristic values for a plurality of times, putting all the characteristic values into a detection layer for fusion, and then carrying out non-maximum suppression operation.
5. The hand motion recognition method based on the depth image and the color image as claimed in claim 1, wherein: in step 5, the non-maximum suppression process is as follows: setting a prediction frame as B during image preprocessing, setting a confidence score set corresponding to the prediction frame as S, and finally keeping the output prediction frame set as F, firstly sequencing the prediction frames of B from large to small according to the confidence score in S, taking a first prediction frame as an inhibition prediction frame, putting the inhibition prediction frame into the output-keeping prediction frame set F, then removing the prediction frame from the set B, then calculating IoU values of the residual prediction frame and the first prediction frame, removing IoU prediction frames larger than a given threshold value T from the set B, and repeating the steps until the number of the prediction frames in B is equal to 0;
wherein:
Figure FDA0002223923600000029
wherein d represents a detection frame and p represents a preselection detection frame; IoU is the coincidence degree of the frame d predicted by the network framework and the frame p marked in the original picture, the calculation method is that the intersection of the detection result d and the marked sample p is compared with the union, namely the detection accuracy rate IoU;
introducing a non-maximum suppression algorithm of proportional penalty coefficients, endowing corresponding penalty coefficients to a prediction box according to the IoU value of the prediction box, reducing the confidence score of the prediction box by turns through the penalty coefficients and removing the prediction box with lower confidence score through multiple iterations, when the IoU values of a detection box and a preselected detection box are more than or equal to a given threshold value T, calculating the corresponding penalty coefficients α according to the IoU value, and when IoU is less than the given threshold value, the penalty coefficients are 1, and the proportional penalty coefficients are the prediction boxes B with the highest original confidence scores through the prediction box and the original confidence scoremaxIoU, and the penalty coefficient is calculated in the following specific calculation mode:
α=1 IoU(Bmax,Bj)<T;
α=1-lg(IoU(Bmax,Bj)+1) IoU(Bmax,Bj)≥T;
wherein IoU (B)max,Bj) Is a detection framePresetting IoU value of the box with highest confidence score, and adding penalty factor when IoU (B)max,Bj) If the result is less than T, the penalty coefficient is 1, and the confidence coefficient is the same as that of the detection frame corresponding to the initial NMS at the moment; IoU (B)max,Bj) ≧ T, at which time the corresponding detection box confidence of the NMS introducing the penalty factor changes from G (1-lg (IoU (B)max,Bj) +1)), G is the original confidence score of the corresponding detection box;
because the initial non-maximum value suppression algorithm directly removes the prediction frame with the value greater than the given threshold value T from the pre-selection detection frame IoU, the condition of removing the prediction frame is limited by introducing the penalty coefficient, so that the problems of missed detection and false detection of the target are avoided.
6. The method of hand motion recognition based on depth images and color images according to any one of claims 1 to 5, wherein: and 5, inputting the paired depth and color hand test sets into the fused network framework for testing, and performing comparative analysis on the paired depth and color hand test sets and the test results of the hand recognition detection model in the single mode to obtain more effective experimental results.
CN201910945063.7A 2019-09-30 2019-09-30 Hand motion recognition method based on depth image and color image Active CN110796018B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910945063.7A CN110796018B (en) 2019-09-30 2019-09-30 Hand motion recognition method based on depth image and color image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910945063.7A CN110796018B (en) 2019-09-30 2019-09-30 Hand motion recognition method based on depth image and color image

Publications (2)

Publication Number Publication Date
CN110796018A true CN110796018A (en) 2020-02-14
CN110796018B CN110796018B (en) 2023-04-28

Family

ID=69438923

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910945063.7A Active CN110796018B (en) 2019-09-30 2019-09-30 Hand motion recognition method based on depth image and color image

Country Status (1)

Country Link
CN (1) CN110796018B (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111339985A (en) * 2020-03-06 2020-06-26 南京理工大学 Gesture detection method based on mixed convolution
CN111652110A (en) * 2020-05-28 2020-09-11 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium
CN111860448A (en) * 2020-07-30 2020-10-30 北京华捷艾米科技有限公司 Hand washing action recognition method and system
CN111968087A (en) * 2020-08-13 2020-11-20 中国农业科学院农业信息研究所 Plant disease area detection method
CN112784810A (en) * 2021-02-08 2021-05-11 风变科技(深圳)有限公司 Gesture recognition method and device, computer equipment and storage medium
CN113221744A (en) * 2021-05-12 2021-08-06 天津大学 Monocular image 3D object detection method based on deep learning
CN115035119A (en) * 2022-08-12 2022-09-09 山东省计算中心(国家超级计算济南中心) Glass bottle bottom flaw image detection and removal device, system and method
CN115138059A (en) * 2022-09-06 2022-10-04 南京市觉醒智能装备有限公司 Pull-up standard counting method, pull-up standard counting system and storage medium of pull-up standard counting system

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140079297A1 (en) * 2012-09-17 2014-03-20 Saied Tadayon Application of Z-Webs and Z-factors to Analytics, Search Engine, Learning, Recognition, Natural Language, and Other Utilities
CN107038424A (en) * 2017-04-20 2017-08-11 华中师范大学 A kind of gesture identification method
CN107808143A (en) * 2017-11-10 2018-03-16 西安电子科技大学 Dynamic gesture identification method based on computer vision
CN107808131A (en) * 2017-10-23 2018-03-16 华南理工大学 Dynamic gesture identification method based on binary channel depth convolutional neural networks
CN108875595A (en) * 2018-05-29 2018-11-23 重庆大学 A kind of Driving Scene object detection method merged based on deep learning and multilayer feature
CN109409246A (en) * 2018-09-30 2019-03-01 中国地质大学(武汉) Acceleration robust features bimodal gesture based on sparse coding is intended to understanding method

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140079297A1 (en) * 2012-09-17 2014-03-20 Saied Tadayon Application of Z-Webs and Z-factors to Analytics, Search Engine, Learning, Recognition, Natural Language, and Other Utilities
CN107038424A (en) * 2017-04-20 2017-08-11 华中师范大学 A kind of gesture identification method
CN107808131A (en) * 2017-10-23 2018-03-16 华南理工大学 Dynamic gesture identification method based on binary channel depth convolutional neural networks
CN107808143A (en) * 2017-11-10 2018-03-16 西安电子科技大学 Dynamic gesture identification method based on computer vision
CN108875595A (en) * 2018-05-29 2018-11-23 重庆大学 A kind of Driving Scene object detection method merged based on deep learning and multilayer feature
CN109409246A (en) * 2018-09-30 2019-03-01 中国地质大学(武汉) Acceleration robust features bimodal gesture based on sparse coding is intended to understanding method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
刘壮等: "双通道Faster R-CNN在RGB-D手部检测中的应用" *

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111339985A (en) * 2020-03-06 2020-06-26 南京理工大学 Gesture detection method based on mixed convolution
CN111652110A (en) * 2020-05-28 2020-09-11 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium
CN111860448A (en) * 2020-07-30 2020-10-30 北京华捷艾米科技有限公司 Hand washing action recognition method and system
CN111968087A (en) * 2020-08-13 2020-11-20 中国农业科学院农业信息研究所 Plant disease area detection method
CN111968087B (en) * 2020-08-13 2023-11-07 中国农业科学院农业信息研究所 Plant disease area detection method
CN112784810A (en) * 2021-02-08 2021-05-11 风变科技(深圳)有限公司 Gesture recognition method and device, computer equipment and storage medium
CN113221744A (en) * 2021-05-12 2021-08-06 天津大学 Monocular image 3D object detection method based on deep learning
CN113221744B (en) * 2021-05-12 2022-10-04 天津大学 Monocular image 3D object detection method based on deep learning
CN115035119A (en) * 2022-08-12 2022-09-09 山东省计算中心(国家超级计算济南中心) Glass bottle bottom flaw image detection and removal device, system and method
CN115138059A (en) * 2022-09-06 2022-10-04 南京市觉醒智能装备有限公司 Pull-up standard counting method, pull-up standard counting system and storage medium of pull-up standard counting system
CN115138059B (en) * 2022-09-06 2022-12-02 南京市觉醒智能装备有限公司 Pull-up standard counting method, pull-up standard counting system and storage medium of pull-up standard counting system

Also Published As

Publication number Publication date
CN110796018B (en) 2023-04-28

Similar Documents

Publication Publication Date Title
CN110796018A (en) Hand motion recognition method based on depth image and color image
CN109299274B (en) Natural scene text detection method based on full convolution neural network
CN109325454B (en) Static gesture real-time recognition method based on YOLOv3
WO2020182121A1 (en) Expression recognition method and related device
US10445602B2 (en) Apparatus and method for recognizing traffic signs
CN113158862B (en) Multitasking-based lightweight real-time face detection method
CN111401293B (en) Gesture recognition method based on Head lightweight Mask scanning R-CNN
CN110555420B (en) Fusion model network and method based on pedestrian regional feature extraction and re-identification
CN107247952B (en) Deep supervision-based visual saliency detection method for cyclic convolution neural network
CN111259940A (en) Target detection method based on space attention map
CN108073851B (en) Grabbing gesture recognition method and device and electronic equipment
CN110674741A (en) Machine vision gesture recognition method based on dual-channel feature fusion
CN110705412A (en) Video target detection method based on motion history image
CN115328319B (en) Intelligent control method and device based on light-weight gesture recognition
CN112163508A (en) Character recognition method and system based on real scene and OCR terminal
CN111368634B (en) Human head detection method, system and storage medium based on neural network
CN113297956A (en) Gesture recognition method and system based on vision
CN114549557A (en) Portrait segmentation network training method, device, equipment and medium
CN114445853A (en) Visual gesture recognition system recognition method
CN110472673B (en) Parameter adjustment method, fundus image processing device, fundus image processing medium and fundus image processing apparatus
CN113487610B (en) Herpes image recognition method and device, computer equipment and storage medium
Sun et al. IRDCLNet: Instance segmentation of ship images based on interference reduction and dynamic contour learning in foggy scenes
CN115223239A (en) Gesture recognition method and system, computer equipment and readable storage medium
CN111368637B (en) Transfer robot target identification method based on multi-mask convolutional neural network
CN111078008B (en) Control method of early education robot

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
EE01 Entry into force of recordation of patent licensing contract
EE01 Entry into force of recordation of patent licensing contract

Application publication date: 20200214

Assignee: JINGMEN WUSAN MACHINERY EQUIPMENT MANUFACTURING Co.,Ltd.

Assignor: WUHAN University OF SCIENCE AND TECHNOLOGY

Contract record no.: X2023420000176

Denomination of invention: A Method for Hand Motion Recognition Based on Depth and Color Images

Granted publication date: 20230428

License type: Common License

Record date: 20230614