CN112749664A - Gesture recognition method, device, equipment, system and storage medium - Google Patents

Gesture recognition method, device, equipment, system and storage medium Download PDF

Info

Publication number
CN112749664A
CN112749664A CN202110053294.4A CN202110053294A CN112749664A CN 112749664 A CN112749664 A CN 112749664A CN 202110053294 A CN202110053294 A CN 202110053294A CN 112749664 A CN112749664 A CN 112749664A
Authority
CN
China
Prior art keywords
image
hand
gesture recognition
carrying
recognition
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110053294.4A
Other languages
Chinese (zh)
Inventor
张坡
阎汉生
朱腾
周宇
黎少
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong College of Industry and Commerce
Original Assignee
Guangdong College of Industry and Commerce
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong College of Industry and Commerce filed Critical Guangdong College of Industry and Commerce
Priority to CN202110053294.4A priority Critical patent/CN112749664A/en
Publication of CN112749664A publication Critical patent/CN112749664A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/155Segmentation; Edge detection involving morphological operators
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/01Indexing scheme relating to G06F3/01
    • G06F2203/012Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention relates to a gesture recognition method, a gesture recognition device, gesture recognition equipment, a gesture recognition system and a storage medium. The method comprises the following steps: acquiring a hand image shot by a binocular camera system; carrying out distortion correction on the hand image to obtain a corrected image; carrying out color space processing on the corrected image to obtain an HSV image; carrying out threshold segmentation on the HSV image to obtain a segmented image; performing frame selection on the hand in the separation image to obtain a hand outline; and classifying and identifying the hand images by using a pre-trained classifier according to the hand contours to obtain an identification result. According to the method, the accuracy of gesture recognition is greatly improved through distortion correction and classifier recognition.

Description

Gesture recognition method, device, equipment, system and storage medium
Technical Field
The invention relates to the technical field of gesture recognition, in particular to a gesture recognition method, a gesture recognition device, gesture recognition equipment, a gesture recognition system and a storage medium.
Background
With the continuous development of Virtual Reality (VR) technology, interactive gesture recognition technology has come to work. In order to improve the sensing effect, the gesture recognition technology generally adopts an infrared LED gray camera mode to realize accurate recognition.
At present, generally, when an interactive gesture recognition technology of an infrared LED gray camera type is used for recognition, the coverage angle of the camera is concentrated in the range of 40-140 degrees of a forward horizontal plane of a corresponding lens, gesture interactive tracking at a higher angle cannot be realized, and if the camera is changed into a wide-angle camera, the camera has a distortion effect, and gesture segmentation and recognition cannot be accurately carried out.
Disclosure of Invention
In view of the above, the present invention provides a gesture recognition method, device, apparatus, system and storage medium to overcome the disadvantages of the prior art.
In order to achieve the purpose, the invention adopts the following technical scheme:
a gesture recognition method, comprising:
acquiring a hand image shot by a binocular camera system;
carrying out distortion correction on the hand image to obtain a corrected image;
carrying out color space processing on the corrected image to obtain an HSV image;
carrying out threshold segmentation on the HSV image to obtain a segmented image;
performing frame selection on the hand in the separation image to obtain a hand outline;
and carrying out classification and identification in the hand image by utilizing a pre-trained classifier according to the hand contour to obtain an identification result.
Optionally, the method further includes:
and performing morphological closed operation on the image obtained after the threshold segmentation of the HSV image.
Optionally, the performing distortion correction on the hand image to obtain a corrected image includes:
calibrating the binocular camera system by adopting a black and white chessboard calibration plate, and solving internal parameters and external parameters of a camera in the binocular camera system;
and correcting by using a preset AI correction algorithm according to the internal parameters and the external parameters to obtain the corrected image.
Optionally, the classifying and recognizing the hand image by using a pre-trained classifier according to the hand contour to obtain a recognition result includes:
intercepting the hand contour in the hand image to obtain an intercepted image;
and carrying out image recognition on the intercepted image by using the classifier to obtain the recognition result.
Optionally, the process of training the classifier includes:
acquiring a set number of hand unfolding images and hand fist making images from a preset experiment platform;
setting labels for the hand unfolding image and the hand fist making image in a self-defined mode respectively;
training the classifier in conjunction with the label, the hand-unfolded image, and the hand-clenched image.
A gesture recognition apparatus comprising:
the hand image acquisition module is used for acquiring a hand image shot by the binocular camera system;
the distortion correction module is used for carrying out distortion correction on the hand image to obtain a corrected image;
the color processing module is used for carrying out color space processing on the corrected image to obtain an HSV image;
the threshold segmentation module is used for carrying out threshold segmentation on the HSV image to obtain a segmented image;
the framing module is used for framing the hand in the separation image to obtain a hand outline;
and the recognition module is used for carrying out classification recognition in the hand image by using a pre-trained classifier according to the hand contour to obtain a recognition result.
Optionally, the method further includes:
and the closed operation module is used for performing morphological closed operation on the image obtained after the threshold segmentation is performed on the HSV image.
A gesture recognition device, comprising:
a processor, and a memory coupled to the processor;
the memory is used for storing a computer program, and the computer program is at least used for executing the gesture recognition method;
the processor is used for calling and executing the computer program in the memory.
A gesture recognition system, comprising:
binocular camera system, and with binocular camera system communication connection as above gesture recognition equipment.
A storage medium storing a computer program which, when executed by a processor, implements the steps of the gesture recognition method as described above.
The technical scheme provided by the application can comprise the following beneficial effects:
the application discloses a gesture recognition method, which comprises the following steps: acquiring a hand image shot by a binocular camera system; carrying out distortion correction on the hand image to obtain a corrected image; carrying out color space processing on the corrected image to obtain an HSV image; carrying out threshold segmentation on the HSV image to obtain a segmented image; performing frame selection on the hand in the separation image to obtain a hand outline; and carrying out classification and identification in the hand image by utilizing a pre-trained classifier according to the hand contour to obtain an identification result. According to the method, the binocular camera system is used as input signal change judgment equipment, and on the basis of guaranteeing that the coverage angle is captured in a traditional mode and enlarged, the novel orthodontic algorithm is used for completing orthodontics so as to reduce the influence on subsequent VR gesture recognition, and therefore the accuracy of gesture recognition is greatly improved.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to the drawings without creative efforts.
FIG. 1 is a flow chart of a gesture recognition method according to an embodiment of the present invention;
FIG. 2 is a block diagram of a gesture recognition apparatus according to an embodiment of the present invention;
FIG. 3 is a block diagram of a gesture recognition apparatus according to an embodiment of the present invention;
fig. 4 is a block diagram of a gesture recognition system according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the technical solutions of the present invention will be described in detail below. It is to be understood that the described embodiments are merely exemplary of the invention, and not restrictive of the full scope of the invention. All other embodiments, which can be derived by a person skilled in the art from the examples given herein without any inventive step, are within the scope of the present invention.
Fig. 1 is a flowchart of a gesture recognition method according to an embodiment of the present invention. Referring to fig. 1, a gesture recognition method includes:
step 101: and acquiring a hand image shot by the binocular camera system. The binocular stereo camera system is mainly used in the prior art, the coverage angle of the camera is in the range of 40-140 degrees, the binocular stereo camera system can shoot videos in a wider range and can shoot images in the range of 360 degrees, and therefore the situation that omission exists in gesture interactive tracking due to limitation of shooting angles is avoided.
Step 102: and carrying out distortion correction on the hand image to obtain a corrected image. Video or image that the camera was shot all can have the condition of distortion, consequently before carrying out gesture recognition, need carry out the distortion correction processing with the hand image earlier, the specific process of distortion correction processing in this application is as follows:
and calibrating the binocular camera system by adopting a black and white chessboard calibration plate, and solving the internal parameters and the external parameters of the camera in the binocular camera system. The internal parameters include focal length, lens radial distortion coefficient and the like, and the external parameters include camera position and direction, scanning angle, inclination angle and the like. And then correcting by using a preset AI correction algorithm according to the internal parameters and the external parameters to obtain the corrected image.
In order to solve the 4 internal parameters and 6 external parameters of the two monocular camera parameters, it is assumed that the black-and-white chessboard calibration plate used in the present application has N angular points, then the calibration plate is placed through M different poses, and the world coordinate is set as the first angular point of the chessboard calibration plate, so the restriction conditions for obtaining the images of the chessboard at different poses are as follows:
2NM≥6M+4;
although theoretically, N is 5, and M is 1, the requirement of the corner points can be met, all targets representing a plane projection view field only need 4 points, so that no matter how many corner points exist in a plane, only 4 pieces of corner point information are available, but in consideration of noise and stability, more corner points and pose requirements exist, and more calibration plate images with different poses need to be collected during calibration.
The coordinates of the checkerboard corner points and the coordinates of the checkerboard corner point image can determine key points, and a corresponding unit matrix is solved for each checkerboard image through least squares. And finally obtaining the internal parameters and the external parameters of the camera.
Step 103: and carrying out color space processing on the corrected image to obtain an HSV image.
Step 104: and carrying out threshold segmentation on the HSV image to obtain a segmented image. In the application, after color space conversion is carried out on the corrected image, self-adaptive threshold segmentation is carried out on the corrected image, and the hand and the background are separated to obtain a segmented image. Meanwhile, in order to obtain a better segmentation effect, morphological closed operation is carried out on the image segmented by the threshold value.
Step 105: and performing frame selection on the hand in the separated image to obtain a hand outline. Specifically, the hand frame is selected by detecting a rectangle and a convex hull which are externally connected with the outline.
Step 106: and carrying out classification and identification in the hand image by utilizing a pre-trained classifier according to the hand contour to obtain an identification result. The classifier adopted in the present application is an SVM classifier, and it should be noted that the type selection of the classifier is not fixed and can be determined according to the actual situation. And finally, intercepting the frame selection part in the original image and carrying out classification and identification by utilizing a trained SVM classifier. Specifically, the hand contour is firstly intercepted from the hand image, and then the intercepted image is identified by an SVM classifier to obtain an identification result.
The SVM classifier needs to be trained before an experiment, images of the unfolded hands and images of the fist made by the hands are obtained through an experiment platform, and tags are marked, for example, the image tag of the unfolded hands is 1, and the image tag of the fist made by the hands is 0. Then training is carried out through a large number of images with labels, and an SVM classifier is obtained.
After the SVM classifier is trained, the classification recognition result of the classifier needs to be verified. The application discloses a process for training an SVM classifier by a self-learning prediction optimization updating algorithm, wherein the optimization algorithm is defined as follows:
Figure BDA0002899773510000061
wherein the content of the first and second substances,
Figure BDA0002899773510000062
Figure BDA0002899773510000063
in the formula: c is a constant term, m is the number of samples, and n is a characteristic number; thetaTA vector of features; (x)(i),y(i)) Is the ith sample; h isθ(x) And (4) corresponding to a self-optimization model for the vector machine.
In the application, the classification object of the linear SVM classifier is an image with a gesture, and the classification result is 0 or 1, so that a calibrated picture can be used in a training sample of the SVM classifier, and data predicted by key points are not needed. However, the data + images of the keypoint predictions may be used as samples for training the SVM classifier.
If the data and the images predicted by the key points are used as samples to train the SVM classifier, the processing can be carried out according to the following gesture key point prediction algorithm and key point calibration, specifically as follows:
the probability that the predicted key point position of a certain point P in the graph is within a distance threshold value sigma of the real position of the predicted key point position is set, and the probability is defined as:
Figure BDA0002899773510000064
in the formula:
Figure BDA0002899773510000065
for detector d0The probability of correct keypoints at point P, σ represents the threshold, T represents the test set,
Figure BDA0002899773510000066
represents the predicted location of the p-th keypoint,
Figure BDA0002899773510000067
represents the actual position of the p-th key point, and delta is a preset value.
By setting different internal parameters n and the number V of cameras with different viewing angles, the probability of correct key points is represented by a correct number TP of correct cases and a wrong number FP of wrong cases. And meanwhile, calculating a correct recognition rate FDR, which is defined as:
Figure BDA0002899773510000071
meanwhile, the hand key point detector used in the key point calibration algorithm in the application is a convolution attitude machine heterogeneous optimization algorithm. The algorithm applies a key point detection algorithm based on deep learning to human body posture analysis, takes an image with a human body posture as input data, and outputs a confidence distribution map of human body posture characteristics through a set convolutional neural network.
According to the embodiment, the traditional monocular wide-angle camera is combined into the binocular stereo camera system, the binocular stereo camera system is used as input signal change judging equipment, a novel orthodontic algorithm is used on the basis of guaranteeing to enlarge the traditional mode and capture the coverage angle, and the influence of orthodontics on subsequent VR gesture recognition is reduced, and the binocular stereo camera system is used simultaneously. The AI correction algorithm is used as a judgment core, and the characteristics of low delay and accurate capture are met.
Corresponding to the gesture recognition method provided by the embodiment of the invention, the embodiment of the invention also provides a gesture recognition device. Please see the examples below.
Fig. 2 is a block diagram of a gesture recognition apparatus according to an embodiment of the present invention. Referring to fig. 2, a gesture recognition apparatus includes:
and a hand image acquisition module 201, configured to acquire a hand image captured by the binocular imaging system.
And the distortion correction module 202 is configured to perform distortion correction on the hand image to obtain a corrected image.
And the color processing module 203 is configured to perform color space processing on the corrected image to obtain an HSV image.
And the threshold segmentation module 204 is configured to perform threshold segmentation on the HSV image to obtain a segmented image.
And the frame selection module 205 is configured to select a hand in the separation image to obtain a hand contour.
And the recognition module 206 is configured to perform classification recognition on the hand image by using a pre-trained classifier according to the hand contour to obtain a recognition result.
On this basis, the device in this application still includes: and the closed operation module is used for performing morphological closed operation on the image obtained after the threshold segmentation is performed on the HSV image.
Wherein the distortion correction module 202 is specifically configured to: calibrating the binocular camera system by adopting a black and white chessboard calibration plate, and solving internal parameters and external parameters of a camera in the binocular camera system; and correcting by using a preset AI correction algorithm according to the internal parameters and the external parameters to obtain the corrected image.
The identification module 206 is specifically configured to: intercepting the hand contour in the hand image to obtain an intercepted image; and carrying out image recognition on the intercepted image by using the classifier to obtain the recognition result.
Above-mentioned embodiment uses novel just abnormal algorithm and is guaranteeing to enlarge traditional mode and catch the coverage angle basis on, accomplishes just abnormal in order to reduce to subsequent VR gesture recognition influence, uses simultaneously. The AI correction algorithm is used as a judgment core, and the characteristics of low delay and accurate capture are met.
In order to more clearly introduce a hardware system for implementing the embodiment of the present invention, a gesture recognition apparatus and a system are also provided in the embodiment of the present invention, corresponding to the gesture recognition method provided in the embodiment of the present invention. Please see the examples below.
Fig. 3 is a block diagram of a gesture recognition apparatus according to an embodiment of the present invention. Referring to fig. 3, a gesture recognition apparatus includes:
a processor 301, and a memory 302 connected to the processor 301;
the memory 302 is used for storing a computer program at least for executing the gesture recognition method;
the processor 301 is used for calling and executing the computer program in the memory 302.
Meanwhile, on the basis, the application also discloses a system, and fig. 4 is a structural diagram of the gesture recognition system provided by the embodiment of the invention. Referring to fig. 4, a gesture recognition system includes:
a binocular camera system 401, and a gesture recognition device 402 as described above in communication with the binocular camera system 401.
Meanwhile, the application also discloses a storage medium, wherein the storage medium stores a computer program, and when the computer program is executed by a processor, the steps in the gesture recognition method are realized.
According to the embodiment, the binocular stereo camera system is adopted, the shooting range of the camera is enlarged, and orthodontic algorithm is used for finishing orthodontic to reduce the influence on subsequent VR gesture recognition and simultaneously on the basis of ensuring that the coverage angle is captured in a traditional manner. The gesture recognition precision is greatly improved.
It is understood that the same or similar parts in the above embodiments may be mutually referred to, and the same or similar parts in other embodiments may be referred to for the content which is not described in detail in some embodiments.
It should be noted that the terms "first," "second," and the like in the description of the present invention are used for descriptive purposes only and are not to be construed as indicating or implying relative importance. Further, in the description of the present invention, the meaning of "a plurality" means at least two unless otherwise specified.
Any process or method descriptions in flow charts or otherwise described herein may be understood as representing modules, segments, or portions of code which include one or more executable instructions for implementing specific logical functions or steps of the process, and alternate implementations are included within the scope of the preferred embodiment of the present invention in which functions may be executed out of order from that shown or discussed, including substantially concurrently or in reverse order, depending on the functionality involved, as would be understood by those reasonably skilled in the art of the present invention.
It should be understood that portions of the present invention may be implemented in hardware, software, firmware, or a combination thereof. In the above embodiments, the various steps or methods may be implemented in software or firmware stored in memory and executed by a suitable instruction execution system. For example, if implemented in hardware, as in another embodiment, any one or combination of the following techniques, which are known in the art, may be used: a discrete logic circuit having a logic gate circuit for implementing a logic function on a data signal, an application specific integrated circuit having an appropriate combinational logic gate circuit, a Programmable Gate Array (PGA), a Field Programmable Gate Array (FPGA), or the like.
It will be understood by those skilled in the art that all or part of the steps carried by the method for implementing the above embodiments may be implemented by hardware related to instructions of a program, which may be stored in a computer readable storage medium, and when the program is executed, the program includes one or a combination of the steps of the method embodiments.
In addition, functional units in the embodiments of the present invention may be integrated into one processing module, or each unit may exist alone physically, or two or more units are integrated into one module. The integrated module can be realized in a hardware mode, and can also be realized in a software functional module mode. The integrated module, if implemented in the form of a software functional module and sold or used as a stand-alone product, may also be stored in a computer readable storage medium.
The storage medium mentioned above may be a read-only memory, a magnetic or optical disk, etc.
In the description herein, references to the description of the term "one embodiment," "some embodiments," "an example," "a specific example," or "some examples," etc., mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above do not necessarily refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
Although embodiments of the present invention have been shown and described above, it is understood that the above embodiments are exemplary and should not be construed as limiting the present invention, and that variations, modifications, substitutions and alterations can be made to the above embodiments by those of ordinary skill in the art within the scope of the present invention.

Claims (10)

1. A gesture recognition method, comprising:
acquiring a hand image shot by a binocular camera system;
carrying out distortion correction on the hand image to obtain a corrected image;
carrying out color space processing on the corrected image to obtain an HSV image;
carrying out threshold segmentation on the HSV image to obtain a segmented image;
performing frame selection on the hand in the separation image to obtain a hand outline;
and carrying out classification and identification in the hand image by utilizing a pre-trained classifier according to the hand contour to obtain an identification result.
2. The method of claim 1, further comprising:
and performing morphological closed operation on the image obtained after the threshold segmentation of the HSV image.
3. The method of claim 1, wherein said distortion correcting said hand image to obtain a corrected image comprises:
calibrating the binocular camera system by adopting a black and white chessboard calibration plate, and solving internal parameters and external parameters of a camera in the binocular camera system;
and correcting by using a preset AI correction algorithm according to the internal parameters and the external parameters to obtain the corrected image.
4. The method of claim 1, wherein the performing classification recognition on the hand image by using a pre-trained classifier according to the hand contour to obtain a recognition result comprises:
intercepting the hand contour in the hand image to obtain an intercepted image;
and carrying out image recognition on the intercepted image by using the classifier to obtain the recognition result.
5. The method of claim 1, wherein training the classifier comprises:
acquiring a set number of hand unfolding images and hand fist making images from a preset experiment platform;
setting labels for the hand unfolding image and the hand fist making image in a self-defined mode respectively;
training the classifier in conjunction with the label, the hand-unfolded image, and the hand-clenched image.
6. A gesture recognition apparatus, comprising:
the hand image acquisition module is used for acquiring a hand image shot by the binocular camera system;
the distortion correction module is used for carrying out distortion correction on the hand image to obtain a corrected image;
the color processing module is used for carrying out color space processing on the corrected image to obtain an HSV image;
the threshold segmentation module is used for carrying out threshold segmentation on the HSV image to obtain a segmented image;
the framing module is used for framing the hand in the separation image to obtain a hand outline;
and the recognition module is used for carrying out classification recognition in the hand image by using a pre-trained classifier according to the hand contour to obtain a recognition result.
7. The apparatus of claim 6, further comprising:
and the closed operation module is used for performing morphological closed operation on the image obtained after the threshold segmentation is performed on the HSV image.
8. A gesture recognition device, comprising:
a processor, and a memory coupled to the processor;
the memory for storing a computer program for at least performing the gesture recognition method of any one of claims 1-5;
the processor is used for calling and executing the computer program in the memory.
9. A gesture recognition system, comprising:
a binocular camera system, and the gesture recognition apparatus of claim 8 communicatively coupled to the binocular camera system.
10. A storage medium, characterized in that the storage medium stores a computer program which, when executed by a processor, performs the steps of the gesture recognition method according to any one of claims 1-5.
CN202110053294.4A 2021-01-15 2021-01-15 Gesture recognition method, device, equipment, system and storage medium Pending CN112749664A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110053294.4A CN112749664A (en) 2021-01-15 2021-01-15 Gesture recognition method, device, equipment, system and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110053294.4A CN112749664A (en) 2021-01-15 2021-01-15 Gesture recognition method, device, equipment, system and storage medium

Publications (1)

Publication Number Publication Date
CN112749664A true CN112749664A (en) 2021-05-04

Family

ID=75652127

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110053294.4A Pending CN112749664A (en) 2021-01-15 2021-01-15 Gesture recognition method, device, equipment, system and storage medium

Country Status (1)

Country Link
CN (1) CN112749664A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117373114A (en) * 2023-09-19 2024-01-09 海智合芯科技(深圳)有限公司 Gesture recognition method, system, device and storage medium based on follow-up image
CN117392759A (en) * 2023-12-11 2024-01-12 成都航空职业技术学院 Action recognition method based on AR teaching aid

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160171340A1 (en) * 2014-12-11 2016-06-16 Intel Corporation Labeling component parts of objects and detecting component properties in imaging data
CN107813310A (en) * 2017-11-22 2018-03-20 浙江优迈德智能装备有限公司 One kind is based on the more gesture robot control methods of binocular vision
CN109961484A (en) * 2017-12-22 2019-07-02 比亚迪股份有限公司 Camera calibration method, device and vehicle
CN111435429A (en) * 2019-01-15 2020-07-21 北京伟景智能科技有限公司 Gesture recognition method and system based on binocular stereo data dynamic cognition

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160171340A1 (en) * 2014-12-11 2016-06-16 Intel Corporation Labeling component parts of objects and detecting component properties in imaging data
CN107813310A (en) * 2017-11-22 2018-03-20 浙江优迈德智能装备有限公司 One kind is based on the more gesture robot control methods of binocular vision
CN109961484A (en) * 2017-12-22 2019-07-02 比亚迪股份有限公司 Camera calibration method, device and vehicle
CN111435429A (en) * 2019-01-15 2020-07-21 北京伟景智能科技有限公司 Gesture recognition method and system based on binocular stereo data dynamic cognition

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117373114A (en) * 2023-09-19 2024-01-09 海智合芯科技(深圳)有限公司 Gesture recognition method, system, device and storage medium based on follow-up image
CN117392759A (en) * 2023-12-11 2024-01-12 成都航空职业技术学院 Action recognition method based on AR teaching aid
CN117392759B (en) * 2023-12-11 2024-03-12 成都航空职业技术学院 Action recognition method based on AR teaching aid

Similar Documents

Publication Publication Date Title
US11503275B2 (en) Camera calibration system, target, and process
CN107909600B (en) Unmanned aerial vehicle real-time moving target classification and detection method based on vision
US8467596B2 (en) Method and apparatus for object pose estimation
CN111144207B (en) Human body detection and tracking method based on multi-mode information perception
US10216979B2 (en) Image processing apparatus, image processing method, and storage medium to detect parts of an object
CN105453546B (en) Image processing apparatus, image processing system and image processing method
CN109934065B (en) Method and device for gesture recognition
US10733705B2 (en) Information processing device, learning processing method, learning device, and object recognition device
WO2010099034A1 (en) Capturing and recognizing hand postures using inner distance shape contexts
CN112287867B (en) Multi-camera human body action recognition method and device
CN112200056B (en) Face living body detection method and device, electronic equipment and storage medium
CN107766864B (en) Method and device for extracting features and method and device for object recognition
CN107527368B (en) Three-dimensional space attitude positioning method and device based on two-dimensional code
CN112784712B (en) Missing child early warning implementation method and device based on real-time monitoring
CN112749664A (en) Gesture recognition method, device, equipment, system and storage medium
JP7230345B2 (en) Information processing device and information processing program
CN110598647A (en) Head posture recognition method based on image recognition
CN113723432B (en) Intelligent identification and positioning tracking method and system based on deep learning
CN113591548B (en) Target ring identification method and system
Wang et al. A detection and tracking system for fisheye videos from traffic intersections
US20230009925A1 (en) Object detection method and object detection device
CN113033256B (en) Training method and device for fingertip detection model
CN112766065A (en) Mobile terminal examinee identity authentication method, device, terminal and storage medium
CN115210758A (en) Motion blur robust image feature matching
CN112351181A (en) Intelligent camera based on CMOS chip and ZYNQ system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210504

RJ01 Rejection of invention patent application after publication