CN111079618A - Three-dimensional tracking gesture observation likelihood modeling method and device - Google Patents

Three-dimensional tracking gesture observation likelihood modeling method and device Download PDF

Info

Publication number
CN111079618A
CN111079618A CN201911258862.3A CN201911258862A CN111079618A CN 111079618 A CN111079618 A CN 111079618A CN 201911258862 A CN201911258862 A CN 201911258862A CN 111079618 A CN111079618 A CN 111079618A
Authority
CN
China
Prior art keywords
gesture
foreground
edge
observation likelihood
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911258862.3A
Other languages
Chinese (zh)
Inventor
周智
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Unicloud Nanjing Digital Technology Co Ltd
Original Assignee
Unicloud Nanjing Digital Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Unicloud Nanjing Digital Technology Co Ltd filed Critical Unicloud Nanjing Digital Technology Co Ltd
Priority to CN201911258862.3A priority Critical patent/CN111079618A/en
Publication of CN111079618A publication Critical patent/CN111079618A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • G06V40/28Recognition of hand or arm movements, e.g. recognition of deaf sign language
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Psychiatry (AREA)
  • Social Psychology (AREA)
  • Human Computer Interaction (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

A three-dimensional tracked gesture observation likelihood modeling method and device, the method comprising: acquiring a gesture image and preprocessing the gesture image; performing gesture observation likelihood measurement based on image edge characteristics and gesture observation likelihood measurement based on gesture foreground characteristics according to the preprocessed gesture image; and combining the gesture observation likelihood measurement based on the image edge characteristic and the gesture observation likelihood measurement based on the gesture foreground characteristic to construct a gesture observation likelihood model. The method can accurately extract the depth information in three-dimensional stereoscopic vision tracking, and can obtain better gesture understanding and recognition effects.

Description

Three-dimensional tracking gesture observation likelihood modeling method and device
Technical Field
The invention belongs to the field of computers, and particularly relates to a method, a device and a computer readable storage medium.
Background
With the development and popularization of artificial intelligence technology, the modeling and recognition of gesture gestures are increasingly applied to human emotion recognition and intelligent traffic control. An efficient gesture model has to be established to obtain a better understanding and recognition result, wherein the gesture tracking is most widely applied to gesture understanding and recognition especially in the gesture state tracking, but the currently used gesture modeling method still follows the principle of gradient descent in later specific recognition, so that the embarrassment of local minimum can not be avoided, and the recognition effect is poor.
Gesture modeling is the most basic step in artificial intelligence emotion recognition and computer vision gesture tracking, but the field classical three-dimensional gesture modeling method fills up the defect of fast generation of a two-hand interaction model based on hand actions and gestures to a certain extent, but shows the defect of overhigh linear saturation in three-dimensional stereoscopic vision tracking, and particularly cannot accurately extract three-dimensional scene depth information.
Disclosure of Invention
In view of the above-mentioned deficiencies of the prior art, one of the objectives of the present invention is to solve the problem that the existing gesture modeling method cannot accurately extract depth information in the face of three-dimensional stereoscopic tracking.
The embodiment of the invention discloses a three-dimensional tracking gesture observation likelihood modeling method and a three-dimensional tracking gesture observation likelihood modeling device, wherein the method comprises the following steps: acquiring a gesture image and preprocessing the gesture image; performing gesture observation likelihood measurement based on image edge characteristics and gesture observation likelihood measurement based on gesture foreground characteristics according to the preprocessed gesture image; and combining the gesture observation likelihood measurement based on the image edge characteristic and the gesture observation likelihood measurement based on the gesture foreground characteristic to construct a gesture observation likelihood model. The method can accurately extract the depth information in three-dimensional stereoscopic vision tracking, and can obtain better gesture understanding and recognition effects.
In one possible embodiment, the gesture observation likelihood model is as follows:
Figure BDA0002311052100000021
wherein x isiDenotes the ith particle, wedge(Z|xi) The weight of the ith particle after sampling is measured for the gesture observation likelihood based on the image edge characteristics; w is aforeground(Z|xi) Weight of the ith particle for gesture observation likelihood metric based on gesture foreground feature, dedgeAnd dforegroundRespectively represent according to wedge(Z|xi) And wforeground(Z|xi) The variance between the obtained particle weights.
In one possible embodiment, the pre-processing comprises: and carrying out serialization processing on the gesture image.
In one possible embodiment, the performing the gesture observation likelihood metric based on the image edge feature and the gesture observation likelihood metric based on the gesture foreground feature includes: acquiring edge characteristics of a gesture frame image by using edge detection, acquiring a two-dimensional gesture outline of a hand model according to projection coordinates of each joint point of the model on a two-dimensional image plane, and calculating the similarity between the state of the model and an observed value according to the edge and the gesture outline; and comparing the similarity of the model state and the current observed value according to the coincidence degree between the gesture foreground and the projection of the model in the two-dimensional plane.
In one possible embodiment, the image edge feature-based gesture observation likelihood metric is represented as: p is a radical ofedge=exp(-dchamfef(edge,counter)),
The gesture observation likelihood metric based on the gesture foreground features is expressed as:
Pforeground=exp{-[Sforeground∪Sprojection]-[Sforeground∩Sprojection]},
wherein, Chamfer represents the Chamfer distance of the image edge and the gesture model outline, SforegroundRepresenting the foreground region of the gesture, SprojectionRepresenting a three-dimensional gesture model projection.
In one possible embodiment, the method further comprises: adjusting the weight wedge(Z|xi) And wforeground(Z|xi) So as to obtain a modeling result conforming to a preset accuracy.
The embodiment of the invention also discloses a three-dimensional tracking gesture observation likelihood modeling device, which comprises: the preprocessing module is used for acquiring a gesture image and preprocessing the gesture image; the observation likelihood measurement module is used for performing gesture observation likelihood measurement based on image edge characteristics and gesture observation likelihood measurement based on gesture foreground characteristics according to the preprocessed gesture image; and the construction module is used for constructing a gesture observation likelihood model by combining the gesture observation likelihood metric based on the image edge characteristics and the gesture observation likelihood metric based on the gesture foreground characteristics.
In one possible embodiment, the gesture observation likelihood model is as follows:
Figure BDA0002311052100000031
wherein x isiDenotes the ith particle, wedge(Z|xi) The weight of the ith particle after sampling is measured for the gesture observation likelihood based on the image edge characteristics; w is aforeground(Z|xi) Weight of the ith particle for gesture observation likelihood metric based on gesture foreground feature, dedgeAnd dforegroundRespectively represent according to wedge(Z|xi) And wforeground(Z|xi) The variance between the obtained particle weights.
In one possible embodiment, the preprocessing module is further configured to: and carrying out serialization processing on the gesture image.
In a possible embodiment, the observation likelihood measurement module is further configured to acquire an edge feature of the gesture frame image by using edge detection, acquire a two-dimensional gesture contour of the hand model according to projection coordinates of each joint point of the model on a two-dimensional image plane, and calculate similarity between the model state and the observation value according to the edge and the gesture contour; and comparing the similarity of the model state and the current observed value according to the coincidence degree between the gesture foreground and the projection of the model in the two-dimensional plane.
In one possible embodiment, the method further comprises: the adjusting module is used for adjusting the weight wedge(Z|xi) And wforeground(Z|xi) So as to obtain a modeling result conforming to a preset accuracy.
The invention also discloses a computer readable storage medium, wherein a computer program is stored on the computer readable storage medium, and when the computer program is executed by a processor, the computer program realizes any one of the three-dimensional tracking gesture observation likelihood modeling methods.
The invention has the beneficial effects that: the method provides a gesture observation likelihood modeling method based on three-dimensional tracking of particle filtering through preprocessing of gesture image information, image information gesture similarity measurement and similarity measurement based on foreground gesture features. The method improves the precision of the particle filter algorithm and can obtain better gesture understanding and recognition effects. And the conclusion that the result of the gesture observation likelihood model obtained by observing and fusing the two new three-dimensional tracking can be well matched with the original gesture information of the three-dimensional tracking gesture observation likelihood model basic platform is obtained through the verification of the three-dimensional tracking gesture observation likelihood model, so that a better modeling effect is achieved.
Drawings
FIG. 1 is a flow chart of a method according to an embodiment of the present invention;
FIG. 2 is a flowchart illustrating an image preprocessing method according to an embodiment of the present invention;
FIG. 3 is a schematic diagram illustrating a result of detecting an edge of a gesture image according to an embodiment of the present invention;
FIG. 4 is a schematic diagram of a model verification effect according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of an apparatus according to an embodiment of the present invention.
Detailed Description
In order to facilitate understanding of those skilled in the art, the present invention will be further described with reference to the following examples and drawings, which are not intended to limit the present invention.
The embodiment of the invention discloses a three-dimensional tracking gesture observation likelihood modeling method, which comprises the following steps of:
s101, acquiring a gesture image and preprocessing the gesture image.
In one embodiment, the gesture images may be acquired by a camera.
In one embodiment, the pre-processing comprises: and carrying out serialization processing on the gesture image.
Specifically, basic preprocessing is performed according to the expression of a figure, the gesture original signal is obtained from a calibration camera, however, interference events such as overexposure and non-gaussian environment intervention can occur in the process of acquiring an image by the camera, so that the acquired gesture digital image has serious distortion, and the distortion can influence the recognition of the gesture posture in the later period to a great extent, so that preprocessing operation must be performed on the interference signal in advance to meet the requirement of high-quality digital image processing.
Fig. 2 shows the flow of the pre-processing. The first step in the pre-processing process is gesture image acquisition. The scheme can be established under an OpenCV library function of a C + + integrated development environment, a CreatCaputewineWindow () method is sequentially called to establish a capture window of a data source, a CraetWebDriver () method is called to establish connection between a gesture camera and a VS, a DlgCammmorsource () method is called to complete setting of relevant parameters of a camera capture gesture image, finally a callback pointer of a CameraCallbackImg () method is used to point to the capture window to complete transmission of a digital image, after each transmission, the VS automatically judges whether a captured object meets a minimum frame frequency rule or not, if the captured object meets the minimum frame frequency rule, the digital signal is continuously captured, otherwise, sampling and extracting are carried out again on the image.
Gesture image serialization is the addition of a fourth time dimension to the three-dimensional RGB digital image signal. The method is generally set by adopting a settimecapacity () method, and after the image has a fourth dimension related to time, various digital image processing methods such as segmentation and enhancement can be adopted to perform corresponding processing. Digital image processing is based on a gray scale image, and a color gesture image can be grayed by using Laplace gray scale transformation as shown in a formula (3-1):
Gray(x,y)=0.286*R(x,y)+0.584*G(x,y)+0.128*B(x,y) (3-1)
and then Canny edge detection is carried out, a Gaussian filter is used for smoothing an image, the amplitude and the direction of a gradient are calculated by using finite difference of first-order partial derivatives, then depth information is extracted from the corner points, non-maximum suppression is carried out on the combined Canny detection image, finally, the edge is detected and connected by using a dual-threshold algorithm, and hysteresis threshold processing is carried out, so that high-dimensional well depth information of edge detection can be obtained. Although the depth information of the background can be extracted to the maximum extent by edge detection, the problem of edge discontinuity still can be caused by insufficient extraction of corner points, a Chamfer distance transformation formed according to hand-type topological information can be adopted, and missing contours can be completed by utilizing the complementary action of topology. The edge detection results are shown in fig. 3.
And S102, performing gesture observation likelihood measurement based on image edge characteristics and gesture observation likelihood measurement based on gesture foreground characteristics according to the preprocessed gesture image.
Specifically, edge characteristics of a gesture frame image can be obtained by using edge detection, a two-dimensional gesture outline of the hand model is obtained according to projection coordinates of each joint point of the model on a two-dimensional image plane, and similarity between a model state and an observed value is calculated according to the edge and the gesture outline; furthermore, the similarity of the model state and the current observed value can be compared according to the coincidence degree between the gesture foreground and the projection of the model in the two-dimensional plane.
In one embodiment, performing an image edge feature-based gesture similarity metric comprises:
first two data points are defined
Figure BDA0002311052100000061
And
Figure BDA0002311052100000062
the Chamfer distance transformation can be performed on the binarized discontinuous edge as follows, and the Chamfer distance transformation is defined as shown in equation (3-2):
Figure BDA0002311052100000063
then constructing the scanning and traversing process of each digital primitive: definition P and
Figure BDA0002311052100000064
respectively representing a high-dimensional depth-of-field pixel set and a three-dimensional gesture modeling pixel set in a binary image, and performing one-time traversal by using a binary primitive method shown as a formula (3-3), namely, when a point coordinate element of a next edge image does not belong to an edge any more, a result of one-time scanning is marked as 0, and when the point coordinate element of the next edge image belongs to the edgeThe result of one scan is +1 on the last state and finally adopts a minimization process, which is to ensure that the Chamfer distance has a certain digital frame gradient.
f1(p)=min{f1(q)+1:q∈B(p)}fp∈<P>(3-3)
Figure BDA0002311052100000065
Then, defining the coordinate of p as (x, y), four element points of (x +1, y), (x-1, y), (x, y +1) and (x, y-1) exist in the four adjacent domains at the point, and for the edge point after the first scanning, a second scanning is performed as shown in a formula (3-4), so that the processing can ensure that the well depth of the second-order gradient can be matched with the model obtained by modeling of the classical three-dimensional gesture. In order to ensure that the accuracy of the two scans can be superimposed on each other, the second scan is performed in the opposite direction to the first scan.
f2(p)=min{f1(p),f2(q)+1:q∈A(p)} (3-4)
After the edge is introduced into the Chamfer distance transformation, a gesture similarity measurement of the edge and the gesture outline can be defined, as shown in the formula (3-5), so that three-dimensional tracked gesture observation likelihood information can be obtained:
pedge=exp(-dchamfef(edge,counter)) (3-5)
in one embodiment, performing a similarity metric based on the gesture foreground features comprises: the gesture foreground image can be obtained through a Gaussian mixture model or a background difference combined Bayesian method. With the camera parameters and model state parameters known, a model projection in the image plane of the gesture state can be obtained.
Defining a similarity likelihood function according to the gesture foreground image and the model projection information, as shown in a formula (3-6), wherein a union represents a maximum combination pixel area of a classic three-dimensional gesture model foreground information pixel point and a high-order projection gesture information pixel point, and an intersection represents a common area of the classic three-dimensional gesture model foreground information pixel point and the high-order projection gesture information pixel point, so that formed similarity is obtainedThe measure of sexual activity is denoted PforegroundThe expression shows that the similarity measurement of the latter adds high-order projection gesture information on the basis of the classical model.
Pforeground=exp{-[Sforeground∪Sprojection]-[Sforeground∩Sprojection]} (3-6)
S103, combining the gesture observation likelihood measurement based on the image edge feature and the gesture observation likelihood measurement based on the gesture foreground feature to construct a gesture observation likelihood model.
After the gesture observation likelihood measurement based on the image edge characteristics and the gesture observation likelihood measurement based on the gesture foreground characteristics are obtained, a three-dimensional tracking gesture observation likelihood model can be established, and the method is mainly used for redistributing the weight after sampling of the gesture observation likelihood measurement based on the image edge characteristics and the weight of the gesture observation likelihood measurement based on the gesture foreground characteristics to obtain a more ideal model effect.
The gesture observation likelihood model is as follows:
Figure BDA0002311052100000081
wherein x isiDenotes the ith particle, wedge(Z|xi) The weight of the ith particle after sampling is measured for the gesture observation likelihood based on the image edge characteristics; w is aforeground(Z|xi) Weight of the ith particle for gesture observation likelihood metric based on gesture foreground feature, dedgeAnd dforegroundRespectively represent according to wedge(Z|xi) And wforeground(Z|xi) The variance between the obtained particle weights.
The method further comprises the following steps: adjusting the weight wedge(Z|xi) And wforeground(Z|xi) So as to obtain a modeling result conforming to a preset accuracy.
The adjustment of the weight is shown in the formula (3-8) and the formula (3-9),
Figure BDA0002311052100000082
Figure BDA0002311052100000083
the method provides a gesture observation likelihood modeling method based on three-dimensional tracking of particle filtering through preprocessing of gesture image information, image information gesture similarity measurement and similarity measurement based on foreground gesture features. The method improves the precision of the particle filter algorithm, can accurately extract depth information in three-dimensional stereoscopic vision tracking, and can obtain better gesture understanding and recognition effects. And the conclusion that the result of the gesture observation likelihood model obtained by observing and fusing the two new three-dimensional tracking can be well matched with the original gesture information of the three-dimensional tracking gesture observation likelihood model basic platform is obtained through the verification of the three-dimensional tracking gesture observation likelihood model, so that a better modeling effect is achieved.
In the verification stage, the camera acquires gestures, the gesture of the palm is kept unchanged, the gesture of the index finger is generated to move in a translation mode according to an x coordinate axis, in a slow rotation mode according to a Y coordinate axis and in a slow rotation mode along a z coordinate axis, and then three groups of representative hand-shaped motions are selected according to experimental characteristics to respectively represent the translation of the palm and the rotation of the fingers around two axes. As shown in FIG. 4, the left side of each graph represents the classical three-dimensional tracked gesture and the high-order projection gesture information, and the result that the classical three-dimensional tracked gesture and the high-order projection gesture information are fused in the graph can be well matched with the original gesture information of the three-dimensional tracked gesture observation likelihood model basic platform, so that a good modeling effect is achieved.
The embodiment of the invention also discloses a three-dimensional tracking gesture observation likelihood modeling device 10, which comprises: the preprocessing module 101 is configured to acquire a gesture image and preprocess the gesture image; the observation likelihood measurement module 102 is configured to perform gesture observation likelihood measurement based on image edge features and gesture observation likelihood measurement based on gesture foreground features according to the preprocessed gesture image; and the constructing module 103 is used for constructing a gesture observation likelihood model by combining the gesture observation likelihood metric based on the image edge feature and the gesture observation likelihood metric based on the gesture foreground feature.
In one embodiment the gesture observation likelihood model is as follows:
Figure BDA0002311052100000091
wherein x isiDenotes the ith particle, wedge(Z|xi) The weight of the ith particle after sampling is measured for the gesture observation likelihood based on the image edge characteristics; w is aforeground(Z|xi) Weight of the ith particle for gesture observation likelihood metric based on gesture foreground feature, dedgeAnd dforegroundRespectively represent according to wedge(Z|xi) And wforeground(Z|xi) The variance between the obtained particle weights.
In one embodiment the pre-processing module is further configured to: and carrying out serialization processing on the gesture image.
In one embodiment, the observation likelihood measurement module is further configured to acquire an edge feature of a gesture frame image by using edge detection, acquire a two-dimensional gesture contour of the hand model according to projection coordinates of each joint point of the model on a two-dimensional image plane, and calculate similarity between a model state and an observation value according to the edge and the gesture contour; and comparing the similarity of the model state and the current observed value according to the coincidence degree between the gesture foreground and the projection of the model in the two-dimensional plane.
In one embodiment the apparatus 10 further comprises: the adjusting module 104 is used for adjusting the weight wedge(Z|xi) And wforeground(Z|xi) So as to obtain a modeling result conforming to a preset accuracy.
For the specific implementation of the apparatus 10, reference may be made to the method embodiment, which is not described in detail.
In the several embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the above-described embodiments are merely illustrative, and for example, a division of a unit is merely a division of a logic function, and an actual implementation may have another division, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed.
Units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units can be selected according to actual needs to achieve the purpose of the solution of the embodiment of the present invention.
In addition, functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit.
While the invention has been described in terms of its preferred embodiments, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the spirit and scope of the invention.

Claims (11)

1. A three-dimensional tracked gesture observation likelihood modeling method is characterized by comprising the following steps:
acquiring a gesture image and preprocessing the gesture image;
performing gesture observation likelihood measurement based on image edge characteristics and gesture observation likelihood measurement based on gesture foreground characteristics according to the preprocessed gesture image;
and combining the gesture observation likelihood measurement based on the image edge characteristic and the gesture observation likelihood measurement based on the gesture foreground characteristic to construct a gesture observation likelihood model.
2. The method of claim 1, wherein the gesture observation likelihood model is as follows:
Figure FDA0002311052090000011
wherein x isiDenotes the ith particle, wedge(Z|xi) The weight of the ith particle after sampling is measured for the gesture observation likelihood based on the image edge characteristics; w is aforeground(Z|xi) Weight of the ith particle for gesture observation likelihood metric based on gesture foreground feature, dedgeAnd dforegroundRespectively represent according to wedge(Z|xi) And wforeground(Z|xi) The variance between the obtained particle weights.
3. The method of claim 1, wherein the pre-processing comprises: and carrying out serialization processing on the gesture image.
4. The method of claim 1 or 2, wherein performing the image edge feature based gesture observation likelihood metric and the gesture foreground feature based gesture observation likelihood metric comprises: acquiring edge characteristics of a gesture frame image by using edge detection, acquiring a two-dimensional gesture outline of a hand model according to projection coordinates of each joint point of the model on a two-dimensional image plane, and calculating the similarity between the state of the model and an observed value according to the edge and the gesture outline; and comparing the similarity of the model state and the current observed value according to the coincidence degree between the gesture foreground and the projection of the model in the two-dimensional plane.
5. The method of claim 4, wherein the image edge feature based gesture observation likelihood metric is expressed as: p is a radical ofedge=exp(-dchamfef(edge,counter)),
The gesture observation likelihood metric based on the gesture foreground features is expressed as:
Pforeground=exp{-[Sforeground∪Sprojection]-[Sforeground∩Sprojection]},
wherein, Chamfer represents the Chamfer distance of the image edge and the gesture model outline, SforegroundRepresenting the foreground region of the gesture, SprojectionRepresenting a three-dimensional gesture model projection.
6. The method of claim 1, further comprising: adjusting the weight wedge(Z|xi) And wforeground(Z|xi) So as to obtain a modeling result conforming to a preset accuracy.
7. A three-dimensional tracked gesture observation likelihood modeling apparatus, comprising: the preprocessing module is used for acquiring a gesture image and preprocessing the gesture image; the observation likelihood measurement module is used for performing gesture observation likelihood measurement based on image edge characteristics and gesture observation likelihood measurement based on gesture foreground characteristics according to the preprocessed gesture image; and the construction module is used for constructing a gesture observation likelihood model by combining the gesture observation likelihood metric based on the image edge characteristics and the gesture observation likelihood metric based on the gesture foreground characteristics.
8. The apparatus of claim 7, wherein the gesture observation likelihood model is as follows:
Figure FDA0002311052090000021
wherein x isiDenotes the ith particle, wedge(Z|xi) The weight of the ith particle after sampling is measured for the gesture observation likelihood based on the image edge characteristics; w is aforeground(Z|xi) Weight of the ith particle for gesture observation likelihood metric based on gesture foreground feature, dedgeAnd dforegroundRespectively represent according to wedge(Z|xi) And wforeground(Z|xi) The variance between the obtained particle weights.
9. The apparatus of claim 7, wherein the preprocessing module is further configured to: and carrying out serialization processing on the gesture image.
10. The apparatus of claim 7 or 8, wherein the observation likelihood metric module is further configured to obtain edge features of the gesture frame image by using edge detection, obtain a two-dimensional gesture contour of the hand model according to projection coordinates of each joint point of the model in a two-dimensional image plane, and calculate similarity between the model state and the observation value according to the edge and the gesture contour; and comparing the similarity of the model state and the current observed value according to the coincidence degree between the gesture foreground and the projection of the model in the two-dimensional plane.
11. The apparatus of claim 7, further comprising: the adjusting module is used for adjusting the weight wedge(Z|xi) And wforeground(Z|xi) So as to obtain a modeling result conforming to a preset accuracy.
CN201911258862.3A 2019-12-10 2019-12-10 Three-dimensional tracking gesture observation likelihood modeling method and device Pending CN111079618A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911258862.3A CN111079618A (en) 2019-12-10 2019-12-10 Three-dimensional tracking gesture observation likelihood modeling method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911258862.3A CN111079618A (en) 2019-12-10 2019-12-10 Three-dimensional tracking gesture observation likelihood modeling method and device

Publications (1)

Publication Number Publication Date
CN111079618A true CN111079618A (en) 2020-04-28

Family

ID=70313581

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911258862.3A Pending CN111079618A (en) 2019-12-10 2019-12-10 Three-dimensional tracking gesture observation likelihood modeling method and device

Country Status (1)

Country Link
CN (1) CN111079618A (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103116895A (en) * 2013-03-06 2013-05-22 清华大学 Method and device of gesture tracking calculation based on three-dimensional model
CN108256461A (en) * 2018-01-11 2018-07-06 深圳市鑫汇达机械设计有限公司 A kind of gesture identifying device for virtual reality device

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103116895A (en) * 2013-03-06 2013-05-22 清华大学 Method and device of gesture tracking calculation based on three-dimensional model
CN108256461A (en) * 2018-01-11 2018-07-06 深圳市鑫汇达机械设计有限公司 A kind of gesture identifying device for virtual reality device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
刘炳超: "手势三维跟踪中的观测似然模型研究", 《中国优秀硕士学位论文全文数据库 (信息科技辑)》 *

Similar Documents

Publication Publication Date Title
CN110675487B (en) Three-dimensional face modeling and recognition method and device based on multi-angle two-dimensional face
WO2019228473A1 (en) Method and apparatus for beautifying face image
KR100682889B1 (en) Method and Apparatus for image-based photorealistic 3D face modeling
JP4597391B2 (en) Facial region detection apparatus and method, and computer-readable recording medium
JP2015522200A (en) Human face feature point positioning method, apparatus, and storage medium
JP2009525543A (en) 3D face reconstruction from 2D images
WO2023071964A1 (en) Data processing method and apparatus, and electronic device and computer-readable storage medium
CN111160291B (en) Human eye detection method based on depth information and CNN
CN112464847B (en) Human body action segmentation method and device in video
CN111553284A (en) Face image processing method and device, computer equipment and storage medium
US20230230305A1 (en) Online streamer avatar generation method and apparatus
Chen et al. A particle filtering framework for joint video tracking and pose estimation
CN112101208A (en) Feature series fusion gesture recognition method and device for elderly people
JP2021517281A (en) Multi-gesture fine division method for smart home scenes
CN110895683A (en) Kinect-based single-viewpoint gesture and posture recognition method
KR20110021500A (en) Method for real-time moving object tracking and distance measurement and apparatus thereof
CN112233161B (en) Hand image depth determination method and device, electronic equipment and storage medium
CN116580169B (en) Digital man driving method and device, electronic equipment and storage medium
Bauer et al. Refining the fusion of pepper robot and estimated depth maps method for improved 3d perception
KR101923405B1 (en) System for tracking and modeling human faces using AAM with geometric transformation and method therefor
CN111079618A (en) Three-dimensional tracking gesture observation likelihood modeling method and device
CN112183155B (en) Method and device for establishing action posture library, generating action posture and identifying action posture
Ahmed et al. A comparative analysis of time coherent 3D animation reconstruction methods from RGB-D video data
CN111241971A (en) Three-dimensional tracking gesture observation likelihood modeling method
Zhang et al. A 3D Face Modeling and Recognition Method Based on Binocular Stereo Vision and Depth-Sensing Detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20200428

WD01 Invention patent application deemed withdrawn after publication