WO2021237875A1 - 基于图卷积网络的手部数据识别方法、系统和存储介质 - Google Patents
基于图卷积网络的手部数据识别方法、系统和存储介质 Download PDFInfo
- Publication number
- WO2021237875A1 WO2021237875A1 PCT/CN2020/099766 CN2020099766W WO2021237875A1 WO 2021237875 A1 WO2021237875 A1 WO 2021237875A1 CN 2020099766 W CN2020099766 W CN 2020099766W WO 2021237875 A1 WO2021237875 A1 WO 2021237875A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- hand
- dimensional
- coordinates
- convolutional network
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 50
- 239000013598 vector Substances 0.000 claims abstract description 44
- 238000000605 extraction Methods 0.000 claims description 3
- 230000008569 process Effects 0.000 abstract description 20
- 230000003993 interaction Effects 0.000 abstract description 11
- 238000011176 pooling Methods 0.000 description 6
- 238000004364 calculation method Methods 0.000 description 5
- 238000005070 sampling Methods 0.000 description 5
- 230000009286 beneficial effect Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004913 activation Effects 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 210000002478 hand joint Anatomy 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 230000001568 sexual effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
- G06V40/28—Recognition of hand or arm movements, e.g. recognition of deaf sign language
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/32—Normalisation of the pattern dimensions
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
Definitions
- the present invention relates to the field of computer vision technology, in particular to a method, system and storage medium for hand data recognition based on graph convolutional networks.
- the hand gesture recognition process is to wear a specific glove on the hand to make the specific glove track the hand posture data.
- the virtual device receives the real-time posture of the hand and performs it in the virtual reality interface. Tracking display to improve the sense of realism in the virtual reality interface.
- specific gloves and their supporting facilities severely limit the scope of application, making virtual devices unable to be effectively promoted.
- the purpose of the present invention is to provide a hand data recognition method, system and storage medium based on graph convolutional network, which can broaden application scenarios to a certain extent.
- a hand data recognition method based on graph convolutional network includes the following steps:
- the extracting the key point coordinates and the two-dimensional thermal image of the hand image includes:
- the combining the characteristic image and the two-dimensional thermal image to generate a characteristic vector includes:
- a feature vector is calculated through a convolutional network.
- the generating three-dimensional joint point position coordinates according to the feature vector and the key point coordinates includes:
- the three-dimensional joint point position coordinates are calculated according to the vertex coordinates and the key point coordinates.
- vertex coordinates of the three-dimensional grid obtained by calculating according to the feature vector are specifically:
- the graph convolution network is used to calculate the coordinates of all the vertices of the three-dimensional grid.
- a linear graph convolution network is used to regress the three-dimensional joint point position coordinates.
- a hand data recognition system based on graph convolutional network including:
- the acquisition module is used to acquire a hand image in a preset state
- An extraction module for extracting feature images, key point coordinates and two-dimensional thermal images of the hand image
- the combining module is used to combine the feature image and the two-dimensional thermal image to generate a feature vector
- a generating module configured to generate three-dimensional joint point position coordinates according to the feature vector and the key point coordinates
- the restoration module is used to restore the hand posture according to the position coordinates of the three-dimensional joint points.
- a hand data recognition system based on graph convolutional network including:
- At least one memory for storing programs
- At least one processor is configured to load the program to execute the hand data recognition method based on the graph convolutional network.
- a computer-readable storage medium in which instructions executable by a processor are stored, and the instructions executable by the processor are used to implement the hand data recognition method based on graph convolutional network when executed by the processor.
- the beneficial effect of the embodiment of the present invention is that the embodiment of the present invention obtains a hand image in a preset state, and handles the feature image, key point coordinates, and two-dimensional thermal image of the hand image, and then combines the feature image and the two-dimensional thermal image After the combination, the feature vector is generated, and then the three-dimensional joint point position coordinates are generated according to the feature vector and the key point coordinates, and finally the hand posture is restored according to the three-dimensional joint point position coordinates, so that during the virtual interaction process, the interactor can complete the process without wearing special gloves.
- the interaction process thereby simplifying the application equipment of the virtual interaction process, in order to broaden the application scenarios to a certain extent.
- FIG. 1 is a flowchart of a method for hand data recognition based on a graph convolutional network according to a specific embodiment of the present invention
- FIG. 2 is a schematic diagram of a stacked hourglass network structure according to a specific embodiment
- Fig. 3 is a schematic diagram of the distribution of 21 joint nodes in a specific embodiment.
- an embodiment of the present invention provides a hand data recognition method based on a graph convolutional network.
- This embodiment is applied to a control server, and the control server can communicate with multiple terminal devices.
- the terminal device may be a camera, a virtual display device, and so on.
- This embodiment includes steps S11-S15:
- the preset state means that the hand is at the center of the image in the shooting scene, and the hand occupies a moderate proportion of the image.
- the stacked hourglass network is a symmetrical network architecture.
- its multi-scale features are used to recognize gestures, and for each network layer in the process of obtaining low-resolution features, there will be a corresponding network layer during the up-sampling process.
- the overall network architecture first uses convolution and pooling operations to reduce the features to a very low resolution, such as 4*4.
- the network will add a new convolution branch, which is used to directly extract features from the original resolution before pooling, similar to the residual operation, and extracted from the subsequent upsampling operation Feature fusion.
- the network After reaching the lowest resolution, the network starts to up-sampling the features, that is, nearest neighbor interpolation, and combines the information at different scales, and then adds the previously connected features according to the element position.
- the output of the final network is a set of key point heat maps, which are used to predict the probability of the existence of 21 key points in each pixel as shown in Figure 3.
- the resolution of the feature map is gradually reduced, and C1a, C2a, C3a, and C4a are a backup of the corresponding feature maps before down-sampling.
- the feature map that reaches the lowest resolution is gradually up-sampled, and then the restored feature map and the corresponding backup original feature map are combined to obtain C1b, C2b, C3b, and C4b. Under different feature maps, correspondingly extracting different key points of the hand can achieve better accuracy.
- step S13 can be implemented through the following steps:
- the size of the two-dimensional thermal image is converted into the size of the feature image; it can use 1*1 convolution to convert the size of the two-dimensional thermal image containing the key points into the size of the feature image.
- the feature vector is calculated through the convolutional network.
- the structure of the convolutional network is similar to resnet18 and consists of 8 residual layers and 4 pooling layers.
- the convolutional network is used to perform feature vector calculations to improve the accuracy of the calculation results.
- this step is to first calculate the vertex coordinates of the three-dimensional grid according to the feature vector, and then calculate the three-dimensional joint point position coordinates according to the vertex coordinates and the key point coordinates.
- the vertex coordinates of the three-dimensional grid are calculated according to the feature vector, which can be specifically implemented by the following steps:
- the graph convolution network is used to calculate the coordinates of all the vertices of the three-dimensional grid.
- the key point feature vector is input to the graph convolutional network.
- the graph convolutional network outputs the 3D coordinates of all vertices in the 3D grid through a series of network layer calculations, and uses the 3D coordinates of the vertices in the 3D grid to reconstruct the hand surface 3D grid.
- the pre-defined map structure of the triangle grid that identifies the hand surface On the pre-defined map structure of the triangle grid that identifies the hand surface, first perform the image coarsening operation, similar to the process of convolutional neural network pooling, using the Graclus multi-level clustering algorithm to coarsen the image vector, and create The tree structure stores the correspondence between the vertices in the graph vectors of adjacent coarsening levels, and the forward propagation device in the graph convolution will upsample the vertex features in the coarsened graph vectors to the corresponding sub-vertices in the graph structure, Finally, the graph convolution is performed to update the features in the graph network, and the parameter K of all graph convolution layers is set to 3.
- the feature vector extracted from the hourglass network is used as the input of the graph convolution.
- the feature vector is converted into 80 vertices with 64-dimensional features during the graph coarsening process, and these features are then used in the convolution process.
- the up-sampling is transformed from low-dimensional to high-dimensional.
- the network outputs the 3D coordinates of 1280 mesh vertices.
- the three-dimensional joint point position coordinates are calculated according to the vertex coordinates and the key point coordinates, which can be implemented in the following ways:
- a linear graph convolution network is used to return the three-dimensional joint point position coordinates.
- a simplified linear graph convolution can be specifically used to linearly regress the position coordinates of the 3D hand joint points from the coordinates of the vertices of the three-dimensional hand grid.
- the vertex coordinates of the three-dimensional mesh include the coordinates of the key points of the entire hand, from which the three-dimensional coordinates of 21 joint nodes can be directly filtered.
- 21 joint points from 0 joint points to 20 joint points.
- a joint node covers the entire hand posture.
- a two-layer graph convolutional network without a nonlinear activation module is used to directly estimate the 3D joint depth information from the 3D mesh vertices, and then use the previously obtained 2D key points to generate 3D joint position coordinates.
- the coordinates of the joint points covering the entire hand posture can be extracted, thereby improving the accuracy of the virtual hand posture synchronization process in virtual reality.
- S15 Restore the hand posture according to the position coordinates of the three-dimensional joint points. Specifically, it restores the hand posture corresponding to the hand image in the virtual reality interface according to the three-dimensional joint point position coordinates, so that the hand posture data in the virtual reality can be synchronized with the actual hand posture to the greatest extent, and the synchronization in the virtual interaction process is enhanced sex.
- this embodiment obtains a hand image in a preset state, and handles the feature image, key point coordinates, and two-dimensional thermal image of the hand image, and then combines the feature image and the two-dimensional thermal image to generate the feature Vector, and then generate the three-dimensional joint point position coordinates according to the feature vector and the key point coordinates, and finally restore the hand posture according to the three-dimensional joint point position coordinates, so that in the virtual interaction process, the interactor can complete the virtual interaction process without wearing special gloves.
- Application equipment that simplifies the virtual interaction process to expand application scenarios to a certain extent.
- the embodiment of the present invention provides a hand data recognition system based on a graph convolutional network corresponding to the method in FIG. 1, including:
- the acquisition module is used to acquire a hand image in a preset state
- An extraction module for extracting feature images, key point coordinates and two-dimensional thermal images of the hand image
- the combining module is used to combine the feature image and the two-dimensional thermal image to generate a feature vector
- a generating module configured to generate three-dimensional joint point position coordinates according to the feature vector and the key point coordinates
- the restoration module is used to restore the hand posture according to the position coordinates of the three-dimensional joint points.
- the embodiment of the present invention provides a hand data recognition system based on graph convolutional network, including:
- At least one memory for storing programs
- At least one processor is configured to load the program to execute the hand data recognition method based on the graph convolutional network.
- an embodiment of the present invention provides a computer-readable storage medium, in which instructions executable by a processor are stored, and the instructions executable by the processor are used to implement the graph-based convolution when executed by the processor. Hand data recognition method of the network.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computing Systems (AREA)
- Evolutionary Biology (AREA)
- Biomedical Technology (AREA)
- Psychiatry (AREA)
- Social Psychology (AREA)
- Human Computer Interaction (AREA)
- Processing Or Creating Images (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (10)
- 一种基于图卷积网络的手部数据识别方法,其特征在于,包括以下步骤:获取预设状态的手部图像;提取所述手部图像的特征图像、关键点坐标和二维热图像;将所述特征图像和所述二维热图像进行结合,生成特征向量;根据所述特征向量和所述关键点坐标生成三维关节点位置坐标;根据所述三维关节点位置坐标还原手部姿态。
- 根据权利要求1所述的一种基于图卷积网络的手部数据识别方法,其特征在于,所述提取所述手部图像的关键点坐标和二维热图像,包括:采用堆叠沙漏网络从所述第一图像中提取关键点特征位置;根据所述关键点特征位置预测所述二维热图,以及确定所述关键点坐标。
- 根据权利要求1所述的一种基于图卷积网络的手部数据识别方法,其特征在于,所述将所述特征图像和所述二维热图像进行结合,生成特征向量,包括:将所述二维热图像的尺寸大小转换为所述特征图像的尺寸大小;根据所述特征图像和尺寸转化后的所述二维热图通过卷积网络计算得到特征向量。
- 根据权利要求1所述的一种基于图卷积网络的手部数据识别方法,其特征在于,所述根据所述特征向量和所述关键点坐标生成三维关节点位置坐标,包括:根据所述特征向量计算得到三维网格的顶点坐标;根据所述顶点坐标和所述关键点坐标计算得到三维关节点位置坐标。
- 根据权利要求4所述的一种基于图卷积网络的手部数据识别方法,其特征在于,所述根据所述特征向量计算得到三维网格的顶点坐标,其具体为:根据所述特征向量采用图卷积网络计算得到三维网格的所有顶点坐标。
- 根据权利要求4所述的一种基于图卷积网络的手部数据识别方法,其特征在于,所述根据所述顶点坐标和所述关键点坐标计算得到三维关节点位置坐标,其具体为:根据所述顶点坐标和所述关键点坐标采用线性图卷积网络回归三维关节点位置坐标。
- 根据权利要求1所述的一种基于图卷积网络的手部数据识别方法,其特征在于,所述根据所述三维关节点位置坐标还原手部姿态,其具体为:根据所述三维关节点位置坐标在虚拟现实界面中还原手部图像对应的手部姿态。
- 一种基于图卷积网络的手部数据识别系统,其特征在于,包括:获取模块,用于获取预设状态的手部图像;提取模块,用于提取所述手部图像的特征图像、关键点坐标和二维热图像;结合模块,用于将所述特征图像和所述二维热图像进行结合,生成特征向量;生成模块,用于根据所述特征向量和所述关键点坐标生成三维关节点位置坐标;还原模块,用于根据所述三维关节点位置坐标还原手部姿态。
- 一种基于图卷积网络的手部数据识别系统,其特征在于,包括:至少一个存储器,用于存储程序;至少一个处理器,用于加载所述程序以执行如权利要求1-7任一项所述的基于图卷积网络的手部数据识别方法。
- 一种计算机可读存储介质,其中存储有处理器可执行的指令,其特征在于,所述处理器可执行的指令在由处理器执行时用于实现如权利要求1-7任一项所述的基于图卷积网络的手部数据识别方法。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010473675.3A CN111753669A (zh) | 2020-05-29 | 2020-05-29 | 基于图卷积网络的手部数据识别方法、系统和存储介质 |
CN202010473675.3 | 2020-05-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021237875A1 true WO2021237875A1 (zh) | 2021-12-02 |
Family
ID=72674421
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/099766 WO2021237875A1 (zh) | 2020-05-29 | 2020-07-01 | 基于图卷积网络的手部数据识别方法、系统和存储介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111753669A (zh) |
WO (1) | WO2021237875A1 (zh) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111260774A (zh) * | 2020-01-20 | 2020-06-09 | 北京百度网讯科技有限公司 | 生成3d关节点回归模型的方法和装置 |
CN114565815A (zh) * | 2022-02-25 | 2022-05-31 | 包头市迪迦科技有限公司 | 一种基于三维模型的视频智能融合方法及系统 |
CN116486489A (zh) * | 2023-06-26 | 2023-07-25 | 江西农业大学 | 基于语义感知图卷积的三维手物姿态估计方法及系统 |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113095254B (zh) * | 2021-04-20 | 2022-05-24 | 清华大学深圳国际研究生院 | 一种人体部位关键点的定位方法及系统 |
CN113724393B (zh) * | 2021-08-12 | 2024-03-19 | 北京达佳互联信息技术有限公司 | 三维重建方法、装置、设备及存储介质 |
CN115909406A (zh) * | 2022-11-30 | 2023-04-04 | 广东海洋大学 | 一种基于多类分类的手势识别方法 |
CN116243803B (zh) * | 2023-05-11 | 2023-12-05 | 南京鸿威互动科技有限公司 | 一种基于vr技术的动作评估方法、系统、设备及可读存储介质 |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170083752A1 (en) * | 2015-09-18 | 2017-03-23 | Yahoo! Inc. | Face detection |
CN108830150A (zh) * | 2018-05-07 | 2018-11-16 | 山东师范大学 | 一种基于三维人体姿态估计方法及装置 |
CN109821239A (zh) * | 2019-02-20 | 2019-05-31 | 网易(杭州)网络有限公司 | 体感游戏的实现方法、装置、设备及存储介质 |
CN110427877A (zh) * | 2019-08-01 | 2019-11-08 | 大连海事大学 | 一种基于结构信息的人体三维姿态估算的方法 |
CN110874865A (zh) * | 2019-11-14 | 2020-03-10 | 腾讯科技(深圳)有限公司 | 三维骨架生成方法和计算机设备 |
CN110991319A (zh) * | 2019-11-29 | 2020-04-10 | 广州市百果园信息技术有限公司 | 手部关键点检测方法、手势识别方法及相关装置 |
CN111062261A (zh) * | 2019-11-25 | 2020-04-24 | 维沃移动通信(杭州)有限公司 | 一种图像处理方法及装置 |
-
2020
- 2020-05-29 CN CN202010473675.3A patent/CN111753669A/zh active Pending
- 2020-07-01 WO PCT/CN2020/099766 patent/WO2021237875A1/zh active Application Filing
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170083752A1 (en) * | 2015-09-18 | 2017-03-23 | Yahoo! Inc. | Face detection |
CN108830150A (zh) * | 2018-05-07 | 2018-11-16 | 山东师范大学 | 一种基于三维人体姿态估计方法及装置 |
CN109821239A (zh) * | 2019-02-20 | 2019-05-31 | 网易(杭州)网络有限公司 | 体感游戏的实现方法、装置、设备及存储介质 |
CN110427877A (zh) * | 2019-08-01 | 2019-11-08 | 大连海事大学 | 一种基于结构信息的人体三维姿态估算的方法 |
CN110874865A (zh) * | 2019-11-14 | 2020-03-10 | 腾讯科技(深圳)有限公司 | 三维骨架生成方法和计算机设备 |
CN111062261A (zh) * | 2019-11-25 | 2020-04-24 | 维沃移动通信(杭州)有限公司 | 一种图像处理方法及装置 |
CN110991319A (zh) * | 2019-11-29 | 2020-04-10 | 广州市百果园信息技术有限公司 | 手部关键点检测方法、手势识别方法及相关装置 |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111260774A (zh) * | 2020-01-20 | 2020-06-09 | 北京百度网讯科技有限公司 | 生成3d关节点回归模型的方法和装置 |
CN111260774B (zh) * | 2020-01-20 | 2023-06-23 | 北京百度网讯科技有限公司 | 生成3d关节点回归模型的方法和装置 |
CN114565815A (zh) * | 2022-02-25 | 2022-05-31 | 包头市迪迦科技有限公司 | 一种基于三维模型的视频智能融合方法及系统 |
CN114565815B (zh) * | 2022-02-25 | 2023-11-03 | 包头市迪迦科技有限公司 | 一种基于三维模型的视频智能融合方法及系统 |
CN116486489A (zh) * | 2023-06-26 | 2023-07-25 | 江西农业大学 | 基于语义感知图卷积的三维手物姿态估计方法及系统 |
CN116486489B (zh) * | 2023-06-26 | 2023-08-29 | 江西农业大学 | 基于语义感知图卷积的三维手物姿态估计方法及系统 |
Also Published As
Publication number | Publication date |
---|---|
CN111753669A (zh) | 2020-10-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021237875A1 (zh) | 基于图卷积网络的手部数据识别方法、系统和存储介质 | |
CN111160085A (zh) | 一种人体图像关键点姿态估计方法 | |
CN112330729B (zh) | 图像深度预测方法、装置、终端设备及可读存储介质 | |
Wang et al. | Laplacian pyramid adversarial network for face completion | |
CN111598993A (zh) | 基于多视角成像技术的三维数据重建方法、装置 | |
WO2022100419A1 (zh) | 一种图像处理方法及相关设备 | |
CN112837215B (zh) | 一种基于生成对抗网络的图像形状变换方法 | |
CN112991350A (zh) | 一种基于模态差异缩减的rgb-t图像语义分割方法 | |
WO2023083030A1 (zh) | 一种姿态识别方法及其相关设备 | |
CN113593001A (zh) | 目标对象三维重建方法、装置、计算机设备和存储介质 | |
CN114332415A (zh) | 基于多视角技术的输电线路廊道的三维重建方法及装置 | |
CN116097307A (zh) | 图像的处理方法及相关设备 | |
WO2022213623A1 (zh) | 图像生成、三维人脸模型生成的方法、装置、电子设备及存储介质 | |
CN114693557A (zh) | 基于姿态引导的行人视频修复方法、系统、设备和介质 | |
EP4292059A1 (en) | Multiview neural human prediction using implicit differentiable renderer for facial expression, body pose shape and clothes performance capture | |
Su et al. | Virtualpose: Learning generalizable 3d human pose models from virtual data | |
Zhou et al. | A superior image inpainting scheme using Transformer-based self-supervised attention GAN model | |
Wang et al. | Learning continuous depth representation via geometric spatial aggregator | |
CN117422851A (zh) | 虚拟换衣方法及其装置、电子设备 | |
US20230104702A1 (en) | Transformer-based shape models | |
KR20230083212A (ko) | 객체 자세 추정 장치 및 방법 | |
CN112508776B (zh) | 动作迁移方法、装置和电子设备 | |
CN114723973A (zh) | 大尺度变化鲁棒的图像特征匹配方法及装置 | |
CN115049764A (zh) | Smpl参数预测模型的训练方法、装置、设备及介质 | |
CN117036658A (zh) | 一种图像处理方法及相关设备 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20938377 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20938377 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 19.06.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20938377 Country of ref document: EP Kind code of ref document: A1 |