CN112365589B - Virtual three-dimensional scene display method, device and system - Google Patents

Virtual three-dimensional scene display method, device and system Download PDF

Info

Publication number
CN112365589B
CN112365589B CN202011387657.XA CN202011387657A CN112365589B CN 112365589 B CN112365589 B CN 112365589B CN 202011387657 A CN202011387657 A CN 202011387657A CN 112365589 B CN112365589 B CN 112365589B
Authority
CN
China
Prior art keywords
human body
point cloud
dimensional
model
preset
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202011387657.XA
Other languages
Chinese (zh)
Other versions
CN112365589A (en
Inventor
李小波
甘健
蔡小禹
马伟振
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Oriental Dream Virtual Reality Technology Co ltd
Original Assignee
Oriental Dream Virtual Reality Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Oriental Dream Virtual Reality Technology Co ltd filed Critical Oriental Dream Virtual Reality Technology Co ltd
Priority to CN202011387657.XA priority Critical patent/CN112365589B/en
Publication of CN112365589A publication Critical patent/CN112365589A/en
Application granted granted Critical
Publication of CN112365589B publication Critical patent/CN112365589B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Graphics (AREA)
  • Software Systems (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Geometry (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The application discloses a virtual three-dimensional scene display method, device and system. The method comprises the steps of creating a human body preset model, and predicting to obtain a three-dimensional human body preset model point cloud according to the human body preset model; acquiring a human body depth image acquired by somatosensory equipment, and converting the human body depth image into a three-dimensional human body point cloud; matching the converted three-dimensional human body point cloud with the predicted three-dimensional human body preset model point cloud, and calculating the pose of somatosensory equipment; according to the calculated pose of the somatosensory equipment, fusing the three-dimensional human body point cloud into the existing human body preset model point cloud to obtain a three-dimensional fusion model; and fusing the three-dimensional human body model with the cloud of the preset environment points to obtain three-dimensional scene display, and outputting the three-dimensional scene display to a high-definition display for display. According to the application, the preset human body model is adjusted through the image acquired by the somatosensory equipment, so that the use of the somatosensory equipment is reduced, the reconstruction process of the three-dimensional model is reduced, and the modeling speed of the three-dimensional scene is improved.

Description

Virtual three-dimensional scene display method, device and system
Technical Field
The application relates to the field of virtual somatosensory interaction, in particular to a virtual three-dimensional scene display method, device and system.
Background
The existing virtual three-dimensional scene display system has the following defects:
1. The use is difficult: to improve the accuracy of the final model, multiple motion sensing devices are often required to be identified from multiple angles, which can cause interference among the multiple devices, and the quality of data in their scan overlap areas can be greatly reduced. Meanwhile, when a plurality of devices are used, each device is required to be calibrated independently, and the coordinate system of the plurality of devices after calibration is ensured to be completely consistent. When multiple devices are used simultaneously, the devices are complicated to set, so that the operation of a common user is difficult.
2. The time consumption is large: modeling using a Kinect somatosensory device requires a certain time to scan the user data to build a user model. In the process, the user needs to be ensured to be static, the limbs move, the thoracic cavity is fluctuated due to respiration, and the like can generate noise, so that the accuracy of the model is affected. At the same time, the user's clothing and body shadows can also affect model refinement. The user needs a certain time to generate a high-precision model without too much interference.
3. The calculation cost is large: when the Kinect somatosensory equipment is used for modeling, point cloud modeling is generally adopted, and because the data volume of the point cloud is huge, redundant data and noise interference exist, the calculation complexity is increased. Therefore, points capable of reflecting curved surface characteristics in the point cloud data are often required to be extracted, the data are simplified, and noise is removed, so that the accuracy and efficiency of model reconstruction are improved.
Disclosure of Invention
The application provides a virtual three-dimensional scene display method, device and system, which are used for adjusting a preset human body model through an image acquired by a motion sensing device, reducing the use of the motion sensing device, reducing the reconstruction process of the three-dimensional model and improving the modeling speed of the three-dimensional scene.
A virtual three-dimensional scene display method comprises the following steps:
creating a human body preset model, and predicting according to the human body preset model to obtain a three-dimensional human body preset model point cloud;
acquiring a human body depth image acquired by somatosensory equipment, and converting the human body depth image into a three-dimensional human body point cloud;
Matching the converted three-dimensional human body point cloud with the predicted three-dimensional human body preset model point cloud, and calculating the pose of somatosensory equipment;
according to the calculated pose of the somatosensory equipment, fusing the three-dimensional human body point cloud into the existing human body preset model point cloud to obtain a three-dimensional fusion model;
And fusing the three-dimensional human body model with the cloud of the preset environment points to obtain three-dimensional scene display, and outputting the three-dimensional scene display to a high-definition display for display.
The virtual three-dimensional scene display method comprises the following steps of:
Acquiring point cloud sets which are matched in point cloud positions and have the same number from the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud, and calculating centroids of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud in the point cloud sets;
constructing an error function according to the mass centers of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud;
And (3) minimizing the value of the error function, calculating an optimal rotation matrix and a translation vector, and determining the pose of the somatosensory equipment according to the optimal rotation matrix and the translation vector.
The virtual three-dimensional scene display method comprises the following concrete steps of:
Obtaining human skin color according to the RGB image acquired from the somatosensory equipment, and attaching the human skin color on the three-dimensional human model;
selecting a preset environment point cloud with the lowest color similarity with the human skin color from a preset environment library according to the human skin color;
And carrying out fusion registration on the three-dimensional human body model and the selected preset environment point cloud to obtain three-dimensional scene display, and outputting the three-dimensional scene display to a high-definition display for display.
The virtual three-dimensional scene display method comprises the following steps of fusing a three-dimensional human body point cloud into a human body preset model to obtain a three-dimensional fused model, wherein the method specifically comprises the following steps of:
Preprocessing the three-dimensional human body point cloud and the human body preset model point cloud;
carrying out model registration and fusion on the preprocessed three-dimensional human body point cloud and a human body preset model;
and (3) carrying out boundary corrosion deburring on the fused model to enable the boundary to be smooth.
The application also provides a virtual three-dimensional scene display device, which comprises:
the human body preset model creation module is used for creating a human body preset model, and predicting to obtain a three-dimensional human body preset model point cloud according to the human body preset model;
the three-dimensional human body point cloud construction module is used for acquiring a human body depth image acquired by the somatosensory equipment and converting the human body depth image into a three-dimensional human body point cloud;
The posture calculation module of the somatosensory equipment is used for matching the converted three-dimensional human body point cloud with the predicted three-dimensional human body preset model point cloud to calculate the posture of the somatosensory equipment;
the three-dimensional fusion model construction module is used for fusing the three-dimensional human body point cloud into the existing human body preset model point cloud according to the calculated pose of the somatosensory equipment to obtain a three-dimensional fusion model;
The three-dimensional scene display construction module is used for fusing the three-dimensional human body model with the preset environment point cloud to obtain three-dimensional scene display, and outputting the three-dimensional scene display to the high-definition display for display.
The virtual three-dimensional scene display device comprises a motion sensing equipment pose calculation module, a motion sensing equipment pose calculation module and a motion sensing equipment pose calculation module, wherein the motion sensing equipment pose calculation module is specifically used for obtaining point cloud sets which are matched in position and have the same number from a three-dimensional human body point cloud and a three-dimensional human body preset model point cloud, and calculating centroids of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud in the point cloud sets; constructing an error function according to the mass centers of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud; and (3) minimizing the value of the error function, calculating an optimal rotation matrix and a translation vector, and determining the pose of the somatosensory equipment according to the optimal rotation matrix and the translation vector.
The virtual three-dimensional scene display device is characterized in that the three-dimensional scene display construction module is specifically used for obtaining human skin colors according to RGB images acquired from somatosensory equipment and attaching the human skin colors on the three-dimensional human model; selecting a preset environment point cloud with the lowest color similarity with the human skin color from a preset environment library according to the human skin color; and carrying out fusion registration on the three-dimensional human body model and the selected preset environment point cloud to obtain three-dimensional scene display, and outputting the three-dimensional scene display to a high-definition display for display.
The virtual three-dimensional scene display device is characterized in that the three-dimensional fusion model construction module is specifically used for preprocessing a three-dimensional human body point cloud and a human body preset model point cloud; carrying out model registration and fusion on the preprocessed three-dimensional human body point cloud and a human body preset model; and (3) carrying out boundary corrosion deburring on the fused model to enable the boundary to be smooth.
The application also provides a virtual three-dimensional scene display system, which comprises: the virtual three-dimensional scene display device also comprises somatosensory equipment and a high-definition display.
The virtual three-dimensional scene showing system as described above, wherein the somatosensory device is used for acquiring human depth images and RGB images.
The beneficial effects achieved by the application are as follows:
(1) The application only uses one motion sensing device, so that the problems of the same coordinate system, interference among multiple devices and the like do not need to be considered, and simultaneously, the model generated when the multiple devices are scanned can be obtained with the accuracy comparable to that of the model generated when the multiple devices are scanned.
(2) And adjusting the human body preset model according to the acquired actual human body image, reducing the reconstruction process of the three-dimensional model, and improving the modeling speed of the three-dimensional scene.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the embodiments or the description of the prior art will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments described in the present invention, and other drawings may be obtained according to these drawings for a person having ordinary skill in the art.
Fig. 1 is a schematic diagram of a virtual three-dimensional scene display system according to a first embodiment of the present application;
FIG. 2 is a flow chart of a method for a virtual three-dimensional scene showing device to perform virtual three-dimensional scene showing;
fig. 3 is a schematic diagram of a virtual three-dimensional scene display device.
Detailed Description
The following description of the embodiments of the present invention will be made clearly and fully with reference to the accompanying drawings, in which it is evident that the embodiments described are some, but not all embodiments of the invention. All other embodiments, which can be made by those skilled in the art based on the embodiments of the invention without making any inventive effort, are intended to be within the scope of the invention.
Example 1
The embodiment of the application provides a virtual three-dimensional scene display system, as shown in fig. 1, which comprises a Kinect2 somatosensory device for acquiring user data, a device (such as a PC) for erecting front-end user image data acquisition and performing virtual three-dimensional somatosensory modeling on the rear end, two display processes and a high-definition display for outputting and displaying processed images; the application only uses one motion sensing device, so that the problems of the same coordinate system, interference among multiple devices and the like do not need to be considered, and simultaneously, the model generated when the multiple devices are scanned can be obtained with the accuracy comparable to that of the model generated when the multiple devices are scanned.
Specifically, in the virtual three-dimensional scene display system, the virtual three-dimensional scene display device executes a virtual three-dimensional scene display method, as shown in fig. 2, and the virtual three-dimensional scene display method specifically includes the following steps:
Step 210, creating a human body preset model, and predicting according to the human body preset model to obtain a three-dimensional human body preset model point cloud;
In the embodiment of the application, the human body preset model is created in the PC in advance, and the model is a binding body of bones and muscles conforming to human body construction, wherein the bones and the muscles conforming to human body construction can be standard statures of different humanoid varieties, different sexes and different ages, and then the human body preset model is adjusted in real time according to the acquired actual human body data, so that the reconstruction process of the three-dimensional model is reduced, and the speed of three-dimensional human body modeling is improved.
Step 220, acquiring a human body depth image acquired by somatosensory equipment, and converting the human body depth image into a three-dimensional human body point cloud;
specifically, a skeleton event is created in Kinect2 somatosensory equipment, a skeleton tracking function is opened, and a human body depth image and an RGB image containing skeleton data are acquired; constructing a front-end image acquisition system and a flow control system by using Unity in a PC, and rapidly acquiring a human body depth image acquired by Kinect2 somatosensory equipment through the front-end image acquisition system and the flow control system, wherein the pixel coordinate points of the human body depth image are [ u, v, w ];
in the embodiment of the application, after the PC acquires the human depth image, the human depth image is converted into the point cloud, specifically:
Specifically, the three-dimensional world coordinate points [ x, y, z ] of the human body of the Kinect2 somatosensory device and the pixel coordinate points [ u, v, w ] of the acquired human body depth image have the following relationship:
w=z·s (1)
wherein f x、fy is the focal length of the Kinect2 somatosensory device on the x and y axes, c z、cy is the aperture center of the Kinect2 somatosensory device, and s is the scaling factor of the human depth image;
The PC derives the relationship between the human body depth image and the three-dimensional world coordinate according to the formula after the depth image is acquired as follows:
Performing point cloud construction according to the formula (2); specifically, f x、fy、cz、cy in the formula (2) is defined as an internal reference matrix C of the Kinect2 somatosensory device, and the spatial position and pixel coordinates of each point are expressed as follows using a matrix model:
wherein R and t are the postures of Kinect2 somatosensory equipment, R is a rotation matrix, t is a displacement vector, and s is the ratio of the data of the human depth map to the actual distance; if the Kinect2 somatosensory equipment is set to be in a static state, namely, rotation and translation are not performed, setting R as an identity matrix I and t as 0;
Each point in the converted point cloud picture is used for defining a size position point of a human body, a reference space coordinate axis of the point cloud picture is established at an intersection point of the central position of the human body and the ground, the human body is divided into two parts by adopting a straight line with the center perpendicular to the ground, and each point in the generated point cloud picture has own coordinate position.
Step 230, matching the converted three-dimensional human body point cloud with the predicted three-dimensional human body preset model point cloud, and calculating the pose of the somatosensory equipment;
Specifically, a back-end image processing system is built in a PC through OpenCV and Python, collected data is processed in real time through the back-end processing system, virtual three-dimensional somatosensory modeling is conducted, and a user three-dimensional model is output;
in the embodiment of the application, the pose of the somatosensory equipment is calculated, and the method specifically comprises the following substeps:
Step 231, acquiring point cloud sets with the same number and matched point cloud positions from the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud, and calculating centroids of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud in the point cloud sets;
specifically, the number of point clouds with matched point cloud positions obtained from the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud is n, and the obtained point cloud set is D= { D 1,d′1,d2,d′2......dn,d′n };
Defining a three-dimensional human body point cloud as d i and a three-dimensional human body preset model point cloud as d' i, wherein the mass center point clouds of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud are respectively as follows:
wherein, p i is the centroid point cloud of the three-dimensional human body point cloud, and p' i is the centroid point cloud of the three-dimensional human body preset model.
Step 232, constructing an error function according to the mass centers of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud;
Specifically, by calculation Transforming the three-dimensional human body preset model point cloud into a three-dimensional human body point cloud coordinate system, and constructing an error function of two point clouds as follows:
wherein J (i) is the error function of the construction, R is the rotation matrix, Is a translation vector.
Step 233, enabling the value of the error function to be minimum, calculating an optimal rotation matrix and a translation vector, and determining the pose of the somatosensory equipment according to the optimal rotation matrix and the translation vector;
specifically, bringing the above formula (4) into formula (5) minimizes the value of the error function to obtain:
Setting R in the formula (6) to be 0, and solving the obtained translation vector Is an optimal value; similarly set translation vector/>0, Solving the obtained rotation matrix R as an optimal value; then, the pose of the somatosensory equipment is calculated as follows:
Wherein D is a point cloud set, R is a rotation matrix, Is a modulus of the translation vector.
Referring back to fig. 2, step 240, according to the calculated pose of the somatosensory device, fusing the three-dimensional human body point cloud into the existing human body preset model point cloud to obtain a three-dimensional fused model;
In the embodiment of the application, a three-dimensional human body point cloud is fused into a human body preset model to obtain a three-dimensional fusion model, which specifically comprises the following sub-steps:
Step 241, preprocessing the three-dimensional human body point cloud and the human body preset model point cloud;
Wherein the preprocessing includes geometric distortion correction, noise suppression and filtering;
the geometric distortion correction is to establish a corresponding mathematical model according to the distortion cause, extract required information from the polluted or distorted signal, restore the original appearance along the distortion inverse process, calculate the estimation value of the real model from the distortion model by using a filter, and enable the estimation value to approach the real model to the maximum extent according to the pre-specified error criterion;
the noise suppression specifically adopts a mean value filtering or median filtering method; the average filtering is specifically to select a plurality of adjacent pixels of the current pixel to be processed to form a template, and the average value of the pixels in the template is used for replacing the value of the original pixel; the median filtering is specifically to sort the to-be-processed models according to the pixel values, and generate monotonically ascending or descending two-dimensional data sequences;
The filtering is to filter out the frequency of a specific wave band in the signal to achieve the effects of inhibiting and preventing interference, and specifically adopts the following formula to filter:
Wherein I (x, y, z) is an input three-dimensional human body point cloud or human body preset model point cloud, I' (x, y, z) is a three-dimensional human body point cloud image or human body preset model point cloud image output after filtering, Ω is a neighborhood range of 2n x 2n size with (x, y, z) as a center point, w (I, j, k) is a weight of the filter at (I, j, k), and (I, j, k) is a point in the field range, and w is a normalization coefficient.
Step 242, performing model registration and fusion on the preprocessed three-dimensional human body point cloud and a human body preset model;
The application carries out model registration and fusion treatment, and specifically comprises the following steps: firstly, respectively calculating gradient fields of a three-dimensional human body point cloud and a human body preset model, and replacing the gradient fields at the corresponding positions of the human body preset model by using the gradient fields of the three-dimensional human body point cloud to obtain a fused model gradient field;
specifically, the gradient fields of the three-dimensional human body point cloud, the human body preset model and the fused model are respectively calculated by adopting the following steps:
wherein, Is x-direction unit vector,/>Is a y-direction unit vector,/>Partial derivative of fused model in X direction for three-dimensional human body point cloud/human body preset model,/>The partial derivative of the model in the Y direction after the fusion of the three-dimensional human body point cloud/human body preset model is obtained; grad (u) is a gradient field vector of a three-dimensional human body point cloud/human body preset model/fused model; h 1 and h 2 are scale factors.
Then calculating the divergence of the fused model according to the gradient field of the fused model, and calculating a pixel value matrix of the fused model according to the divergence of the fused image;
Specifically, a second derivative is obtained on the gradient field of the fused model, so as to obtain the divergence of the fused model; then calculating pixel values of the fused model according to the divergence of the fused model and the coefficient matrix of the fused model; the coefficient matrix of the fused model is specifically: calculating data of the central position of the coefficient matrix according to the matrix corresponding to the boundary pixel points of the fused model, and then setting data of two sides of the central position of the coefficient matrix as 1, and setting the forward diagonal data as 1 to obtain the coefficient matrix;
243, carrying out boundary corrosion deburring on the fused model to enable the boundary to be smooth;
and after boundary corrosion deburring is carried out on the fused graph model, carrying out normalization processing on the fused graph model, and converting the pixel value of the fused graph model into a numerical value between 0 and 1.
In the embodiment of the application, after the three-dimensional human body model is obtained through fusion, the three-dimensional human body model is output to the high-definition display for display.
Referring back to fig. 2, step 250, fusing the three-dimensional human body model with the preset environmental point cloud to obtain a three-dimensional scene display, and outputting the three-dimensional scene display to a high-definition display for display;
Specifically, the three-dimensional human body model is fused with a preset environmental point cloud to obtain a three-dimensional scene display, and the method specifically comprises the following substeps:
step1, obtaining human skin color according to an RGB image acquired from somatosensory equipment, and attaching the human skin color on a three-dimensional human model;
Specifically, a face image is separated from an RGB image, then, the average value of the color values of the face image is calculated to obtain the average skin color, and the average skin color is used as the attached skin color of the three-dimensional human body model.
Step2, selecting a preset environment point cloud with the lowest color similarity with the human skin color from a preset environment library according to the human skin color;
Specifically, the similarity between the human skin color and the preset environmental point cloud is calculated by adopting the following formula:
Wherein sim is similarity, A is human skin color, B i (x, y, z) is preset environmental point cloud, n is total number of the preset environmental point cloud, and the minimum similarity is calculated, namely min (sim) is calculated to obtain the selected preset environmental point cloud.
Step3, fusing and registering the three-dimensional human body model and the selected preset environment point cloud to obtain three-dimensional scene display, and outputting the three-dimensional scene display to a high-definition display for display.
Example two
A second embodiment of the present application provides a virtual three-dimensional scene display device, as shown in FIG. 3, including:
the human body preset model creation module 310 is configured to create a human body preset model, and predict and obtain a three-dimensional human body preset model point cloud according to the human body preset model;
The three-dimensional human body point cloud construction module 320 is configured to acquire a human body depth image acquired by the somatosensory device, and convert the human body depth image into a three-dimensional human body point cloud;
The posture calculation module 330 of the somatosensory device is configured to match the converted three-dimensional human body point cloud with the predicted three-dimensional human body preset model point cloud, and calculate the posture of the somatosensory device;
The three-dimensional fusion model construction module 340 is configured to fuse the three-dimensional human body point cloud into the existing human body preset model point cloud according to the calculated pose of the somatosensory device, so as to obtain a three-dimensional fusion model;
the three-dimensional scene display construction module 350 is configured to fuse the three-dimensional human model with a preset environmental point cloud to obtain three-dimensional scene display, and output the three-dimensional scene display to the high-definition display for display.
The three-dimensional scene display construction module is specifically used for obtaining human skin colors according to RGB images acquired from somatosensory equipment and attaching the human skin colors on the three-dimensional human model; selecting a preset environment point cloud with the lowest color similarity with the human skin color from a preset environment library according to the human skin color; and carrying out fusion registration on the three-dimensional human body model and the selected preset environment point cloud to obtain three-dimensional scene display, and outputting the three-dimensional scene display to a high-definition display for display.
Specifically, the posture calculation module 330 of the somatosensory device is specifically configured to obtain point cloud sets with the same number and matching point cloud positions from the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud, and calculate centroids of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud in the point cloud sets; constructing an error function according to the mass centers of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud; and (3) minimizing the value of the error function, calculating an optimal rotation matrix and a translation vector, and determining the pose of the somatosensory equipment according to the optimal rotation matrix and the translation vector.
The three-dimensional fusion model construction module 340 specifically includes a preprocessing sub-module 341, a registration fusion sub-module 342, and a boundary processing sub-module 343;
The preprocessing sub-module 341 preprocesses the three-dimensional human body point cloud and the human body preset model point cloud; the registration fusion sub-module 342 performs model registration and fusion on the preprocessed three-dimensional human body point cloud and a human body preset model; the boundary processing sub-module 343 performs boundary corrosion deburring on the fused model to smooth the boundary.
Specifically, the preprocessing sub-module 341 is specifically configured to perform geometric distortion correction, noise suppression, and filtering on the three-dimensional human body point cloud and the human body preset model point cloud; the registration fusion sub-module 342 is specifically configured to calculate gradient fields of the three-dimensional human body point cloud and the human body preset model, and replace the gradient field at the corresponding position of the human body preset model with the gradient field of the three-dimensional human body point cloud to obtain a fused model gradient field; and calculating the divergence of the fused model according to the gradient field of the fused model, and calculating the pixel value matrix of the fused model according to the divergence of the fused image. After the boundary processing sub-module 343 performs boundary corrosion deburring on the fused graph model, the method further comprises the step of performing normalization processing on the fused graph model and converting the pixel value of the fused graph model into a value between 0 and 1.
The above examples are only specific embodiments of the present application, and are not intended to limit the scope of the present application, but it should be understood by those skilled in the art that the present application is not limited thereto, and that the present application is described in detail with reference to the foregoing examples: any person skilled in the art may modify or easily conceive of the technical solution described in the foregoing embodiments, or perform equivalent substitution of some of the technical features, while remaining within the technical scope of the present disclosure; such modifications, changes or substitutions do not depart from the spirit and scope of the corresponding technical solutions. Are intended to be encompassed within the scope of the present application. Therefore, the protection scope of the present application shall be subject to the protection scope of the claims.

Claims (8)

1. The virtual three-dimensional scene display method is characterized by comprising the following steps of:
creating a human body preset model, and predicting according to the human body preset model to obtain a three-dimensional human body preset model point cloud;
acquiring a human body depth image acquired by somatosensory equipment, and converting the human body depth image into a three-dimensional human body point cloud;
Matching the converted three-dimensional human body point cloud with the predicted three-dimensional human body preset model point cloud, and calculating the pose of somatosensory equipment;
according to the calculated pose of the somatosensory equipment, fusing the three-dimensional human body point cloud into the existing human body preset model point cloud to obtain a three-dimensional fusion model;
fusing the three-dimensional human body model with a preset environmental point cloud to obtain three-dimensional scene display, and outputting the three-dimensional scene display to a high-definition display for display;
the method comprises the following steps of fusing a three-dimensional human body model with a preset environmental point cloud to obtain a three-dimensional scene display, and specifically comprises the following sub-steps:
Obtaining human skin color according to the RGB image acquired from the somatosensory equipment, and attaching the human skin color on the three-dimensional human model;
selecting a preset environment point cloud with the lowest color similarity with the human skin color from a preset environment library according to the human skin color;
and carrying out fusion registration on the three-dimensional human body model and the selected preset environment point cloud to obtain three-dimensional scene display.
2. The virtual three-dimensional scene showing method according to claim 1, wherein the step of calculating the pose of the somatosensory device comprises the following steps:
Acquiring point cloud sets which are matched in point cloud positions and have the same number from the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud, and calculating centroids of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud in the point cloud sets;
constructing an error function according to the mass centers of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud;
And (3) minimizing the value of the error function, calculating an optimal rotation matrix and a translation vector, and determining the pose of the somatosensory equipment according to the optimal rotation matrix and the translation vector.
3. The virtual three-dimensional scene display method according to claim 1, wherein the three-dimensional human body point cloud is fused into a human body preset model to obtain a three-dimensional fusion model, and the method specifically comprises the following sub-steps:
Preprocessing the three-dimensional human body point cloud and the human body preset model point cloud;
carrying out model registration and fusion on the preprocessed three-dimensional human body point cloud and a human body preset model;
and (3) carrying out boundary corrosion deburring on the fused model to enable the boundary to be smooth.
4. A virtual three-dimensional scene display device, comprising:
the human body preset model creation module is used for creating a human body preset model, and predicting to obtain a three-dimensional human body preset model point cloud according to the human body preset model;
the three-dimensional human body point cloud construction module is used for acquiring a human body depth image acquired by the somatosensory equipment and converting the human body depth image into a three-dimensional human body point cloud;
The posture calculation module of the somatosensory equipment is used for matching the converted three-dimensional human body point cloud with the predicted three-dimensional human body preset model point cloud to calculate the posture of the somatosensory equipment;
the three-dimensional fusion model construction module is used for fusing the three-dimensional human body point cloud into the existing human body preset model point cloud according to the calculated pose of the somatosensory equipment to obtain a three-dimensional fusion model;
The three-dimensional scene display construction module is used for fusing the three-dimensional human body model with the preset environment point cloud to obtain three-dimensional scene display, and outputting the three-dimensional scene display to the high-definition display for display;
The three-dimensional scene display construction module is specifically used for obtaining human skin colors according to RGB images acquired from somatosensory equipment and attaching the human skin colors on the three-dimensional human model; selecting a preset environment point cloud with the lowest color similarity with the human skin color from a preset environment library according to the human skin color; and carrying out fusion registration on the three-dimensional human body model and the selected preset environment point cloud to obtain three-dimensional scene display.
5. The virtual three-dimensional scene display device according to claim 4, wherein the motion sensing device pose calculation module is specifically configured to obtain point cloud sets with the same number and matching positions from a three-dimensional human body point cloud and a three-dimensional human body preset model point cloud, and calculate centroids of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud in the point cloud sets; constructing an error function according to the mass centers of the three-dimensional human body point cloud and the three-dimensional human body preset model point cloud; and (3) minimizing the value of the error function, calculating an optimal rotation matrix and a translation vector, and determining the pose of the somatosensory equipment according to the optimal rotation matrix and the translation vector.
6. The virtual three-dimensional scene display device according to claim 4, wherein the three-dimensional fusion model construction module is specifically configured to preprocess a three-dimensional human body point cloud and a human body preset model point cloud; carrying out model registration and fusion on the preprocessed three-dimensional human body point cloud and a human body preset model; and (3) carrying out boundary corrosion deburring on the fused model to enable the boundary to be smooth.
7. A virtual three-dimensional scene display system, comprising the virtual three-dimensional scene display device according to any of claims 4-6, further comprising a somatosensory device and a high definition display.
8. The virtual three-dimensional scene presentation system of claim 7, wherein the motion sensing device is configured to capture human depth images and RGB images.
CN202011387657.XA 2020-12-01 2020-12-01 Virtual three-dimensional scene display method, device and system Active CN112365589B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011387657.XA CN112365589B (en) 2020-12-01 2020-12-01 Virtual three-dimensional scene display method, device and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011387657.XA CN112365589B (en) 2020-12-01 2020-12-01 Virtual three-dimensional scene display method, device and system

Publications (2)

Publication Number Publication Date
CN112365589A CN112365589A (en) 2021-02-12
CN112365589B true CN112365589B (en) 2024-04-26

Family

ID=74536944

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011387657.XA Active CN112365589B (en) 2020-12-01 2020-12-01 Virtual three-dimensional scene display method, device and system

Country Status (1)

Country Link
CN (1) CN112365589B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113555083B (en) * 2021-08-19 2022-05-24 遨博(北京)智能科技有限公司 Massage track generation method
CN114004939B (en) * 2021-12-31 2022-04-19 深圳奥雅设计股份有限公司 Three-dimensional model optimization method and system based on modeling software script

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102074019A (en) * 2010-12-28 2011-05-25 深圳泰山在线科技有限公司 Human tracking method and system
CN103955963A (en) * 2014-04-30 2014-07-30 崔岩 Digital human body three-dimensional reconstruction method and system based on Kinect device
CN104618819A (en) * 2015-03-05 2015-05-13 广州新节奏智能科技有限公司 Television terminal-based 3D somatosensory shopping system and method
CN104992441A (en) * 2015-07-08 2015-10-21 华中科技大学 Real human body three-dimensional modeling method specific to personalized virtual fitting
CN105843386A (en) * 2016-03-22 2016-08-10 宁波元鼎电子科技有限公司 Virtual fitting system in shopping mall
WO2016123913A1 (en) * 2015-02-04 2016-08-11 华为技术有限公司 Data processing method and apparatus
CN108629831A (en) * 2018-04-10 2018-10-09 清华大学 3 D human body method for reconstructing and system based on parametric human body template and inertia measurement
CN108961393A (en) * 2018-06-29 2018-12-07 杭州光珀智能科技有限公司 A kind of human body modeling method and device based on point cloud data stream
CN110175897A (en) * 2019-06-03 2019-08-27 广东元一科技实业有限公司 A kind of 3D synthesis fitting method and system
CN110223387A (en) * 2019-05-17 2019-09-10 武汉奥贝赛维数码科技有限公司 A kind of reconstructing three-dimensional model technology based on deep learning
CN111274909A (en) * 2020-01-16 2020-06-12 重庆邮电大学 Human body point cloud framework extraction method based on deep learning
CN111598998A (en) * 2020-05-13 2020-08-28 腾讯科技(深圳)有限公司 Three-dimensional virtual model reconstruction method and device, computer equipment and storage medium

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102074019A (en) * 2010-12-28 2011-05-25 深圳泰山在线科技有限公司 Human tracking method and system
CN103955963A (en) * 2014-04-30 2014-07-30 崔岩 Digital human body three-dimensional reconstruction method and system based on Kinect device
WO2016123913A1 (en) * 2015-02-04 2016-08-11 华为技术有限公司 Data processing method and apparatus
CN104618819A (en) * 2015-03-05 2015-05-13 广州新节奏智能科技有限公司 Television terminal-based 3D somatosensory shopping system and method
CN104992441A (en) * 2015-07-08 2015-10-21 华中科技大学 Real human body three-dimensional modeling method specific to personalized virtual fitting
CN105843386A (en) * 2016-03-22 2016-08-10 宁波元鼎电子科技有限公司 Virtual fitting system in shopping mall
CN108629831A (en) * 2018-04-10 2018-10-09 清华大学 3 D human body method for reconstructing and system based on parametric human body template and inertia measurement
CN108961393A (en) * 2018-06-29 2018-12-07 杭州光珀智能科技有限公司 A kind of human body modeling method and device based on point cloud data stream
CN110223387A (en) * 2019-05-17 2019-09-10 武汉奥贝赛维数码科技有限公司 A kind of reconstructing three-dimensional model technology based on deep learning
CN110175897A (en) * 2019-06-03 2019-08-27 广东元一科技实业有限公司 A kind of 3D synthesis fitting method and system
CN111274909A (en) * 2020-01-16 2020-06-12 重庆邮电大学 Human body point cloud framework extraction method based on deep learning
CN111598998A (en) * 2020-05-13 2020-08-28 腾讯科技(深圳)有限公司 Three-dimensional virtual model reconstruction method and device, computer equipment and storage medium

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
使用Kinect快速重建三维人体;周瑾;潘建江;童晶;刘利刚;潘志庚;;计算机辅助设计与图形学学报(第06期);全文 *
基于Kinect的人体模板化三维模型拟合重建;袁仁奇;徐增波;;丝绸;第54卷(第10期);29-36 *
基于深度相机的人体点云获取与配准;姚慧;;电脑与信息技术(第05期);全文 *
袁仁奇 ; 徐增波 ; .基于Kinect的人体模板化三维模型拟合重建.丝绸.2017,第54卷(第10期),29-36. *

Also Published As

Publication number Publication date
CN112365589A (en) 2021-02-12

Similar Documents

Publication Publication Date Title
Jiang et al. Selfrecon: Self reconstruction your digital avatar from monocular video
CN111414798B (en) Head posture detection method and system based on RGB-D image
CN109636831B (en) Method for estimating three-dimensional human body posture and hand information
CN107292965B (en) Virtual and real shielding processing method based on depth image data stream
CN106803267B (en) Kinect-based indoor scene three-dimensional reconstruction method
CN106780619B (en) Human body size measuring method based on Kinect depth camera
Hasler et al. Multilinear pose and body shape estimation of dressed subjects from image sets
CN106204718B (en) A kind of simple and efficient 3 D human body method for reconstructing based on single Kinect
CN103733226B (en) Quickly there is the tracking of joint motions
CN109544677A (en) Indoor scene main structure method for reconstructing and system based on depth image key frame
CN111160164B (en) Action Recognition Method Based on Human Skeleton and Image Fusion
CN104376596B (en) A kind of three-dimensional scene structure modeling and register method based on single image
CN111932678B (en) Multi-view real-time human motion, gesture, expression and texture reconstruction system
CN104794722A (en) Dressed human body three-dimensional bare body model calculation method through single Kinect
CN113012293A (en) Stone carving model construction method, device, equipment and storage medium
CN113610889B (en) Human body three-dimensional model acquisition method and device, intelligent terminal and storage medium
CN103106688A (en) Indoor three-dimensional scene rebuilding method based on double-layer rectification method
JP2019096113A (en) Processing device, method and program relating to keypoint data
CN112907631B (en) Multi-RGB camera real-time human body motion capture system introducing feedback mechanism
Collins et al. [POSTER] realtime shape-from-template: System and applications
CN112365589B (en) Virtual three-dimensional scene display method, device and system
Leizea et al. Real-time deformation, registration and tracking of solids based on physical simulation
CN112330813A (en) Wearing three-dimensional human body model reconstruction method based on monocular depth camera
CN110232664A (en) A kind of mask restorative procedure of exorcising based on augmented reality
Cheng et al. An integrated approach to 3D face model reconstruction from video

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant