CN116805354A - Rare bird three-dimensional model reconstruction method and device based on nerve radiation field - Google Patents

Rare bird three-dimensional model reconstruction method and device based on nerve radiation field Download PDF

Info

Publication number
CN116805354A
CN116805354A CN202311063067.5A CN202311063067A CN116805354A CN 116805354 A CN116805354 A CN 116805354A CN 202311063067 A CN202311063067 A CN 202311063067A CN 116805354 A CN116805354 A CN 116805354A
Authority
CN
China
Prior art keywords
dimensional model
ndc
rare
viewpoint image
light field
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202311063067.5A
Other languages
Chinese (zh)
Other versions
CN116805354B (en
Inventor
江红星
刘畅
吕梦雪
路峰
岳修鹏
张树岩
邱钧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute Of Forest Ecological Environment And Nature Conservation Chinese Academy Of Forestry World Natural Heritage Conservation Research Center State Forestry And Grassland Administration
Shandong Yellow River Delta National Nature Reserve Management Committee
Beijing Information Science and Technology University
Original Assignee
Institute Of Forest Ecological Environment And Nature Conservation Chinese Academy Of Forestry World Natural Heritage Conservation Research Center State Forestry And Grassland Administration
Shandong Yellow River Delta National Nature Reserve Management Committee
Beijing Information Science and Technology University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute Of Forest Ecological Environment And Nature Conservation Chinese Academy Of Forestry World Natural Heritage Conservation Research Center State Forestry And Grassland Administration, Shandong Yellow River Delta National Nature Reserve Management Committee, Beijing Information Science and Technology University filed Critical Institute Of Forest Ecological Environment And Nature Conservation Chinese Academy Of Forestry World Natural Heritage Conservation Research Center State Forestry And Grassland Administration
Priority to CN202311063067.5A priority Critical patent/CN116805354B/en
Publication of CN116805354A publication Critical patent/CN116805354A/en
Application granted granted Critical
Publication of CN116805354B publication Critical patent/CN116805354B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Graphics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Image Processing (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The invention discloses a rare bird three-dimensional model reconstruction method and device based on a nerve radiation field, wherein the method comprises the following steps: step 1, collecting discrete rotating light field data of rare bird targets; step 2, estimating camera pose corresponding to a scene multi-viewpoint image corresponding to the discrete rotation light field data; step 3, according to the camera pose corresponding to the scene multi-viewpoint image obtained in the step 2, obtaining a transformation matrix of the camera pose of the multi-viewpoint image, and further transforming the discrete rotating light field data of the rare bird target object acquired in the step 1 into an NDC space through homogeneous coordinate transformation and ray transformation in the NDC; step 4, generating a new viewpoint image of the rotating light field under the NDC space of the discrete rotating light field data by utilizing the nerve radiation field; and 5, reconstructing a three-dimensional model of the rare birds according to the new viewpoint image of the rotating light field generated in the step 4. The present invention belongs to the technology of computer vision and computer graphics. The method is used for reconstructing the three-dimensional model of the rare birds.

Description

Rare bird three-dimensional model reconstruction method and device based on nerve radiation field
Technical Field
The invention relates to the technical field of three-dimensional image reconstruction, in particular to a rare bird three-dimensional model reconstruction method and device based on a nerve radiation field.
Background
The three-dimensional image reconstruction method of the bird target is based on the technology of computer vision and computer graphics, and aims to extract the three-dimensional shape and structure information of birds from image or video data. The method has important application value in the fields of researching ecology, behaviours, chemical advance and the like of birds.
Techniques involved in reconstructing three-dimensional images of avian targets include: (1) multiview geometry: the position, posture and shape information of the bird target in the three-dimensional space are deduced by utilizing the multi-view geometrical principle through the image or video data acquired from different view angles. (2) feature extraction and matching: in three-dimensional reconstruction of avian objects, features are extracted from image or video data and feature matching is performed to determine correspondence. Common features include corner points, edges, textures and the like, and feature points of the bird body can be identified through feature extraction and matching algorithms. (3) three-dimensional model generation: by utilizing the information obtained by the multi-view geometry and feature matching, different shape reconstruction methods, such as triangular mesh reconstruction, voxel reconstruction and the like, can be adopted to generate the three-dimensional model of the bird. Rare birds have the characteristic of complex morphology and feather structure, and the problem that point clouds are sparse and shielding and concave-convex parts cannot be reconstructed is caused. The invention realizes high-quality and high-precision three-dimensional scene reconstruction by using a neural radiation field.
Disclosure of Invention
It is an object of the present invention to provide a method and apparatus for three-dimensional model reconstruction of rare birds based on a neuro-radiation field that overcomes or at least alleviates at least one of the above-mentioned drawbacks of the prior art.
In order to achieve the above object, the present invention provides a rare bird three-dimensional model reconstruction method based on a nerve radiation field, comprising:
step 1, collecting discrete rotating light field data of rare bird targets;
step 2, estimating camera pose corresponding to a scene multi-viewpoint image corresponding to the discrete rotation light field data;
step 3, according to the camera pose corresponding to the scene multi-viewpoint image obtained in the step 2, obtaining a transformation matrix of the camera pose of the multi-viewpoint image, and further transforming the discrete rotating light field data of the rare bird target object acquired in the step 1 into an NDC (Normalized Device Coordinate, equipment coordinate normalization) space through homogeneous coordinate transformation and ray transformation in the NDC;
step 4, generating a new viewpoint image of the rotating light field under the NDC space of the discrete rotating light field data by utilizing the nerve radiation field;
and 5, reconstructing a three-dimensional model of the rare birds according to the new viewpoint image of the rotating light field generated in the step 4.
Further, the method of the nerve radiation field in the step 4 specifically includes:
step 41, describing the geometric shape and color information of the bird three-dimensional model by using the nerve field, including the space position vector X and the direction vector d, and then mapping the space position vector X and the direction vector d to N-dimensional space respectively by using the following formula (17) to obtain the coordinate codeAnd direction code->Then input into the nerve radiation field, and output the volume density of the position from the nerve radiation field>And the color value of this position in the d-direction +.>
(17)
Step 42, original ray reaching pixel point P of camera along camera optical centerThe color value of the pixel P is obtained from the following formula (18)>
(18)
In the method, in the process of the invention,for rays->Bulk density at parameter S, +.>For rays->Bulk density at parameter t, +.>And->The closest and furthest points along the original ray r, respectively +.>For rays->Pixel values in the d-direction.
Further, the camera pose corresponding to the scene multi-view image in step 2 includes a rotation matrix R, and the method for acquiring the rotation matrix R specifically includes:
step 21a, rotating the camera about the X-axisThe rotation matrix of the angle is expressed as the following formula (6), rotated about the y-axis +.>The rotation matrix of the angle is expressed as the following formula (7), which rotates around the Z-axis +.>The rotation matrix of the angle is expressed as the following formula (8):
(6)
(7)
(8)
step 22a, the rotation matrix R obtained by multiplying the above formula (6), formula (7) and formula (8) is expressed as the following formula (9):
(9)。
further, in step 3, the method for homogeneous coordinate transformation in NDC specifically includes:
homogeneous coordinate pointCorresponding to coordinate point +.>Is of the following formula (13):
(13)
wherein n and f are the distances from the near and far shear planes to the origin, r andthe right and upper bounds of the scene on the near-clipping plane, respectively.
Further, in step 3, the method of ray transformation in NDC specifically includes:
transforming the original ray r into a ray in NDC space with the origin of the ray in NDC spaceAnd direction->Represented by the following formulas (14), (15), respectively:
(14)
(15)
wherein n and f are respectively the near and far shearing planes to the originDistance, r and->Respectively near-shear on-plane scenesRight and upper bounds->、/>、/>Origin of original rays respectively->Coordinate values in x, y, z direction,/-respectively>、/>Coordinate values of the directions d of the original rays in the x, y and z directions are respectively shown.
The invention also provides a rare bird three-dimensional model reconstruction device based on the nerve radiation field, which comprises:
the data acquisition unit is used for acquiring discrete rotating light field data of rare bird targets;
a camera pose estimation unit for estimating a camera pose corresponding to a scene multi-view image corresponding to the discrete rotation light field data;
the NDC space conversion unit is used for obtaining a conversion matrix of the camera pose of the multi-view image according to the camera pose corresponding to the multi-view image of the scene acquired by the camera pose estimation unit, and further converting the discrete rotating light field data of the rare bird target object acquired by the data acquisition unit into an NDC space through homogeneous coordinate conversion and ray conversion in the NDC;
a new viewpoint image acquisition unit for generating a rotating light field new viewpoint image under a discrete rotating light field data NDC space using a neural radiation field;
and a rare bird three-dimensional model reconstruction unit for reconstructing a rare bird three-dimensional model from the rotated light field new viewpoint image generated by the new viewpoint image acquisition unit.
Further, the method for obtaining the nerve radiation field of the new viewpoint image acquisition unit specifically includes:
a coding subunit for describing geometric shape and color information of the bird three-dimensional model with the nerve field, including a spatial position vector X and a direction vector d, and then mapping the spatial position vector X and the direction vector d to N-dimensional space respectively using the following formula (17) to obtain coordinate codesAnd direction code->Then input into the nerve radiation field, and output the volume density of the position from the nerve radiation field>And the pixel value of this position in the d-direction +.>
(17)
A pixel color value calculation subunit for raw rays reaching a pixel point P of the camera along the camera optical centerThe color value of the pixel P is obtained from the following formula (18)>
(18)
In the method, in the process of the invention,for rays->Bulk density at parameter S, +.>For rays->Bulk density at parameter t, +.>And->The closest and furthest points along the original ray r, respectively +.>For rays->Pixel values in the d-direction.
Further, the camera pose corresponding to the scene multi-view image of the camera pose estimation unit comprises a rotation matrix R, and the method for acquiring the rotation matrix R specifically comprises the following steps:
step 21a, rotating the camera about the X-axisThe rotation matrix of the angle is expressed as the following formula (6), rotated about the y-axis +.>The rotation matrix of the angle is expressed as the following formula (7), which rotates around the Z-axis +.>The rotation matrix of the angle is expressed as the following formula (8):
(6)
(7)
(8)
step 22a, the rotation matrix R obtained by multiplying the above formula (6), formula (7) and formula (8) is expressed as the following formula (9):
(9)。
further, the NDC space conversion unit specifically includes a homogeneous coordinate transformation subunit for transforming homogeneous coordinate pointsTransformed into coordinate point +.>Expressed by the following formula (13):
(13)
wherein n and f are the distances from the near and far shear planes to the origin, r andthe right and upper bounds of the scene on the near-clipping plane, respectively.
Further, the NDC space conversion unit specifically includes a ray conversion subunit for converting the original ray r into a ray in the NDC space whose origin is the originAnd direction->Represented by the following formulas (14), (15), respectively:
(14)
(15)
wherein n and f are respectively the near and far shearing planes to the originDistance, r and->The right and upper bounds of the scene on the near-clipping plane, respectively, < >>、/>、/>Origin of original rays respectively->Coordinate values in x, y, z direction,/-respectively>、/>Coordinate values of the directions d of the original rays in the x, y and z directions are respectively shown.
According to the invention, the camera pose corresponding to the scene multi-viewpoint image is firstly obtained, then the consistency space is established through NDC space transformation, the scene is subjected to viewpoint super-resolution along the rotating light field through voxel rendering, a new viewpoint image of the rotating light field is obtained, and finally the new viewpoint image is utilized for three-dimensional image reconstruction, so that the problem that the traditional three-dimensional reconstruction point cloud is sparse and the occlusion and concave-convex parts cannot be reconstructed is effectively solved by the three-dimensional scene reconstructed after super-resolution, and the three-dimensional scene reconstruction with high quality and high precision is completed.
Drawings
Fig. 1 is a schematic diagram of the principle of collecting discrete rotating light field data of rare bird targets according to an embodiment of the present invention.
Detailed Description
The present invention will be described in detail with reference to the accompanying drawings and examples.
The rare bird three-dimensional model reconstruction method based on the nerve radiation field specifically comprises the following steps:
and step 1, collecting discrete rotating light field data of rare bird targets.
As shown in fig. 1, from the perspective of the paper, the left box of fig. 1 illustrates an avian target in which the coordinate system is the image (camera) coordinate system XYZ, as can also be appreciated: the image coordinates in the rotating light field can be expressed as,/>Indicating the rotation angle. The right side of fig. 1 shows: the optical axis of the camera is perpendicular to the Z axis of the world coordinate system XYZ, and the camera is used for sampling around the rotation axis Y at equal intervals to obtain discrete rotation light field data of rare bird targets, wherein the data records the light radiation degree of 360 degrees of a three-dimensional scene.
Setting the rotation center of the scene as o, the focal length of the camera as f, and the coordinates of the optical center of the camera in the world coordinate system XYZ asThe distance between the camera (optical center) and the rotation axis Y is R. For a point P in three-dimensional space, the distance to the axis of rotation Y is R, the initial angle is +.>. In the X-Z plane of the world coordinate system, the X and Z components can be represented by polar coordinates as the following formulas (1) and (2), respectively:
(1)
(2)
at a rotation angleWithin the scope of this, the images sampled at intervals form a three-dimensional image volume +.>The characteristic point locus curve can be expressed as the following formula (3), and projections in the X-direction and the y-direction are respectively shown as formulas (4) to (5), thereby obtaining discrete rotation light field data:
(3)
(4)
(5)
the rotation light field consistency characteristic is distributed as a sine curve in a three-dimensional space, the vision consistency characteristic has high-dimensional continuous light intensity distribution characteristic, and the light intensity distribution track contains the relative motion relation between the three-dimensional scene and the camera. The discrete rotating light field data of the rare bird target object collected by the embodiment of the invention can be used for an orthogonal projection model, and is used for accurately estimating the depth of a scene and increasing the foreground and background light information. In the further three-dimensional reconstruction, a high-precision three-dimensional model free from occlusion and noise interference can be generated.
And 2, estimating camera pose corresponding to the scene multi-view image corresponding to the discrete rotation light field data, wherein the camera pose comprises a rotation matrix R and a translation vector t.
In a three-dimensional spaceIs calibrated at a point P of the (c),position information representing point P, coordinates +.>A fixed point in three-dimensional space may be represented. The camera can also be regarded as a three-dimensional point in space, the angle of the camera +.>Plays an important role in reconstructing the projection, so an additional 3 degrees of freedom are introduced +.>To represent the rotation of the camera in three dimensions. Therefore, the positioning of the camera in the world coordinate system, i.e. the pose of the camera, requires 6 degrees of freedom +.>、/>、/>、/>、/>Representation, degree of freedom->、/>、/>Corresponding to the rotation matrix R of the camera in the world coordinate system,>、/>、/>corresponding to the translation vector t of the camera in the world coordinate system.
In one embodiment, the method for obtaining the rotation matrix R specifically includes:
step 21a, the acquisition cameras are rotated around the X-axis respectivelyAngle, rotation about y-axis->Angle, rotation about Z axis->A rotation matrix of angles.
For example: the rotation matrix rotated by an angle a about the X-axis may be represented by, but is not limited to, formula (6) below, rotated about the y-axisThe rotation matrix of angles can be represented by, but not limited to, the following formula (7), rotated about the Z-axis +.>The rotation matrix of angles may be represented, but is not limited to, as the following formula (8):
(6)
(7)
(8)
step 22a, obtaining a final rotation matrix R by multiplying the above formula (6), formula (7) and formula (8) as the following formula (9):
(9)
in one embodiment, the method for obtaining the translation vector t specifically includes:
、/>、/>representing the translation distances of the camera along the X-axis, y-axis and Z-axis directions, respectively, the translation vector t can be learned by obtaining the translation distances of the camera along the X-axis, y-axis and Z-axis directions.
Of course, in addition to selecting the camera pose corresponding to the multi-view image of the scene using the rotation matrix R and the translation vector t of the camera in the world coordinate system as in the above embodiment, the person skilled in the art may also represent the camera pose corresponding to the multi-view image of the scene in a plurality of different manners, such as rotation vectors, quaternions, euler angles, and the like.
Step 3, according to the camera pose corresponding to the scene multi-viewpoint image obtained in step 2, obtaining a transformation matrix T of the camera pose of the multi-viewpoint image, and further obtaining parameters through homogeneous coordinate transformation and ray transformation in NDC (Normalized Device Coordinate, device coordinate normalization)Origin of ray in NDC space +.>Direction of ray in NDC space +.>
In the embodiment, the discrete rotating light field data of the rare bird target object acquired in the step 1 is converted into an NDC space, so that a real scene can be reconstructed in the NDC space later.
In one embodiment, the transformation matrix T of the multi-viewpoint image camera pose is obtained by the following equation (10):
(10)
in one embodiment, the homogeneous coordinate points are transformed by homogeneous coordinate transformation in the NDCTransformed into coordinate point +.>Wherein->For camera view coordinates>Is the NDC space coordinates. The homogeneous coordinate transformation method in the NDC specifically comprises the following steps:
for homogeneous coordinate pointsThe standard three-dimensional perspective projection matrix M under homogeneous coordinates is:
(11)
wherein n and f are the distances from the near and far shearing planes to the origin, r andthe right and upper bounds of the scene on the near-clipping plane, respectively. For transforming a homogeneous coordinate point +.>Multiplying M to the left and then dividing by the fourth coordinate yields the following formula (12):
(12)
thus, the homogeneous coordinate pointCorresponding to coordinate point +.>Is of the following formula (13):
(13)
in one embodiment, the original ray is transformed by a ray in the NDCTransforming into rays in NDC space +.>Wherein->For the origin of the original ray, d is the direction of the original ray, t represents the original ray +.>Parameter t point,/, of (2)>For the origin of the original ray in NDC space, < >>For the direction of the original ray in NDC space, +.>Representing rays +.>Parameter of->And (5) a dot.
The method for transforming the rays in the NDC specifically comprises the following steps:
the original ray is processedTransformed into a ray in NDC space with origin +.>And direction->Represented by the following formulas (14), (15), respectively:
(14)
(15)
(16)
wherein n and f are respectively the near and far shearing planes to the originDistance, r and->The right and upper bounds of the scene on the near-clipping plane, respectively, < >>、/>、/>Origin of original rays respectively->Coordinate values in x, y, z direction,/-respectively>、/>Coordinate values of the direction d of the original ray in the x, y and z directions respectively, +.>、/>、/>Origin +.>Three coordinates transformed to the origin in NDC space,/->、/>、/>Three coordinates of the direction d transformed into NDC space, respectively.
After the original rays are transformed into NDC space rays, the phenomenon that when a neural radiation field expresses a 360-degree scene, rays in the background are captured by mistake in the foreground, so that a large number of fuzzy noise points exist in an output result can be avoided. The present invention thus models a 360 ° scene around a set NDC spatially normalized grid and multi-view image.
And 4, generating a new viewpoint image of the rotating light field in the NDC space by utilizing the nerve radiation field, so that the scene is subjected to viewpoint super-resolution along the rotating light field through the new viewpoint image of the rotating light field. The rotating light field visual consistency characteristic after the viewpoint super resolution has high-dimensional continuous light intensity distribution characteristics, contains a lot of information (such as foreground and background juncture, high-frequency texture detail positions, scene concave-convex and shielding positions) lacking before, and can be further used for completing high-quality and high-precision three-dimensional scene reconstruction.
Under the NDC space of the discrete rotation light field data, the nerve radiation field gradually optimizes the voxels in the training process, and enhances the implicit expression of the voxels so as to obtain the rendering result under the new viewpoint.
The method for nerve radiation field specifically comprises the following steps:
at step 41, the radiation field representation of the three-dimensional model and the neural networking thereof.
The geometric shape and color information of three-dimensional model of birds are described by nerve field, including spatial position vectorAnd a direction vector->,/>Represents polar angle>Represents the azimuth angle, that is, X represents a point of the three-dimensional model, and d represents the direction along the three-dimensional model point X. Mapping the spatial position vector X and the direction vector d to an N-dimensional space using the following formula (17) to obtain a coordinate code +.>And direction code->Then input into the nerve radiation field, and output the volume density of the position from the nerve radiation field>And the pixel value of this position in the d-direction +.>R, g, b denote red, green, and blue component values, respectively, of the pixel value.
In particular, in order to accelerate the convergence speed of the neural network, the position vector and the direction vector are mapped to a space of a higher dimension (N dimension) using high frequency mapping, and then input to the multi-layered perceptron network. Whereby coordinate encoding can be obtainedAnd direction code->Information in a high frequency space is captured. Wherein the high frequency mapping functionWherein V represents the input of the function; />Applied to X and d. Then the following formula (17):
(17)
and step 42, voxel rendering.
According to classical voxel rendering principles, voxel density can be understood as the different probabilities that a light beam ends up in a certain infinitely small particle. Therefore, the voxel density and the color can be integrated along the light, N sampling points are uniformly sampled on the light, and the integration weight is the cumulative transmittance from near to far, so that the voxel rendering result is obtained. Original rays reaching pixel point P of the camera along the camera optical centerThe color value of the pixel P is obtained from the following formula (18)>
(18)
In the method, in the process of the invention,for rays->Bulk density at parameter S, +.>For rays->Bulk density at parameter t, +.>And->The closest point and the furthest point along r, respectively,/->For rays->Pixel values in the d-direction.
After the radiation field is expressed by the nerve radiation field, the voxel rendering can be utilized to perform viewpoint super-resolution on the scene along the rotating light field.
And 5, reconstructing a three-dimensional model of the rare birds according to the new viewpoint image of the rotating light field generated in the step 4.
The embodiment of the invention also provides a rare bird three-dimensional model reconstruction device based on the nerve radiation field, which comprises a data acquisition unit, a camera pose estimation unit, an NDC space conversion unit, a new viewpoint image acquisition unit and a rare bird three-dimensional model reconstruction unit, wherein:
the data acquisition unit is used for acquiring discrete rotating light field data of rare bird targets.
And the camera pose estimation unit is used for estimating the camera pose corresponding to the scene multi-view image corresponding to the discrete rotation light field data.
The NDC space conversion unit is used for obtaining a conversion matrix of the camera pose of the multi-view image according to the camera pose corresponding to the multi-view image of the scene acquired by the camera pose estimation unit, and further converting the discrete rotating light field data of the rare bird target object acquired by the data acquisition unit into an NDC space through homogeneous coordinate conversion and ray conversion in the NDC.
The new viewpoint image acquisition unit is used for generating a new viewpoint image of the rotating light field under the space of the discrete rotating light field data NDC by utilizing the nerve radiation field.
The three-dimensional model reconstruction unit is used for reconstructing the three-dimensional model of the rare bird according to the new viewpoint image of the rotating light field generated by the new viewpoint image acquisition unit.
In one embodiment, the new viewpoint image acquisition unit specifically includes an encoding subunit and an encoding subunit:
the coding subunit is used for describing the geometric shape and color information of the bird three-dimensional model by using the nerve field, and comprises a space position vector X and a direction vector d, and then mapping the space position vector X and the direction vector d into N-dimensional space respectively by using the method (17) to obtain the coordinate codingAnd direction code->Then input into the nerve radiation field, and output the volume density of the position from the nerve radiation field>And the pixel value of this position in the d-direction +.>
The coding subunit is used for reaching the phase along the optical center of the cameraOriginal ray of pixel point P of machineObtaining the color value of the pixel P from equation (18)>
In one embodiment, the camera pose corresponding to the scene multi-view image of the camera pose estimation unit includes a rotation matrix R, and the method for acquiring the rotation matrix R specifically includes:
step 21a, rotating the camera about the X-axisThe rotation matrix of the angle is expressed as formula (6), rotated about the y-axis +.>The rotation matrix of the angle is expressed as formula (7), which rotates about the Z-axis +.>The rotation matrix of the angle is expressed as formula (8).
In step 22a, the rotation matrix R obtained by multiplying expression (6), expression (7) and expression (8) is expressed as expression (9).
In one embodiment, the NDC space conversion unit specifically includes a homogeneous coordinate transformation subunit for transforming homogeneous coordinate pointsTransformed into coordinate point +.>Expressed by formula (13).
In one embodiment, the NDC space conversion unit specifically includes a ray conversion subunit for converting the original rayTransformed into a ray in NDC space with origin +.>And direction->Expressed by formulas (14) and (15), respectively.
Under the NDC space of the discrete rotating light field data, a new viewpoint image of a 360-degree scene can be obtained after the scene is subjected to viewpoint super-resolution, the vision consistency characteristic of the new viewpoint image has high-dimensional continuous light intensity distribution characteristics, and the new viewpoint image contains a lot of information (such as foreground and background boundary points, high-frequency texture detail positions, scene concave-convex and shielding positions) which is lacking before, so that the problem that the traditional three-dimensional reconstruction point cloud is sparse and the shielding and concave-convex positions cannot be reconstructed can be effectively solved by the three-dimensional scene reconstructed after super-resolution, and the high-quality high-precision three-dimensional scene reconstruction method is completed.
In the experiment, the effect of the traditional COLMAP method and the method is compared with the effect of the traditional COLMAP method in three view angles of a front view, a side view and a top view of a bird scene reconstruction result, and the conclusion is that the method can solve the problems of sparse point cloud and incapability of reconstructing shielding and concave-convex parts to a certain extent, and a high-quality high-precision three-dimensional model of rare birds is completed. Moreover, it can be seen that: the COLMAP method loses a large amount of point cloud information at the wings, the neck and the top of the bird, so that a complete bird model cannot be reconstructed in subsequent dense reconstruction, the texture reconstruction effect at the front wings of the bird is poor, a large amount of noise points exist, the material of the bird feathers cannot be accurately restored, and the overall reconstruction effect is poor; after the viewpoint super-resolution is performed by utilizing the nerve radiation field, the generated new viewpoint image complements the missing information of the bird scene, and the data of the concave-convex parts of the bird bodies can be generated at the new viewpoint, so that the abundant texture details of the bird feathers can be successfully restored. The number of the bird scene sparse point clouds is increased, and then the high-precision three-dimensional model is densely reconstructed.
Finally, it should be pointed out that: the above embodiments are only for illustrating the technical solution of the present invention, and are not limiting. Those of ordinary skill in the art will appreciate that: the technical schemes described in the foregoing embodiments may be modified or some of the technical features may be replaced equivalently; such modifications and substitutions do not depart from the spirit and scope of the technical solutions of the embodiments of the present invention.

Claims (10)

1. The three-dimensional model reconstruction method for rare birds based on the nerve radiation field is characterized by comprising the following steps of:
step 1, collecting discrete rotating light field data of rare bird targets;
step 2, estimating camera pose corresponding to a scene multi-viewpoint image corresponding to the discrete rotation light field data;
step 3, according to the camera pose corresponding to the scene multi-viewpoint image obtained in the step 2, obtaining a transformation matrix of the camera pose of the multi-viewpoint image, and further transforming the discrete rotating light field data of the rare bird target object acquired in the step 1 into an NDC space through homogeneous coordinate transformation and ray transformation in the NDC;
step 4, generating a new viewpoint image of the rotating light field under the NDC space of the discrete rotating light field data by utilizing the nerve radiation field;
and 5, reconstructing a three-dimensional model of the rare birds according to the new viewpoint image of the rotating light field generated in the step 4.
2. The method for reconstructing a three-dimensional model of rare birds based on a nerve radiation field according to claim 1, wherein the method for reconstructing the nerve radiation field in step 4 specifically comprises the following steps:
step 41, describing the geometric shape and color information of the bird three-dimensional model by using the nerve field, including the space position vector X and the direction vector d, and then mapping the space position vector X and the direction vector d to N-dimensional space respectively by using the following formula (17) to obtain the coordinate codeAnd direction code->Then input into the nerve radiation field, and output the bit from the nerve radiation fieldBulk density of placementAnd the color value of this position in the d-direction +.>
(17)
Step 42, original ray reaching pixel point P of camera along camera optical centerThe color value of the pixel P is obtained from the following formula (18)>
(18)
In the method, in the process of the invention,for rays->Bulk density at parameter S, +.>For rays->The bulk density at the point of the parameter t,and->Respectively along the original shotThe closest and furthest point of line r, < >>For rays->Pixel values in the d-direction.
3. The method for reconstructing a three-dimensional model of rare birds based on a nerve radiation field according to claim 1 or 2, wherein the camera pose corresponding to the scene multi-viewpoint image in step 2 comprises a rotation matrix R, and the method for acquiring the rotation matrix R specifically comprises:
step 21a, rotating the camera about the X-axisThe rotation matrix of the angle is expressed as the following formula (6), rotated about the y-axis +.>The rotation matrix of the angle is expressed as the following formula (7), which rotates around the Z-axis +.>The rotation matrix of the angle is expressed as the following formula (8):
(6)
(7)
(8)
step 22a, the rotation matrix R obtained by multiplying the above formula (6), formula (7) and formula (8) is expressed as the following formula (9):
(9)。
4. the method for reconstructing a three-dimensional model of rare birds based on a nerve radiation field according to claim 1 or 2, wherein in step 3, the method for homogeneous coordinate transformation in NDC specifically comprises:
homogeneous coordinate pointCorresponding to coordinate point +.>Is of the following formula (13):
(13)
wherein n and f are the distances from the near and far shear planes to the origin, r andthe right and upper bounds of the scene on the near-clipping plane, respectively.
5. The method for reconstructing a three-dimensional model of rare birds based on a nerve radiation field according to claim 1 or 2, wherein in step 3, the method for transforming the rays in NDC specifically comprises:
transforming the original ray r into a ray in NDC space with the origin of the ray in NDC spaceAnd direction->Represented by the following formulas (14), (15), respectively:
(14)
(15)
wherein n and f are respectively the near and far shearing planes to the originDistance, r and->The right and upper bounds of the scene on the near-clipping plane, respectively, < >>、/>、/>Origin of original rays respectively->Coordinate values in x, y, z direction,/-respectively>、/>、/>Coordinate values of the directions d of the original rays in the x, y and z directions are respectively shown.
6. A rare bird three-dimensional model reconstruction device based on a nerve radiation field, comprising:
the data acquisition unit is used for acquiring discrete rotating light field data of rare bird targets;
a camera pose estimation unit for estimating a camera pose corresponding to a scene multi-view image corresponding to the discrete rotation light field data;
the NDC space conversion unit is used for obtaining a conversion matrix of the camera pose of the multi-view image according to the camera pose corresponding to the multi-view image of the scene acquired by the camera pose estimation unit, and further converting the discrete rotating light field data of the rare bird target object acquired by the data acquisition unit into an NDC space through homogeneous coordinate conversion and ray conversion in the NDC;
a new viewpoint image acquisition unit for generating a rotating light field new viewpoint image under a discrete rotating light field data NDC space using a neural radiation field;
and a rare bird three-dimensional model reconstruction unit for reconstructing a rare bird three-dimensional model from the rotated light field new viewpoint image generated by the new viewpoint image acquisition unit.
7. The three-dimensional model reconstruction device for rare birds based on a nerve radiation field according to claim 6, wherein the new viewpoint image acquisition unit specifically comprises:
a coding subunit for describing geometric shape and color information of the bird three-dimensional model with the nerve field, including a spatial position vector X and a direction vector d, and then mapping the spatial position vector X and the direction vector d to N-dimensional space respectively using the following formula (17) to obtain coordinate codesAnd direction code->Then input into the nerve radiation field, and output the volume density of the position from the nerve radiation field>And the pixel value of this position in the d-direction +.>
(17)
A pixel color value calculation subunit for raw rays reaching a pixel point P of the camera along the camera optical centerThe color value of the pixel P is obtained from the following formula (18)>
(18)
In the method, in the process of the invention,for rays->Bulk density at parameter S, +.>For rays->Bulk density at parameter t, +.>And->The closest and furthest points along the original ray r, respectively +.>For rays->Pixel values in the d-direction.
8. The three-dimensional model reconstruction device for rare birds based on a nerve radiation field according to claim 6 or 7, wherein the camera pose corresponding to the scene multi-viewpoint image of the camera pose estimation unit comprises a rotation matrix R, and the method for acquiring the rotation matrix R specifically comprises:
step 21a, rotating the camera about the X-axisThe rotation matrix of the angle is expressed as the following formula (6), rotated about the y-axis +.>The rotation matrix of the angle is expressed as the following formula (7), which rotates around the Z-axis +.>The rotation matrix of the angle is expressed as the following formula (8):
(6)
(7)
(8)
step 22a, the rotation matrix R obtained by multiplying the above formula (6), formula (7) and formula (8) is expressed as the following formula (9):
(9)。
9. the three-dimensional model reconstruction device for rare birds based on nerve radiation field as claimed in claim 6 or 7, wherein the NDC space conversion unit comprises a homogeneous coordinate transformation subunit for transforming homogeneous coordinate pointsTransformed into coordinate point +.>Expressed by the following formula (13):
(13)
wherein n and f are the distances from the near and far shear planes to the origin, r andthe right and upper bounds of the scene on the near-clipping plane, respectively.
10. The three-dimensional model reconstruction device for rare birds based on a nerve radiation field according to claim 6 or 7, wherein the NDC space conversion unit comprises a ray conversion subunit for converting an original ray r into a ray in an NDC space having an origin of the rayAnd direction->Represented by the following formulas (14), (15), respectively:
(14)
(15)
wherein n and f are respectively the near and far shearing planes to the originDistance, r and->The right and upper bounds of the scene on the near-clipping plane, respectively, < >>、/>、/>Origin of original rays respectively->Coordinate values in x, y, z direction,/-respectively>、/>、/>Coordinate values of the directions d of the original rays in the x, y and z directions are respectively shown.
CN202311063067.5A 2023-08-23 2023-08-23 Rare bird three-dimensional model reconstruction method and device based on nerve radiation field Active CN116805354B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311063067.5A CN116805354B (en) 2023-08-23 2023-08-23 Rare bird three-dimensional model reconstruction method and device based on nerve radiation field

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311063067.5A CN116805354B (en) 2023-08-23 2023-08-23 Rare bird three-dimensional model reconstruction method and device based on nerve radiation field

Publications (2)

Publication Number Publication Date
CN116805354A true CN116805354A (en) 2023-09-26
CN116805354B CN116805354B (en) 2023-12-29

Family

ID=88079670

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311063067.5A Active CN116805354B (en) 2023-08-23 2023-08-23 Rare bird three-dimensional model reconstruction method and device based on nerve radiation field

Country Status (1)

Country Link
CN (1) CN116805354B (en)

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130038696A1 (en) * 2011-08-10 2013-02-14 Yuanyuan Ding Ray Image Modeling for Fast Catadioptric Light Field Rendering
US20130222633A1 (en) * 2012-02-28 2013-08-29 Lytro, Inc. Light-field processing and analysis, camera control, and user interfaces and interaction on light-field capture devices
CN114004941A (en) * 2022-01-04 2022-02-01 苏州浪潮智能科技有限公司 Indoor scene three-dimensional reconstruction system and method based on nerve radiation field
CN114863038A (en) * 2022-07-07 2022-08-05 杭州像衍科技有限公司 Real-time dynamic free visual angle synthesis method and device based on explicit geometric deformation
WO2022167602A2 (en) * 2021-02-04 2022-08-11 Deepmind Technologies Limited Rendering new images of scenes using geometry-aware neural networks conditioned on latent variables
US20220301257A1 (en) * 2021-03-17 2022-09-22 Microsoft Technology Licensing, Llc High resolution neural rendering
CN115359173A (en) * 2022-07-01 2022-11-18 北京邮电大学 Virtual multi-view video generation method and device, electronic equipment and storage medium
CN115690324A (en) * 2022-11-15 2023-02-03 广州中思人工智能科技有限公司 Neural radiation field reconstruction optimization method and device based on point cloud
CN116310076A (en) * 2022-12-29 2023-06-23 深圳万兴软件有限公司 Three-dimensional reconstruction method, device, equipment and storage medium based on nerve radiation field
CN116418961A (en) * 2023-06-09 2023-07-11 深圳臻像科技有限公司 Light field display method and system based on three-dimensional scene stylization

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20130038696A1 (en) * 2011-08-10 2013-02-14 Yuanyuan Ding Ray Image Modeling for Fast Catadioptric Light Field Rendering
US20130222633A1 (en) * 2012-02-28 2013-08-29 Lytro, Inc. Light-field processing and analysis, camera control, and user interfaces and interaction on light-field capture devices
WO2022167602A2 (en) * 2021-02-04 2022-08-11 Deepmind Technologies Limited Rendering new images of scenes using geometry-aware neural networks conditioned on latent variables
US20220301257A1 (en) * 2021-03-17 2022-09-22 Microsoft Technology Licensing, Llc High resolution neural rendering
CN114004941A (en) * 2022-01-04 2022-02-01 苏州浪潮智能科技有限公司 Indoor scene three-dimensional reconstruction system and method based on nerve radiation field
CN115359173A (en) * 2022-07-01 2022-11-18 北京邮电大学 Virtual multi-view video generation method and device, electronic equipment and storage medium
CN114863038A (en) * 2022-07-07 2022-08-05 杭州像衍科技有限公司 Real-time dynamic free visual angle synthesis method and device based on explicit geometric deformation
CN115690324A (en) * 2022-11-15 2023-02-03 广州中思人工智能科技有限公司 Neural radiation field reconstruction optimization method and device based on point cloud
CN116310076A (en) * 2022-12-29 2023-06-23 深圳万兴软件有限公司 Three-dimensional reconstruction method, device, equipment and storage medium based on nerve radiation field
CN116418961A (en) * 2023-06-09 2023-07-11 深圳臻像科技有限公司 Light field display method and system based on three-dimensional scene stylization

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CHANG LIU: "Adaptive matching norm based disparity estimation from light field data", 《SIGNAL PROCESSING》 *
DAWA DERKSEN: "Shadow Neural Radiance Fields for Multi-view Satellite Photogrammetry", 《IEEE XPLORE》 *
程龙;郭立;袁红星;陈晓琳;: "基于光场渲染的动态3D目标重构技术", 中国科学院研究生院学报, no. 06 *
苗源 等: "基于神经辐射场的光场角度域超分辨", 《光学学报》, vol. 43, no. 14 *

Also Published As

Publication number Publication date
CN116805354B (en) 2023-12-29

Similar Documents

Publication Publication Date Title
CN106803267B (en) Kinect-based indoor scene three-dimensional reconstruction method
WO2019219012A1 (en) Three-dimensional reconstruction method and device uniting rigid motion and non-rigid deformation
CN111899328B (en) Point cloud three-dimensional reconstruction method based on RGB data and generation countermeasure network
CN103268629B (en) Unmarked some real time restoration method of 3 D human body form and attitude
CN114666564B (en) Method for synthesizing virtual viewpoint image based on implicit neural scene representation
CN114863038B (en) Real-time dynamic free visual angle synthesis method and device based on explicit geometric deformation
CN114863035B (en) Implicit representation-based three-dimensional human motion capturing and generating method
CN116310076A (en) Three-dimensional reconstruction method, device, equipment and storage medium based on nerve radiation field
CN116958453B (en) Three-dimensional model reconstruction method, device and medium based on nerve radiation field
CN114842136A (en) Single-image three-dimensional face reconstruction method based on differentiable renderer
CN113178009A (en) Indoor three-dimensional reconstruction method utilizing point cloud segmentation and grid repair
WO2024103890A1 (en) Model construction method and apparatus, reconstruction method and apparatus, and electronic device and non-volatile readable storage medium
CN115761178A (en) Multi-view three-dimensional reconstruction method based on implicit neural representation
CN112365589B (en) Virtual three-dimensional scene display method, device and system
WO2024037562A1 (en) Three-dimensional reconstruction method and apparatus, and computer-readable storage medium
CN117671138A (en) Digital twin modeling method and system based on SAM large model and NeRF
CN116805354B (en) Rare bird three-dimensional model reconstruction method and device based on nerve radiation field
CN117501313A (en) Hair rendering system based on deep neural network
CN117422829A (en) Face image synthesis optimization method based on nerve radiation field
CN112562067A (en) Method for generating large-batch point cloud data sets
CN117218192A (en) Weak texture object pose estimation method based on deep learning and synthetic data
CN116616812A (en) NeRF positioning-based ultrasonic autonomous navigation method
CN113284249B (en) Multi-view three-dimensional human body reconstruction method and system based on graph neural network
CN112419467A (en) Rendering efficiency improving method, device and system based on deep learning
Jäger et al. A comparative Neural Radiance Field (NeRF) 3D analysis of camera poses from HoloLens trajectories and Structure from Motion

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 100091 No. 2 east mansion, Beijing, Haidian District

Applicant after: Institute of Forest Ecological Environment and Nature Conservation, Chinese Academy of Forestry (World Natural Heritage Conservation Research Center, State Forestry and Grassland Administration)

Applicant after: BEIJING INFORMATION SCIENCE AND TECHNOLOGY University

Applicant after: Shandong Yellow River Delta National Nature Reserve Management Committee

Address before: No. 265 Yihe Road, Dongying District, Dongying City, Shandong Province, 257091

Applicant before: Shandong Yellow River Delta National Nature Reserve Management Committee

Applicant before: Institute of Forest Ecological Environment and Nature Conservation, Chinese Academy of Forestry (World Natural Heritage Conservation Research Center, State Forestry and Grassland Administration)

Applicant before: BEIJING INFORMATION SCIENCE AND TECHNOLOGY University

GR01 Patent grant
GR01 Patent grant