CN115457192A - Method for realizing NERF new visual angle synthetic model based on self-integration module - Google Patents
Method for realizing NERF new visual angle synthetic model based on self-integration module Download PDFInfo
- Publication number
- CN115457192A CN115457192A CN202210911889.3A CN202210911889A CN115457192A CN 115457192 A CN115457192 A CN 115457192A CN 202210911889 A CN202210911889 A CN 202210911889A CN 115457192 A CN115457192 A CN 115457192A
- Authority
- CN
- China
- Prior art keywords
- nerf
- self
- scene
- integration module
- rgb color
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 41
- 230000000007 visual effect Effects 0.000 title claims abstract description 14
- 238000009877 rendering Methods 0.000 claims abstract description 25
- 230000015572 biosynthetic process Effects 0.000 claims abstract description 16
- 238000003786 synthesis reaction Methods 0.000 claims abstract description 16
- 238000012549 training Methods 0.000 claims abstract description 9
- 238000012360 testing method Methods 0.000 claims description 11
- 238000005070 sampling Methods 0.000 claims description 10
- 238000002834 transmittance Methods 0.000 claims description 3
- 230000000694 effects Effects 0.000 abstract description 8
- 230000005855 radiation Effects 0.000 abstract description 6
- 210000005036 nerve Anatomy 0.000 abstract description 4
- 230000001537 neural effect Effects 0.000 description 5
- 230000006870 function Effects 0.000 description 4
- 238000005457 optimization Methods 0.000 description 4
- 238000012512 characterization method Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 238000012935 Averaging Methods 0.000 description 1
- 125000003580 L-valyl group Chemical group [H]N([H])[C@]([H])(C(=O)[*])C(C([H])([H])[H])(C([H])([H])[H])[H] 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000013528 artificial neural network Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000005266 casting Methods 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 238000011478 gradient descent method Methods 0.000 description 1
- 235000019692 hotdogs Nutrition 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 230000001737 promoting effect Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 230000002194 synthesizing effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/54—Extraction of image or video features relating to texture
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30168—Image quality inspection
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Software Systems (AREA)
- Image Generation (AREA)
Abstract
The invention discloses a realization method of a NERF new visual angle synthesis model based on a self-integration module. The invention comprises the following steps: step 1: selecting image data, and selecting training data open source in googledrive by an NERF author, wherein the training data comprises generated music height data and scene data, and the music height data and the scene data are respectively stored in two folders; and 2, step: constructing a NERF new perspective synthesis model based on a self-integration module, wherein the NERF new perspective synthesis model comprises the steps of acquiring scene representation of NERF; voxel rendering based on a scene representation; and constructing a self-integration module. According to the invention, the self-integration module is added into the NERF nerve radiation field, so that the peak signal-to-noise ratio of an output image can be improved, and the image quality is improved. The self-integration module provided by the invention has an enhanced effect on reconstructing a 3D new visual angle and can better restore the specific details of a scene.
Description
Technical Field
The invention provides a realization method of a NERF new visual angle synthesis model based on a self-integration module.
Background
New view synthesis can be divided into image-based methods, learning-based methods and geometry-based methods. Image-based methods warp and blend the relevant blobs in the observation frame, generating a new view based on the quality measurements. Learning-based methods predict the effects of hybrid weights and view dependence through neural networks and other heuristics. Deep learning also facilitates methods of predicting new views from a single image, but they typically require a large amount of data to train. Unlike image-based and learning-based methods, geometry-based methods first reconstruct a three-dimensional model and render images from the target pose. For example, aliev et al assign multi-resolution features to a point cloud and then perform neural rendering; thies et al store the neural texture on a three-dimensional grid and then use a conventional graphics pipeline to render a new view. Other geometric representations include multi-planar images, voxel grids, depths, and hierarchical depths. These methods, while producing relatively high quality results, require extensive data and memory for the discrete representation, and the resolution of the rendering is limited by the accuracy of the reconstructed geometry.
Implicit neural representation has proven its usefulness in representing shapes and scenes, which typically utilizes multi-layer perceptrons (MLPs) to encode signed distance field distances fields, occupancy, or volume density. Together with micro-renderable, these methods can reconstruct the geometry and appearance of objects and scenes. Among other things, the neural radiation field (NeRF) achieves a significant effect in synthesizing new views of a static scene given a set of pose input images. The key idea of NeRF is to construct a continuous radiation field using MLP and acquire images by micro-volume rendering, so the optimization process can be achieved by minimizing photometric loss, and more NeRF extensions appear, for example: there is no reconstruction of the input camera pose, modeling non-rigid scene non-rigid scenes, unbounded scene and object class objects. NeRF has also been studied as being able to reactivate, generate, edit and 3D reconstruct.
In addition, mip-NeRF also considers the problem of NeRF solution. They show that NeRF rendering at different resolutions will introduce aliasing artifacts and resolve it by proposing an integrated position coding, with a cone instead of a single point. However, mip-NeRF only considers renderings with low sampling resolution, and there is currently no work to investigate how to improve the resolution of NeRF.
The work of the present invention is also related to super-resolution of images. Classical Single Image Super Resolution (SISR) methods utilize priors, such as image statistics or gradients. The CNN-based approach aims to learn the relationship between HR and LR images in CNN by minimizing the mean square error between the SR image and the ground truth. Generation of countermeasure networks (GANs), which produce high resolution details through countermeasure learning, are also popular in super resolution. Most of these methods derive knowledge from large-scale data sets or existing high-resolution and low-resolution training pairs. Furthermore, these two-dimensional image-based methods, especially GAN-based methods, do not take into account view consistency and are suboptimal for new view synthesis.
The existing method matches the correspondence between patch-match, feature extraction or HR reference of attention and LR input. Inspired by their work, we also focused on learning HR details from a given reference image. However, most of these methods are based on referencing super-resolution one SR input, and our method can refine details from all new views through one reference image.
Disclosure of Invention
The invention aims to provide a realization method of a NERF new visual angle synthesis model based on a self-integration module aiming at the defects of the prior art. The invention develops a new NERF nerve radiation field based on a self-integration module, and provides a new self-integration module to reconstruct a new high-quality visual angle image. The self-integration module is mainly used in a testing stage, and images with better reconstruction effect are used in the testing stage. The self-integrated module-based NERF neural rendering field demonstrated greater enhanced reconstruction capabilities than the original NERF, through extensive testing on the data set sourced by the NERF authors.
The technical scheme adopted by the invention for solving the technical problems is as follows:
step 1: selecting image data
Selecting training data which are open in the google drive by a NERF author, wherein the training data comprise generated music height data and scene data, and the music height data and the scene data are respectively stored in two folders;
step 2: constructing a NERF new perspective synthesis model based on the self-integration module.
Further, step 2 is specifically implemented as follows:
2-1, acquiring scene representation of NERF;
2-2, voxel rendering based on scene representation;
and 2-3, constructing a self-integration module.
Further, the step 2-1 is specifically realized as follows:
the input of the MLP is the three-dimensional space coordinate and the view angle corresponding to the source imageThe output is (R, G, B, sigma), and after the scene representation (R, G, B, sigma) is obtained, a picture corresponding to the ray is rendered; wherein (x, y, z) represents three-dimensional space coordinates,represents a view angle; (R, G, B) representsRGB color values, σ, denote the voxel density.
Further, the step 2-2 is specifically realized as follows:
2-2-1, sampling 64 points for a ray, wherein each sampling point corresponds to one pointObtaining scene representation (R, G, B, sigma) of each sampling point through MLP;
2-2-2, rendering the ray by using a classical rendering formula to obtain a color value C (r) of a pixel corresponding to the ray, wherein the specific rendering process is as follows:
wherein r (t) is a formula of ray, and the value range of t is from a near-end point t n And a far end point t f I.e. the upper and lower limits of the integral; σ (r (t)) is the voxel density of the ray at point t, resulting from the results of the MLP prediction; c (r (t), d) is the RGB color value of the ray at point t, resulting from the prediction by MLP, where the notation d is the direction vector in ray o + td; t (T) is the transmittance of the ray at point T, and is obtained by integrating the voxel density sigma (r (T)), and the specific formula is Solving the RGB color value of a certain pixel point in the picture by using the prediction result of the MLP;
2-2-3. Select the L1 loss, i.e., the mean square error loss function, as follows:
L1=(RGB NeRF -RGB groundtruth ) 2
wherein, RGB NeRF Is the RGB color values of a picture rendered for a scene of a particular perspective using NeRF, RGB groundtruth Is the RGB color value of the real picture corresponding to that view.
Further, the steps 2 to 3 are specifically realized as follows.
The self-integration module sets and trains 8 completely independent models, respectively carries out 3D reconstruction on the same data set, and implicitly stores the same object or the same scene in different parameters of different models.
Furthermore, after each model trains 20000 epochs, 3D reconstruction is performed on the same view angle of the same scene respectively, so as to obtain the rendered RGB color value corresponding to the view angle.
Further, the loss of the rendered RGB color values and the RGB color values (grountrun) of the real picture is calculated and converted into psnr.
Further, the rendered RGB color values obtained by the 8 completely independent models are averaged to obtain the final RGB color value reconstructed under the target view angle.
Further, each completely independent model repeats the operations of the steps 1 and 2, and different visual angles of the same scene in the same data set are rendered to obtain RGB color values.
Furthermore, the self-integration module is mainly embodied in a model testing stage.
The beneficial effects of the invention are realized as follows:
the present invention is a self-integrating process on models based on NERF. The self-integration module can better extract the texture features of the picture, thereby improving the reconstruction precision. According to the invention, the self-integration module is added into the NERF nerve radiation field, so that the peak signal-to-noise ratio of an output image can be improved, and the image quality is improved. Specifically, the invention trains the model on the original NERF, tests the model on the test data sets, and compares the results after self-integration with the NERF and NERF variants, and finds that the self-integration proposed by the invention has good reconstruction effect. The peak signal-to-noise ratio of data set Rain100L increased from 25.73 to 26.73. These phenomena reflect the enhancement of the self-integration module proposed by us on the reconstruction of 3D new viewing angles, which can better restore the details of the scene.
Drawings
FIG. 1 is a schematic diagram of a NERF new view synthesis model MLP prediction based on self-integrated modules.
FIG. 2 is a schematic diagram of NERF new perspective composition model rendering and loss construction based on self-integration modules.
Figure 3 is a diagram of the new perspective composite model effect of NERF from integrated modules.
Detailed Description
The invention will be further explained with reference to the drawings.
As shown in fig. 1-3, the method for implementing the new NERF view synthesis model based on the self-integration module includes the following specific steps:
step 1: selecting image data
Selecting training data which are open source in the google drive by NERF authors, wherein the training data comprise generated music height data and scene data, and the music height data and the scene data are respectively stored in two folders. Take the category "hot dog" in the scene data as an example, which contains a "train/val/test" folder and three json files containing the camera pose at the time of each picture taking. Specifically, in the reading process, information such as image, transform _ matrix and the like is read from the json file by a function of "load _ folder _ data".
Example 1:
in our experiments, a batch size of 4096 rays was used, with each ray sampled at Nc =64 coordinates in the coarse volume and Nf =128 additional coordinates in the fine volume. Using Adam optimizer, learning rate is from 5 × 10 -4 Start and decay exponentially to 5 x 10 -5 During the optimization (other Adam hyper-parameters remain as default values β 1=0.9, β 2= 0.999). Optimization of a single scene typically requires about 100 –300k The number of iterations can converge on a single NVIDIA V100 GPU (approximately 1-2 days).
Step 2: constructing a NERF new visual angle synthetic model based on a self-integration module;
2-1 acquiring scene characterization of NERF
NERF uses a voxel-based characterization, fitting with MLP. As shown in fig. 1.
The input of the MLP is the corresponding three-dimensional space coordinates and view angle of the source imageThe output is (R, G, B, σ), representing RGB color values and voxel densities, respectively. And rendering a picture corresponding to the ray after obtaining the scene representation (R, G, B, sigma). Wherein (x, y, z) represents three-dimensional space coordinates,represents a view angle; (R, G, B) denotes RGB color values, and σ denotes voxel density.
Because of the limitation of MLP extracting high frequency information, adopting the skill of position coding in NLP field to inputAnd the function of a high frequency domain is easier to fit by carrying out position coding, and the performance is improved.
2-2 voxel rendering based on scene characterization
Voxel rendering is based on ray casting and is widely used in new view synthesis tasks. NERF itself is a method based on three-dimensional voxel rendering to convert the characterizing information of a scene into a 2D image. As shown in fig. 2.
2-2-1, first, sampling 64 points for a ray, each sampling point corresponding to oneObtaining scene representation (R, G, B, sigma) of each sampling point through MLP;
2-2-2, rendering the ray by using a classical rendering formula to obtain a color value C (r) of a pixel corresponding to the ray, wherein the specific rendering process is as follows:
wherein r (t) is a formula of ray, and the value range of t is from a near-end point t n And a far end point t f I.e. the upper and lower limits of the integral; σ (r (t)) is the voxel density of the ray at point t, resulting from the results of the MLP prediction; c (r (t), d) isThe RGB color values of the line at point t, resulting from the prediction by MLP, where the symbol d is the direction vector in ray o + td; t (T) is the transmittance of the ray at point T, and is obtained by integrating the voxel density sigma (r (T)), and the specific formula is Therefore, the prediction result of the MLP is utilized to solve the RGB color value of a certain pixel point in the picture.
2-2-3. Select the L1 loss, i.e., the mean square error loss function, as follows:
L1 loss=(RGB NeRF -RGB groundtruth ) 2
wherein, RGB NeRF Is picture color information, RGB, rendered for a scene of a particular perspective using NeRF groundtruth Is the true picture color information corresponding to that viewing angle. And continuously optimizing the NeRF model through gradient calculation and a gradient descent method, so that the quality of the NeRF rendered picture is improved.
And 2-3, constructing a self-integration module.
The self-integration module is mainly embodied in a model testing stage. As shown in fig. 3, the self-integration module sets and trains 8 completely independent models, respectively performs 3D reconstruction on the same data set, and implicitly stores the same object or the same scene in different parameters of different models.
After 20000 epochs are trained by each model, 3D reconstruction is carried out on the same visual angle of the same scene respectively, and the rendered RGB color value corresponding to the visual angle is obtained.
The loss of RGB color values after rendering and RGB color values (grountuth) of the real picture is calculated and converted into psnr.
And averaging the rendered RGB color values obtained by the 8 completely independent models to obtain the reconstructed final RGB color value under the target view angle. The method fully utilizes the randomness of the models and the strong information fusion capability of the ensemble learning to extract and fuse the reconstruction results of each model for the same task to obtain a surprising result.
Furthermore, each completely independent model repeats the operations of the steps 1 and 2, and different visual angles of the same scene in the same data set are rendered to obtain RGB color values.
Comparison of Algorithm Performance in step 3
Experiments were performed on the synthetic rendered data sets of the two objects. The DeepVoxels20 dataset contains four Lambertian objects with simple geometric shapes. Each object appears as 512 × 512 pixels from an upper-hemisphere sampled viewpoint. We render 100 views of each scene as input, 200 for testing, all of which are 800 x 800 pixels. The method is based on a self-integration process of the model on NERF. The integration module in the framework can better extract the texture features of the picture, thereby improving the reconstruction precision. The module is added to the NERF nerve radiation field, so that the peak signal-to-noise ratio of an output image can be improved, and the image quality is improved. In particular, we trained the model on the original NERF and tested the model on these test datasets and compared the results of our self-integration with NERF and NERF variants, as shown in table 1, and found that our proposed self-integration had a very good reconstruction effect. The peak signal-to-noise ratio of the data set increased from 25.7317 to 26.7329, i.e., we improved the performance by about one percent compared to the original NeRF. Although only one percentile, it is a huge step-over for the optimization of the original higher quality picture. Currently, the NeRF variants in the world, such as the Mip-NeRF, can only be improved by approximately 0.2 percentage points. Therefore, the self-integration module provided by the inventor has a quite large effect enhancement amplitude for 3D new visual angle reconstruction, can better restore the specific details of a scene, and has certain referential significance for promoting the progress of a 3D reconstruction task technology.
Table 1: NERF and NERF based on self-integrated modules for comparison
Method | PSNR | SSIM |
NERF | 25.7317 | 0.987 |
NERF based on self-integrated modules | 26.7329 | 0.992 |
。
Claims (10)
1. The method for realizing the NERF new visual angle synthesis model based on the self-integration module is characterized by comprising the following steps of:
step 1: selecting image data
Selecting training data open source in googledrive by a NERF author, wherein the training data comprises generated music height data and scene data, and the music height data and the scene data are respectively stored in two folders;
step 2: constructing a NERF new perspective synthesis model based on the self-integration module.
2. The method of claim 1, wherein step 2 is implemented as follows:
2-1, acquiring scene representation of NERF;
2-2, voxel rendering based on scene representation;
and 2-3, constructing a self-integration module.
3. The method according to claim 2, wherein step 2-1 is implemented as follows:
the input of the MLP is the corresponding three-dimensional space coordinates and view angle of the source imageThe output is (R, G, B, sigma), and after the scene representation (R, G, B, sigma) is obtained, a picture corresponding to the ray is rendered; wherein (x, y, z) represents three-dimensional space coordinates,represents a view angle; (R, G, B) represents RGB color values, and σ represents voxel density.
4. The method for implementing a new angle-of-view synthesis model of NERF based on self-integrated modules as claimed in claim 2, wherein step 2-2 is implemented as follows:
2-2-1, sampling a ray for 64 points, wherein each sampling point corresponds to oneObtaining scene representation (R, G, B, sigma) of each sampling point through MLP;
2-2-2, rendering the ray by using a classical rendering formula to obtain a color value C (r) of a pixel corresponding to the ray, wherein the specific rendering process is as follows:
wherein r (t) is a formula of ray, and the value range of t is from a near-end point t n And a far end point t f I.e. the upper and lower limits of the integral; σ (r (t)) is the voxel density of a ray at point t, derived from the results of MLP prediction; c (r (t), d) is the RGB color value of the ray at point t, resulting from the prediction by MLP, where the notation d is the direction vector in ray o + td; t (T) is the transmittance of the ray at point T, and is obtained by integrating the voxel density sigma (r (T)), and the specific formula is Solving the RGB color value of a certain pixel point in the picture by using the prediction result of the MLP;
2-2-3. Select the L1 loss, i.e., the mean square error loss function, as follows:
L1=(RGB NeRF -RGB groundtruth ) 2
wherein, RGB NeRF Is the RGB color values of a picture rendered for a scene of a particular perspective using NeRF, RGB groundtruth Is the RGB color value of the real picture corresponding to that view.
5. The method of claim 2 wherein step 2-3 is embodied as follows.
8 completely independent models are set and trained by the self-integration module, 3D reconstruction is carried out on the same data set, and the same object or the same scene is stored in different parameters of different models in an implicit mode.
6. The method of claim 5, wherein after 20000 epochs are trained for each new NERF view synthesis model, each new NERF view synthesis model performs 3D reconstruction on the same view of the same scene to obtain the rendered RGB color values corresponding to the view.
7. The method of claim 6 in which the loss of rendered RGB color values from real picture RGB color values (grountruth) is calculated and converted to psnr.
8. The method of claim 7 wherein the rendered RGB color values from 8 fully independent models are averaged to obtain the final RGB color value reconstructed at the target view.
9. The method of claim 7 in which each completely independent model repeats the operations of steps 1 and 2, rendering different views of the same scene in the same dataset to obtain RGB color values.
10. The method of claim 7 wherein the self-integrating module is embodied primarily in the model testing phase.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210911889.3A CN115457192A (en) | 2022-07-29 | 2022-07-29 | Method for realizing NERF new visual angle synthetic model based on self-integration module |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210911889.3A CN115457192A (en) | 2022-07-29 | 2022-07-29 | Method for realizing NERF new visual angle synthetic model based on self-integration module |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115457192A true CN115457192A (en) | 2022-12-09 |
Family
ID=84296242
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210911889.3A Pending CN115457192A (en) | 2022-07-29 | 2022-07-29 | Method for realizing NERF new visual angle synthetic model based on self-integration module |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115457192A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115731336A (en) * | 2023-01-06 | 2023-03-03 | 粤港澳大湾区数字经济研究院(福田) | Image rendering method, image rendering model generation method and related device |
-
2022
- 2022-07-29 CN CN202210911889.3A patent/CN115457192A/en active Pending
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115731336A (en) * | 2023-01-06 | 2023-03-03 | 粤港澳大湾区数字经济研究院(福田) | Image rendering method, image rendering model generation method and related device |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Mildenhall et al. | Nerf: Representing scenes as neural radiance fields for view synthesis | |
Skorokhodov et al. | Epigraf: Rethinking training of 3d gans | |
CN116958453B (en) | Three-dimensional model reconstruction method, device and medium based on nerve radiation field | |
WO2022198684A1 (en) | Methods and systems for training quantized neural radiance field | |
Yuan et al. | Make encoder great again in 3d gan inversion through geometry and occlusion-aware encoding | |
Li et al. | Symmnerf: Learning to explore symmetry prior for single-view view synthesis | |
CN114429538A (en) | Method for interactively editing nerve radiation field geometry | |
Karimpouli et al. | Multistep Super Resolution Double-U-net (SRDUN) for enhancing the resolution of Berea sandstone images | |
CN115457192A (en) | Method for realizing NERF new visual angle synthetic model based on self-integration module | |
Kasten et al. | Point cloud completion with pretrained text-to-image diffusion models | |
CN110322548B (en) | Three-dimensional grid model generation method based on geometric image parameterization | |
Debbagh | Neural Radiance Fields (NeRFs): A Review and Some Recent Developments | |
El Seoud et al. | A comprehensive review on volume rendering techniques | |
CN105321205B (en) | A kind of parameterized human body model method for reconstructing based on sparse key point | |
Zimny et al. | Multiplanenerf: Neural radiance field with non-trainable representation | |
Xie et al. | Multi-head mutual self-attention generative adversarial network for texture synthesis | |
Song et al. | City-on-Web: Real-time Neural Rendering of Large-scale Scenes on the Web | |
Li et al. | Video frame interpolation based on multi-scale convolutional network and adversarial training | |
Feng et al. | ZS-SRT: An efficient zero-shot super-resolution training method for Neural Radiance Fields | |
Chen et al. | Single-view Neural Radiance Fields with Depth Teacher | |
Hahlbohm et al. | PlenopticPoints: Rasterizing Neural Feature Points for High-Quality Novel View Synthesis. | |
Colom et al. | 3D shape reconstruction from non-realistic multiple-view depictions using NVDiffRec | |
Park et al. | Bridging Implicit and Explicit Geometric Transformation for Single-Image View Synthesis | |
KR102648938B1 (en) | Method and apparatus for 3D image reconstruction based on few-shot neural radiance fields using geometric consistency | |
Perche et al. | StyleDEM: A versatile model for authoring terrains |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |