CN110288695A - Single-frame images threedimensional model method of surface reconstruction based on deep learning - Google Patents

Single-frame images threedimensional model method of surface reconstruction based on deep learning Download PDF

Info

Publication number
CN110288695A
CN110288695A CN201910509313.2A CN201910509313A CN110288695A CN 110288695 A CN110288695 A CN 110288695A CN 201910509313 A CN201910509313 A CN 201910509313A CN 110288695 A CN110288695 A CN 110288695A
Authority
CN
China
Prior art keywords
model
layer
deep learning
threedimensional model
frame images
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910509313.2A
Other languages
Chinese (zh)
Other versions
CN110288695B (en
Inventor
杨路
杨经纶
李佑华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Electronic Science and Technology of China
Original Assignee
University of Electronic Science and Technology of China
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by University of Electronic Science and Technology of China filed Critical University of Electronic Science and Technology of China
Priority to CN201910509313.2A priority Critical patent/CN110288695B/en
Publication of CN110288695A publication Critical patent/CN110288695A/en
Application granted granted Critical
Publication of CN110288695B publication Critical patent/CN110288695B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/04Indexing scheme for image data processing or generation, in general involving 3D image data

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Computer Graphics (AREA)
  • General Engineering & Computer Science (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Geometry (AREA)
  • Architecture (AREA)
  • Computer Hardware Design (AREA)
  • Image Analysis (AREA)

Abstract

The present invention provides a kind of single-frame images threedimensional model method of surface reconstruction based on deep learning generates the single-frame images under model true shape point cloud and different points of view different distance comprising steps of being sampled and being rendered to CAD model;Feature extraction is carried out to image by convolutional neural networks, obtains the high-level semantic of two dimensional image;Control point coordinates required for the high-level semantic of acquisition is changed by full Connection Neural Network module converter for three-dimensional reconstruction stage nurbs surface and weighting parameter;It is updated using NURBS model of the obtained control point coordinates and weighting parameter to initialization, gradually carries out three-dimensional reconstruction;Training sample input deep learning model is trained automatically, obtains optimal model parameters, complete three-dimensional reconstruction by training deep learning model.The present invention simply and efficiently can carry out three-dimensional reconstruction to single-frame images, and the threedimensional model of reconstruction has the characteristics that details is abundant, surface is smooth and whole good.

Description

Single-frame images threedimensional model method of surface reconstruction based on deep learning
Technical field
The present invention relates to computer vision field more particularly to a kind of single-frame images threedimensional model tables based on deep learning Face method for reconstructing.
Background technique
Three-dimensional reconstruction refers to the process of the founding mathematical models in a computer of the three-dimension object in real scene, is to calculate One popular research direction of machine visual field.Compared to two dimensional image, threedimensional model is capable of providing the Global Information of object, more Add comprehensive displaying object properties, thus is suffered from widely in multiple fields such as computer animation, human-computer interaction, modern medicines Using.
With the fast development of deep learning, single-frame images three-dimensional reconstruction field achieves new breakthrough.Researcher utilizes Convolutional neural networks carry out feature extraction to single-frame images and rebuild specific expression-form by specific Feature Mapping method Threedimensional model.However between different expression-forms, there is some difference, accuracy, robustness and algorithm complexity all shadows Ring the generation effect of threedimensional model.The expression-form of threedimensional model mainly has voxel, point cloud and grid at present.
Voxel expression-form is similar with the two-dimensional pixel of image, only expands to three-dimensional stand from two-dimensional pixel Cube unit.But in three dimensions, voxel resolution promotion be by cube as unit of it is increased, this will lead to algorithm complexity Degree increases rapidly.Since the memory of current computer limits, the reconstruction of high-resolution three-dimension voxel model is very difficult.
Point cloud expression-form completes three-dimensional reconstruction by way of directly returning three-dimensional space point coordinate, and method is simple, It is easily achieved.But connect with point without part due to putting, point cloud effect when indicating continuous threedimensional model is bad, easy Lead to loss in detail, the threedimensional model of reconstruction usually lacks surface fluency.
Grid expression-form rebuilds threedimensional model by the connection on vertex and side.The expression-form can effectively avoid a little The excessively high problem of freedom degree between point is capable of forming stable, details threedimensional model abundant.But it in the conventional method, pushes up The acceptance region of point is relatively narrow, it means that a vertex cannot effectively receive the feature on other vertex from multiple sides, be easy Vertex is caused to fall into local optimum, the whole inadequate robust of the threedimensional model of reconstruction.
Nurbs surface is a kind of common expression-form in conventional three-dimensional modeling field, and there is good convexity and geometry to connect Continuous property can generate most of basic three-dimension curved surface.Pass through the variation of control point coordinates and weight, the expression-form adjustable three The local shape of curved surface is tieed up, and then preferably completes reconstructing three-dimensional model work.But traditional control point weight and coordinate Setting needs to calculate by a large amount of initialization and manual fine-tuning, this makes the reconstruction of complex-curved threedimensional model more difficult.
Summary of the invention
Technical problem to be solved by the invention is to provide a kind of single-frame images threedimensional model surface based on deep learning Method for reconstructing simply and efficiently can carry out three-dimensional reconstruction to single-frame images, and the threedimensional model of reconstruction is abundant with details, surface Smoothness, whole good feature.
In order to solve the above technical problems, the technical solution adopted by the present invention is that:
A kind of single-frame images threedimensional model method of surface reconstruction based on deep learning, comprising the following steps:
Step 1: sampling CAD model and rendered, and model true shape point cloud and different points of view different distance are generated Under single-frame images;
Step 2: feature extraction is carried out to image by convolutional neural networks, obtains the high-level semantic of two dimensional image;
Step 3: by the high-level semantic of acquisition by full Connection Neural Network module converter be three-dimensional reconstruction stage NURBS Control point coordinates required for curved surface changes and weighting parameter;
Step 4: it is updated using NURBS model of the obtained control point coordinates and weighting parameter to initialization, gradually Carry out three-dimensional reconstruction;
Step 5: training sample input deep learning model is trained automatically, obtains optimal mould by training deep learning model Shape parameter completes three-dimensional reconstruction.
Further, in step 1, CAD model is sampled and is rendered using OpenGL, generate training sample.
Further, in step 2, high level is carried out using identical VGG16 model to the image of each training sample Extraction of semantics carries out feature extraction by input picture of the following formula to each training sample:
In formula, N is positive integer;It indicates in n-th of classification in i-th of CAD training sample produced by jth picture High-level semantic;ExtractVGG16Indicate VGG16 feature extraction network;Indicate i-th of CAD training sample in n-th of classification Middle jth picture.
Further, in step 2, ExtractVGG16Including input layer, convolutional layer, pond layer and output layer, to volume Lamination exports result and carries out non-liner revision;Relationship such as following formula between convolutional layer or between convolutional layer and input layer:
In formula, Conv (i, j) indicates the data of i row j column in kth convolutional layer, Wk-1,k(m, n) indicates that -1 convolutional layer of kth arrives The data that m row n is arranged in the convolution kernel of kth convolutional layer, bk-1,kIndicate linear list of -1 hidden layer of kth to k-th of hidden layer The link biasing of member;I indicates the input picture of input layer, Wi,k(m, n) indicates linear unit of the input layer to k-th of hidden layer Link weight, bi,kIndicate that input layer is biased to the link of the linear unit of k-th of hidden layer;ReLU is defeated to every layer of convolutional layer Result carries out non-liner revision, formula out are as follows:
The following formula of relationship between the input and output of pond layer:
Pool (i, j)=maxm,nI(i+m,i+n)
Pool is the output of pond layer as a result, I is input grapheme, and pond layer takes the maximum value of input data regional area.
Further, the step 4 specifically:
NURBS model initialization: as effective object, there is coordinate (x at each control point at initialization control pointi,yi,zi) With weight w;
NURBS reconstructing three-dimensional model: control point coordinates required for being changed using obtained NURBS and weight parameter are to first Beginning model is updated, and obtains target three-dimensional;Initialization is shown below with model modification:
Wherein, C (μ, v) represents shape function in each section;wiRepresent the weight at control point, CiThat is (xi,yi,zi) represent The coordinate at control point;Ni,d(μ) represents interval function, μ, and v represents the section node of curved surface different directions, and n represents section sum, The corresponding section n+1 node, d represent the order of interval function.
Further, the step 5 specifically:
In the forward propagation process, network convolution kernel and characteristic pattern carry out dot product calculating, gradually obtain high-level semantic, Quan Lian It connects neural network module and parameter needed for obtaining phase of regeneration is returned to high-level semantic, complete the control node coordinate of NURBS With right value update;In back-propagation process, mask convolution core and full articulamentum undated parameter;
Neural network is trained using the threedimensional model of generation and the chamfering distance of true three-dimension model, that is, under passing through Formula is trained neural network:
In formula, loss represents loss function, LCDRepresent seek generate threedimensional model and true three-dimension model chamfering away from From, P represents the threedimensional model generated, and Q represents true shape threedimensional model, | | | | 2 indicate two norms.
Compared with prior art, the beneficial effects of the present invention are: 1) being based on deep learning neural network, training study is completed Three-dimensional reconstruction task improves the efficiency of three-dimensional reconstruction, reduces operation difficulty;2) NURBS conventional three-dimensional modeling method is combined, By the parameter learning to control point coordinates and weight, the recurrence complexity of neural network can be greatly reduced, reduces network Scale promotes network operation speed;3) mathematical constraint of NURBS makes the threedimensional model details rebuild abundant, and surface is smooth, whole Body works well.
Detailed description of the invention
Fig. 1 is the work flow diagram of the embodiment of the present invention;
Fig. 2 is that the training sample of the embodiment of the present invention generates schematic diagram;
Fig. 3 is the data transmission scheme of the image characteristics extraction network of the embodiment of the present invention;
Fig. 4 is the schematic diagram that the image information of the embodiment of the present invention is converted into the control coordinate and weight of nurbs surface;
Fig. 5 is that the initial nurbs surface model gradual deformation of the embodiment of the present invention is the schematic diagram of target three-dimensional.
Specific embodiment
The present invention will be further described in detail below with reference to the accompanying drawings and specific embodiments.The present invention passes through depth It practises neural network and seeks control point coordinates and weight parameter value required for nurbs surface changes, on the one hand, deep learning nerve Network can use its powerful ability in feature extraction and analytical calculation ability, required for the variation of automatic returning nurbs surface Parameter;On the other hand, NURBS method does not need to carry out recurrence calculating to each point of threedimensional model, it is only necessary to seek The weight and coordinate at control point, so that it may realize three-dimensional reconstruction, reduce network parameter, reduce computation complexity.
As shown in Figure 1, the single-frame images threedimensional model method of surface reconstruction based on deep learning includes the following steps:
Step 1: training sample is generated
Sampling is carried out to CAD model using OpenGL and renders generation training sample, the nerve for reconstructing three-dimensional model Network training.Certainly, the method for generating training sample is not limited to OpenGL, and any method that can be realized same technique effect is all It can be used for generating training sample.It is specific as follows:
Select CAD data collection.2015, Princeton University, Stanford University and TTIC researcher issued cooperation Project ShapeNet data set, it is intended to which Yu Jianli mono- has abundant 3D graphic data set annotate, large-scale, for all over the world Researcher these data are provided, to support computer graphics, computer vision, robot technology and other related disciplines Research.The present invention uses a subset of ShapeNet data set, and it is one that it, which includes 50,000 kinds of models and 13 primary categories, The three-dimensional CAD model of a set.Points Sample is carried out to CAD model using OpenGL, obtains mould shapes.It is regarded from initial coordinate Angle is set out, and is sampled using OpenGL to CAD model, each model acquire 4096 points, available CAD model it is true Real shape information, as shown in the left side Fig. 2.
24 width single-frame images are obtained with different distance rendering CAD model from different perspectives using OpenGL, and record every The corresponding camera angle of picture and distance.On OpenGL platform, by mobile virtual camera to the angle of CAD model and away from From the picture of available different points of view is as training sample, as shown in the right side of fig 2.
Data set is divided into trained and test set, wherein 4/5 for training, remaining is 1/5 for testing.In order to facilitate instruction Practice, the present embodiment upsets obtained sample data set, and each classification extracts 4/5 and collectively constitutes training dataset;Remaining 1/5 Data independent storage of all categories, for testing the modelling effect of each type.
Step 2: image characteristics extraction
Feature extraction is carried out to the input picture of each training sample, obtains the semantic feature of input picture.Due to each CAD model has 24 subtended angle degree and uses phase respectively from apart from different input pictures, therefore to 24 pictures of each sample Same VGG16 model carries out the extraction of high-level semantic.Feature can be carried out by input picture of the following formula to each training sample It extracts:
Wherein, N is positive integer, and N takes 13 in the present embodiment, indicates 13 training data classifications;Indicate n-th of class High-level semantic caused by jth picture in i-th of CAD training sample in not, j maximum takes 24 in the present embodiment; ExtractVGG16Indicate VGG16 feature extraction network;Indicate that jth is schemed in i-th of CAD training sample in n-th of classification Piece.
ExtractVGG16Including input layer, convolutional layer, pond layer and output layer.Convolutional layer is 15 in the embodiment of the present invention It is a, as shown in Figure 3.A indicates that input layer, B1 indicate that the first convolutional layer, B15 indicate that the 15th convolutional layer, C1 indicate the first pond Layer, C15 indicate that the 15th pond layer, D indicate output layer.Each of first convolutional layer unit need to only experience input layer part Image-region is experienced without doing to global image, and the second convolutional layer can experience the first convolutional layer regional area again, according to this class It pushes away, in higher convolutional layer, these are experienced different local neurons integrating can be obtained by global information.In order to add The convergence of fast network model and the accuracy for improving result carry out non-liner revision to convolutional layer output result.Convolutional layer it Between or convolutional layer and input layer between relationship such as following formula:
In formula, Conv (i, j) indicates the data of i row j column in kth convolutional layer, Wk-1,k(m, n) indicates that -1 convolutional layer of kth arrives The data that m row n is arranged in the convolution kernel of kth convolutional layer, bk-1,kIndicate linear list of -1 hidden layer of kth to k-th of hidden layer The link biasing of member;I indicates the input picture of input layer, Wi,k(m, n) indicates linear unit of the input layer to k-th of hidden layer Link weight, bi,kIndicate that input layer is biased to the link of the linear unit of k-th of hidden layer;ReLU is defeated to every layer of convolutional layer Result carries out non-liner revision, formula out are as follows:
When semantic layer makes a small amount of translation, the semantic information approximation that pond layer can help convolutional layer to obtain is constant, pond Change the following formula of relationship between the input and output of layer:
Pool (i, j)=maxm,nI(i+m,i+n)
Pool is the output of pond layer as a result, I is input grapheme, and pond layer takes the maximum value of input data regional area.
Step 3: characteristics of image conversion
The picture high-level semantic extracted is returned using full Connection Neural Network module, exports the three-dimensional reconstruction stage Control point coordinates required for nurbs surface changes and weighting parameter.Full Connection Neural Network module includes l layers, defaults l=1, 2,3, suitable for simple, common, complicated reconstructing three-dimensional model task.Convolution obtains high-level semantics features firstly the need of expanding into Then column vector is connected with full articulamentum, is unfolded as follows formula:
Wherein,Vector is unfolded in high-level semantics to extract from input picture, and Conv is the output multi-dimensional matrix of convolutional layer One multi-dimensional matrix is converted into a column vector by semanteme, Flatten ().
Connection such as following formula between full articulamentum and high-level semantics expansion vector and full articulamentum:
Para indicates that output parameter, i indicate input channel, and j indicates output channel,Indicate that vector is unfolded in high-level semantics Or a upper full articulamentum, b (j) represent the link biasing of jth linear unit.
The coordinate and weight parameter at control point required for output parameter dimension changes with nurbs surface match, as follows Formula:
Para(j)→Numct*((x,y,z)+w)
Wherein, NumctThe quantity at control point is represented, (x, y, z) represents the coordinate at control point, and w represents the weight at control point.
Step 4: reconstructing three-dimensional model
The present invention does not directly generate threedimensional model, but will be first with weight parameter by the control point coordinates that conversion obtains Beginning threedimensional model is gradually deformed into target three-dimensional, this is a kind of mapping of three-dimensional to three-dimensional, and it is higher accurate to have Degree and globality.
Step 4 specifically includes:
NURBS model initialization: this example is initialized 7*9 control point as effective object, and each control point has Coordinate (x, y, z) and weight w, as shown in Figure 4.
NURBS reconstructing three-dimensional model: control point coordinates required for being changed using obtained NURBS and weight parameter are to first Beginning model is updated, and gradually obtains target three-dimensional.In the present embodiment, initialization is shown below with model modification:
Wherein, C (μ, v) represents shape function in each section;wiRepresent the weight at control point, CiThat is (xi,yi,zi) represent The coordinate at control point;Ni,d(μ) represents interval function, μ, and v represents the section node of curved surface different directions, and n represents section sum, The corresponding section n+1 node, d represent the order of interval function.
Step 5: deep learning model training
Training set is fully entered into neural network, entire depth learning model is trained.In the forward propagation process, Network convolution kernel and characteristic pattern carry out dot product calculating, gradually obtain high-level semantic, full Connection Neural Network module is to high-level semantic Parameter needed for obtaining phase of regeneration is returned, the control node coordinate and right value update of NURBS are completed;In back-propagation process In, mask convolution core and full articulamentum undated parameter.
Neural network is trained using the threedimensional model of generation and the chamfering distance of true three-dimension model.For each Point, chamfering distance algorithm find nearest adjoint point in another set, and by square summation of distance.The algorithm is as set The function of midpoint, be it is continuous, piecewise smooth, the range searching to each point is independent, therefore can be easy to Ground parallelization.Neural network is trained by following formula:
Wherein, loss represents loss function, LCDRepresent seek generate threedimensional model and true three-dimension model chamfering away from From, P represents the threedimensional model generated, and Q represents true shape threedimensional model, | | | | 2 indicate two norms.
Single-frame images threedimensional model method of surface reconstruction provided by the invention based on deep learning, is different from traditional NURBS three-dimensional rebuilding method and the common deep learning three-dimensional rebuilding method based on voxel, point cloud and grid.The present invention is not necessarily to Various parameters required for NURBS reconstructing three-dimensional model are calculated, therefore need a large amount of manual operation unlike conventional three-dimensional models, Which greatly improves the efficiency of reconstructing three-dimensional model;The sky of point of the present invention without individually predicting each threedimensional model simultaneously Between coordinate, solve model complicated difficult to calculate, the problem of precision is difficult to improve, stronger enhances the smoothness of threedimensional model Degree, ensure that the details integrality of reconstruction model, improves the precision and speed of three-dimensional reconstruction.
Further, since the single-frame images threedimensional model method of surface reconstruction based on deep learning is a kind of based on statistical learning Machine learning method, need a large amount of training sample, that is, demarcated the point cloud data of the shape information of CAD model from it is different The single frames picture of angle and distance.By obtaining the priori knowledge in relation to mould shapes and viewpoint from single frames picture, depth is obtained The weight parameter of learning model is spent, and then obtains the deep neural network of this single-frame images nurbs surface reconstructing three-dimensional model Model.For this reason, the present invention is sampled using OpenGL and rendering tool, flexible progress data generation increase instruction Practice sample size, and then improves the reconstruction ability of deep neural network model.

Claims (6)

1. a kind of single-frame images threedimensional model method of surface reconstruction based on deep learning, which comprises the following steps:
Step 1: sampling CAD model and rendered, and generates under model true shape point cloud and different points of view different distance Single-frame images;
Step 2: feature extraction is carried out to image by convolutional neural networks, obtains the high-level semantic of two dimensional image;
Step 3: by the high-level semantic of acquisition by full Connection Neural Network module converter be three-dimensional reconstruction stage nurbs surface Control point coordinates required for changing and weighting parameter;
Step 4: it is updated using NURBS model of the obtained control point coordinates and weighting parameter to initialization, is gradually carried out Three-dimensional reconstruction;
Step 5: training sample input deep learning model is trained automatically, obtains optimal models ginseng by training deep learning model Number completes three-dimensional reconstruction.
2. the single-frame images threedimensional model method of surface reconstruction based on deep learning as described in claim 1, which is characterized in that In step 1, CAD model is sampled and rendered using OpenGL, generates training sample.
3. the single-frame images threedimensional model method of surface reconstruction based on deep learning as claimed in claim 2, which is characterized in that In step 2, high-level semantic extraction is carried out using identical VGG16 model to the image of each training sample, that is, under passing through Formula carries out feature extraction to the input picture of each training sample:
In formula, N is positive integer;Indicate in n-th of classification height caused by jth picture in i-th of CAD training sample Layer is semantic;ExtractVGG16Indicate VGG16 feature extraction network;It indicates in n-th of classification in i-th of CAD training sample J picture.
4. the single-frame images threedimensional model method of surface reconstruction based on deep learning as claimed in claim 3, which is characterized in that In step 2, ExtractVGG16Including input layer, convolutional layer, pond layer and output layer, convolutional layer output result is carried out Non-liner revision;Relationship such as following formula between convolutional layer or between convolutional layer and input layer:
In formula, Conv (i, j) indicates the data of i row j column in kth convolutional layer, Wk-1,k(m, n) indicates that -1 convolutional layer of kth is rolled up to kth The data that m row n is arranged in the convolution kernel of lamination, bk-1,kIndicate linear unit from -1 hidden layer of kth to k-th of hidden layer chain Connect biasing;I indicates the input picture of input layer, Wi,k(m, n) indicate linear unit from input layer to k-th of hidden layer link Weight, bi,kIndicate that input layer is biased to the link of the linear unit of k-th of hidden layer;ReLU exports result to every layer of convolutional layer Carry out non-liner revision, formula are as follows:
The following formula of relationship between the input and output of pond layer:
Pool (i, j)=maxm,nI(i+m,i+n)
Pool is the output of pond layer as a result, I is that input is semantic, and pond layer takes the maximum value of input data regional area.
5. the single-frame images threedimensional model method of surface reconstruction based on deep learning as described in claim 1, which is characterized in that The step 4 specifically:
NURBS model initialization: as effective object, there are coordinate (x, y, z) and weight in each control point at initialization control point w;
NURBS reconstructing three-dimensional model: control point coordinates required for being changed using obtained NURBS and weight parameter are to introductory die Type is updated, and obtains target three-dimensional;Initialization is shown below with model modification:
Wherein, C (μ, v) represents shape function in each section;wiRepresent the weight at control point, CiThat is (xi,yi,zi) represent control The coordinate of point;Ni,d(μ) represents interval function, μ, and v represents the section node of curved surface different directions, and n represents section sum, corresponding n + 1 section node, d represent the order of interval function.
6. the single-frame images threedimensional model method of surface reconstruction based on deep learning as described in claim 1, which is characterized in that The step 5 specifically:
In the forward propagation process, network convolution kernel and characteristic pattern carry out dot product calculating, gradually obtain high-level semantic, full connection mind Parameter needed for obtaining phase of regeneration is returned to high-level semantic through network module, completes the control node coordinate and power of NURBS Value updates;In back-propagation process, mask convolution core and full articulamentum undated parameter;
Neural network is trained using the threedimensional model of generation and the chamfering distance of true three-dimension model, that is, passes through following formula pair Neural network is trained:
In formula, loss represents loss function, LCDRepresent the chamfering distance of the threedimensional model and true three-dimension model of seeking generating, P The threedimensional model generated is represented, Q represents true shape threedimensional model, | | | | 2 indicate two norms.
CN201910509313.2A 2019-06-13 2019-06-13 Single-frame image three-dimensional model surface reconstruction method based on deep learning Active CN110288695B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910509313.2A CN110288695B (en) 2019-06-13 2019-06-13 Single-frame image three-dimensional model surface reconstruction method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910509313.2A CN110288695B (en) 2019-06-13 2019-06-13 Single-frame image three-dimensional model surface reconstruction method based on deep learning

Publications (2)

Publication Number Publication Date
CN110288695A true CN110288695A (en) 2019-09-27
CN110288695B CN110288695B (en) 2021-05-28

Family

ID=68004750

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910509313.2A Active CN110288695B (en) 2019-06-13 2019-06-13 Single-frame image three-dimensional model surface reconstruction method based on deep learning

Country Status (1)

Country Link
CN (1) CN110288695B (en)

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110992435A (en) * 2019-11-06 2020-04-10 上海东软医疗科技有限公司 Image reconstruction method and device, and imaging data processing method and device
CN111598111A (en) * 2020-05-18 2020-08-28 商汤集团有限公司 Three-dimensional model generation method and device, computer equipment and storage medium
CN111651954A (en) * 2020-06-10 2020-09-11 嘉兴市像景智能装备有限公司 Method for three-dimensional reconstruction of SMT electronic component based on deep learning
CN111768452A (en) * 2020-06-30 2020-10-13 天津大学 Non-contact automatic mapping method based on deep learning
CN112116700A (en) * 2020-08-28 2020-12-22 北京航空航天大学 Monocular view-based three-dimensional reconstruction method and device
CN112150608A (en) * 2020-09-07 2020-12-29 鹏城实验室 Three-dimensional face reconstruction method based on graph convolution neural network
CN112348907A (en) * 2020-10-16 2021-02-09 佛山科学技术学院 Compression method and device for optical coherence tomography three-dimensional data
CN112634438A (en) * 2020-12-24 2021-04-09 北京工业大学 Single-frame depth image three-dimensional model reconstruction method and device based on countermeasure network
CN112884878A (en) * 2019-11-29 2021-06-01 北京航空航天大学 Method for displaying cumulus cloud three-dimensional model
CN113096234A (en) * 2019-12-23 2021-07-09 复旦大学 Method and device for generating three-dimensional grid model by using multiple color pictures
WO2021179593A1 (en) * 2020-03-10 2021-09-16 山东大学 Deep learning-based three-dimensional pipeline reconstruction method, system, medium, and apparatus
CN113706676A (en) * 2021-08-26 2021-11-26 京东鲲鹏(江苏)科技有限公司 Model self-supervision training method and device for point cloud data
CN114049444A (en) * 2022-01-13 2022-02-15 深圳市其域创新科技有限公司 3D scene generation method and device
CN114078152A (en) * 2020-08-20 2022-02-22 北京瓦特曼科技有限公司 Robot carbon block cleaning method based on three-dimensional reconstruction

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102722906A (en) * 2012-05-23 2012-10-10 温州大学 Feature-based top-down image modeling method
CN105427293A (en) * 2015-11-11 2016-03-23 中国科学院深圳先进技术研究院 Indoor scene scanning reconstruction method and apparatus
CN106683181A (en) * 2017-01-06 2017-05-17 厦门大学 Method for reconstructing three-dimensional human body dense surface motion field
CN107403430A (en) * 2017-06-15 2017-11-28 中山大学 A kind of RGBD image, semantics dividing method
CN107862741A (en) * 2017-12-10 2018-03-30 中国海洋大学 A kind of single-frame images three-dimensional reconstruction apparatus and method based on deep learning

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102722906A (en) * 2012-05-23 2012-10-10 温州大学 Feature-based top-down image modeling method
CN105427293A (en) * 2015-11-11 2016-03-23 中国科学院深圳先进技术研究院 Indoor scene scanning reconstruction method and apparatus
CN106683181A (en) * 2017-01-06 2017-05-17 厦门大学 Method for reconstructing three-dimensional human body dense surface motion field
CN107403430A (en) * 2017-06-15 2017-11-28 中山大学 A kind of RGBD image, semantics dividing method
CN107862741A (en) * 2017-12-10 2018-03-30 中国海洋大学 A kind of single-frame images three-dimensional reconstruction apparatus and method based on deep learning

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
YANGLU: "Robust spatial–temporal Bayesian view synthesis for video stitching with occlusion handling", 《MACHINE VISION AND APPLICATIONS VOLUME》 *
史利民等: "基于语义交互的三维重建", 《计算机辅助设计与图形学学报》 *
李彦东: "基于卷积神经网络的计算机视觉关键技术研究", 《中国博士学位论文全文数据库信息科技辑》 *
陈正鸣等: "曲面特征技术研究的现状及展望", 《河海大学常分校学报》 *

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110992435A (en) * 2019-11-06 2020-04-10 上海东软医疗科技有限公司 Image reconstruction method and device, and imaging data processing method and device
CN110992435B (en) * 2019-11-06 2023-10-20 上海东软医疗科技有限公司 Image reconstruction method and device, imaging data processing method and device
CN112884878A (en) * 2019-11-29 2021-06-01 北京航空航天大学 Method for displaying cumulus cloud three-dimensional model
CN113096234A (en) * 2019-12-23 2021-07-09 复旦大学 Method and device for generating three-dimensional grid model by using multiple color pictures
WO2021179593A1 (en) * 2020-03-10 2021-09-16 山东大学 Deep learning-based three-dimensional pipeline reconstruction method, system, medium, and apparatus
CN111598111A (en) * 2020-05-18 2020-08-28 商汤集团有限公司 Three-dimensional model generation method and device, computer equipment and storage medium
CN111598111B (en) * 2020-05-18 2024-01-05 商汤集团有限公司 Three-dimensional model generation method, device, computer equipment and storage medium
CN111651954A (en) * 2020-06-10 2020-09-11 嘉兴市像景智能装备有限公司 Method for three-dimensional reconstruction of SMT electronic component based on deep learning
CN111651954B (en) * 2020-06-10 2023-08-18 嘉兴市像景智能装备有限公司 Method for reconstructing SMT electronic component in three dimensions based on deep learning
CN111768452A (en) * 2020-06-30 2020-10-13 天津大学 Non-contact automatic mapping method based on deep learning
CN111768452B (en) * 2020-06-30 2023-08-01 天津大学 Non-contact automatic mapping method based on deep learning
CN114078152B (en) * 2020-08-20 2023-05-02 北京瓦特曼科技有限公司 Robot carbon block cleaning method based on three-dimensional reconstruction
CN114078152A (en) * 2020-08-20 2022-02-22 北京瓦特曼科技有限公司 Robot carbon block cleaning method based on three-dimensional reconstruction
CN112116700B (en) * 2020-08-28 2022-09-16 北京航空航天大学 Monocular view-based three-dimensional reconstruction method and device
CN112116700A (en) * 2020-08-28 2020-12-22 北京航空航天大学 Monocular view-based three-dimensional reconstruction method and device
CN112150608A (en) * 2020-09-07 2020-12-29 鹏城实验室 Three-dimensional face reconstruction method based on graph convolution neural network
CN112348907A (en) * 2020-10-16 2021-02-09 佛山科学技术学院 Compression method and device for optical coherence tomography three-dimensional data
CN112348907B (en) * 2020-10-16 2024-03-26 佛山科学技术学院 Compression method and device for optical coherence tomography three-dimensional data
CN112634438A (en) * 2020-12-24 2021-04-09 北京工业大学 Single-frame depth image three-dimensional model reconstruction method and device based on countermeasure network
CN113706676A (en) * 2021-08-26 2021-11-26 京东鲲鹏(江苏)科技有限公司 Model self-supervision training method and device for point cloud data
CN113706676B (en) * 2021-08-26 2024-01-16 京东鲲鹏(江苏)科技有限公司 Model self-supervision training method and device for point cloud data
CN114049444A (en) * 2022-01-13 2022-02-15 深圳市其域创新科技有限公司 3D scene generation method and device

Also Published As

Publication number Publication date
CN110288695B (en) 2021-05-28

Similar Documents

Publication Publication Date Title
CN110288695A (en) Single-frame images threedimensional model method of surface reconstruction based on deep learning
CN112396703B (en) Reconstruction method of single-image three-dimensional point cloud model
CN109147048B (en) Three-dimensional mesh reconstruction method by utilizing single-sheet colorful image
CN108921926B (en) End-to-end three-dimensional face reconstruction method based on single image
CN101404091B (en) Three-dimensional human face reconstruction method and system based on two-step shape modeling
Neubert et al. Approximate image-based tree-modeling using particle flows
CN103021017B (en) Three-dimensional scene rebuilding method based on GPU acceleration
CN110458939A (en) The indoor scene modeling method generated based on visual angle
CN111862101A (en) 3D point cloud semantic segmentation method under aerial view coding visual angle
CN111899328B (en) Point cloud three-dimensional reconstruction method based on RGB data and generation countermeasure network
Shen et al. Deepsketchhair: Deep sketch-based 3d hair modeling
CN111028335B (en) Point cloud data block surface patch reconstruction method based on deep learning
CN107481313A (en) A kind of dense three-dimensional object reconstruction method based on study available point cloud generation
CN110349247A (en) A kind of indoor scene CAD 3D method for reconstructing based on semantic understanding
CN111583408B (en) Human body three-dimensional modeling system based on hand-drawn sketch
CN103530907A (en) Complicated three-dimensional model drawing method based on images
CN104778736A (en) Three-dimensional garment animation generation method driven by single video content
Sharma et al. Point cloud upsampling and normal estimation using deep learning for robust surface reconstruction
Eyiyurekli et al. Interactive free-form level-set surface-editing operators
CN115272608A (en) Human hand reconstruction method and equipment
Kamra et al. Lightweight reconstruction of urban buildings: Data structures, algorithms, and future directions
CN113808006B (en) Method and device for reconstructing three-dimensional grid model based on two-dimensional image
He et al. Sketch2cloth: Sketch-based 3d garment generation with unsigned distance fields
Pellis et al. Synthetic data generation and testing for the semantic segmentation of heritage buildings
Hou et al. Depth estimation and object detection for monocular semantic SLAM using deep convolutional network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant