CN109462747A - Based on the DIBR system gap filling method for generating confrontation network - Google Patents

Based on the DIBR system gap filling method for generating confrontation network Download PDF

Info

Publication number
CN109462747A
CN109462747A CN201811512179.3A CN201811512179A CN109462747A CN 109462747 A CN109462747 A CN 109462747A CN 201811512179 A CN201811512179 A CN 201811512179A CN 109462747 A CN109462747 A CN 109462747A
Authority
CN
China
Prior art keywords
empty
image
masking
pixel
target image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811512179.3A
Other languages
Chinese (zh)
Other versions
CN109462747B (en
Inventor
刘然
赵洋
肖迪
郑杨婷
刘亚琼
陈希
张艳珍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shaanxi Hengsheng Bolong Junxin Technology Co ltd
Original Assignee
"CHENGDU MEILYU TECHNOLOGY Co LTD"
Chongqing University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by "CHENGDU MEILYU TECHNOLOGY Co LTD", Chongqing University filed Critical "CHENGDU MEILYU TECHNOLOGY Co LTD"
Priority to CN201811512179.3A priority Critical patent/CN109462747B/en
Publication of CN109462747A publication Critical patent/CN109462747A/en
Application granted granted Critical
Publication of CN109462747B publication Critical patent/CN109462747B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Abstract

The invention discloses a kind of based on the DIBR system gap filling method for generating confrontation network, a collection of reference picture sample and corresponding depth image are collected first, generate training sample, it is trained using this training sample to confrontation network is generated, obtain the generation model filled for DIBR system cavity, then the corresponding tensor of target image to be filled is generated using method identical with training sample, and it generates prospect and inhibits masking-out, processing obtains input tensor, it is input to generation model, finally output result is further processed to obtain the filled target image in cavity.The present invention extracts low-level features and high-level semantics feature using confrontation network is generated, and proposes the consistency of highly filled hole region and remainder pixel.

Description

Based on the DIBR system gap filling method for generating confrontation network
Technical field
The invention belongs to 3-D image converter technique fields, more specifically, are related to a kind of based on generation confrontation network DIBR system gap filling method.
Background technique
View synthesis engine (view synthesis engine) is the important component of 3D display technology, and view closes At there are two main classes in engine view synthesizing method: stereo content generation method and Content_based image based on threedimensional model are raw At the method for stereo-picture, what is generallyd use at present is to draw (depth-image-based based on depth image Rendering, DIBR) algorithm.
Traditional 3D stereo display technique needs to transmit two-way or multichannel view, and generation continuously looks around (continuous Look-around), make one to experience stereoscopic effect.However, DIBR technology can only pass through video and its corresponding depth all the way Figure generates stereo-picture (stereoscopic image).Often there is empty (holes) using the view that DIBR technology synthesizes, The main reason for generating cavity is the visibility of object in view as the change of viewpoint is changed, such as in reference picture By the background object of foreground occlusion, become in the target image as it can be seen that therefore generate virtual visual point image (target image) exist Partial information content missing, to generate cavity.It is to be solved in DIBR technology for how reasonably and effectively filling these cavities Critical issue.In recent years, researcher proposes the method for many cavity fillings, these methods are broadly divided into following 4 class:
(1) depth map pre-processes
The pretreated basic thought of depth map is that cavity occur after reducing or even avoiding View synthesis.The depth in depth map Discontinuous place often generates cavity.If the depth value of two neighbor pixels in reference picture exists certain poor Different, then their displacements accordingly in 3-D image transformation are different, i.e., the two pixels are no longer adjacent in the target image. Depth in reference picture between two pixels is bigger, and empty width is also bigger.Depth image pretreatment is adjacent by reducing Depth difference between pixel reduces cavity, i.e., is first smoothed to depth map before 3-D image transformation.After smooth Depth map depth continuity it is more gentle, therefore generate cavity it is smaller.
Since depth map describes the geometry of scene, each change of depth map can all lead to transformed target The change of scene geometry in image, this, which means that, can generate geometric distortion in the view of synthesis.Since human eye is to vertical It is distorted more sensitive, therefore Zhang et al. is filtered depth image using asymmetric Gaussian filter, is hung down with weakening in background The geometric distortion phenomenon of straight grain information.Daribo et al. is only smooth in the adjacent edges application of depth map, in this way, only cavity Neighbouring region can be just smoothed, and other regions remain unchanged.Hong proposes direction gaussian filtering (DGF, directional Gaussian filter), in pretreatment deep figure, depth map is first expanded, then using asymmetric Gaussian filter to appearance The place in cavity is filtered.However, such methods existed general problem is that they can introduce geometric distortion, so that target figure The quality of picture reduces.
(2) interpolation method (Interpolation)
Gap filling method thought based on interpolation is fairly simple, and this method directly passes through the texture near cavity slotting The method of value travels to hole region.Method based on interpolation mainly includes horizontal interpolation (horizontal ) and background extrapolation (background extrapolation) interpolation.Horizontal interpolation method to hole region line by line Processing fills the empty pixel in the row by the linear interpolation between the non-empty pixel in left and right in the row.Background extrapolation side Method also handles filling cavity line by line, it identifies the non-empty pixel of the row or so first, then passes the pixel value of farther place pixel It is multicast in hole region.Since these interpolation methods are to every a line independent process in view, such methods will lead to row On visual artefacts.
(3) image repair (Image inpainting)
Gap filling method based on image repair depends on bigger reference zone.Image repair method mainly divides at present For two classes: one is the conventional method for being based on patch (patch), another kind is that the depth based on convolutional neural networks generates mould Type.Due to the similitude of cavity filling and image repair problem, many empty filling algorithms based on image repair are by classics Image repair algorithm improvement obtain, but have not yet to see the depth based on convolutional neural networks and generate model and be applied to DIBR In the filling of system cavity.Traditional reparation algorithm based on patch can effectively fill static texture, but non-static in object In the case of filling effect will be greatly reduced.
(4) the cavity filling based on space-time domain information
Above-mentioned gap filling method merely with a certain moment image spatial information, and be based on time-space domain (spatio- Temporal domain) method carry out cavity using the image information of time domain (temporal domain) upper different moments Filling.Hua et al. proposes a kind of drafting (Depth-Video-Based Rendering, DVBR) method based on deep video, This method has content-adaptive characteristic, and for by blocking the reason of, caused macroscopic-void was solved using the relativity of time domain of scene Certainly, but for the small cavity as caused by quantization error image restoration technology is still used.Lin et al. then proposes a kind of using subgraph Generate (sprite generation) algorithm gap filling method, subgraph generating algorithm can be to series of successive frames among Background information (including colored with depth map) merged, subgraph model is established, for empty in DIBR synthesis view Filling.However, the method based on time-space domain cannot be filled up completely cavity, there is still a need for other cavities in the cavity that cannot be filled Fill method is filled.
However, only used low-level features (low-level features) in the studies above, and can not capture higher The semanteme of grade, causes the empty filling effect under complex scene undesirable.
Summary of the invention
It is an object of the invention to overcome the deficiencies of the prior art and provide a kind of based on the DIBR system for generating confrontation network Gap filling method extracts low-level features and high-level semantics feature using confrontation network is generated, mention highly filled hole region and The consistency of remainder pixel.
For achieving the above object, the present invention is based on generate confrontation network DIBR system gap filling method include with Lower step:
S1: training sample is generated using following methods:
S1.1: a collection of reference picture is collectedAnd corresponding depth image Dt, t=1,2 ..., K, K expression reference picture number Amount;
S1.2: by every width reference pictureWith depth image Dt3-D image transformation is carried out, corresponding target image is obtainedAnd generate and target imageThe masking-out mask of identical sizet, masking-out masktFor marking empty pixel, if target ImageIn pixel be empty pixel, then in masking-out masktIn corresponding pixel value be 1, if it is non-empty pixel Point, then in masking-out masktIn corresponding pixel value be 0;
S1.3: by every width reference pictureEach access matrix and corresponding masking-out masktIt is separately converted to batch_ size×Mt×Nt× 1 tensor, then successively splicing obtains batch_size × Mt×NtThe tensor of × (L+1) is as correspondence Training sample, wherein batch_size indicates batch size, Mt×NtIndicate reference pictureSize, L indicate reference picturePort number;
S2: the training sample generated using step S1 is trained to confrontation network is generated, and is obtained empty for DIBR system The generation model of hole filling;
S3: for through reference picture I 'RCarry out the target image I ' to cavity filling that image three-dimensional convertsS, generate With target image I 'SThe non-empty matrix M ' of identical sizeSWith masking-out mask ', wherein non-cavity matrix M 'SFor distinguishing target Image I 'SIn non-empty pixel and empty pixel, for target image I 'SIn pixel be empty pixel, then non- Empty matrix M 'SIt is middle that corresponding element value is set to -1, whereas if being non-empty pixel, then in non-empty matrix M 'SIn Corresponding element value is set to the depth value of the pixel;Masking-out mask ' is for marking empty pixel, if target image I′SIn pixel be empty pixel, then be 1 by corresponding element value in masking-out mask ', if it is non-empty pixel Point, then in masking-out mask ', corresponding element value is set to 0 by;
S4: according to non-empty matrix M 'SIt carries out empty detection and the cavity detected is expanded, the matrix that will be obtained Inhibit masking-out mask ' as prospectFR
S5: using same method in step S1.3 by target image I 'SEach access matrix and masking-out mask ' turn respectively The tensor of batch_size × M ' × N ' × 1 is turned to, wherein M ' × N ' expression target image I 'SSize, then successively splice Batch_size × M ' × N ' × (L+1) tensor R is obtained, input tensor input is then calculated according to the following formula:
Input=R ⊙ (1-mask ')
Wherein, ⊙ indicates that corresponding element is multiplied;
Inhibit masking-out mask ' through prospectFRInput tensor input ' is obtained after processing:
Input '=input ⊙ (1-mask 'FR);
S6: input tensor input ' being input in the generation model that step S2 is obtained, the image output exported, Then it handles to obtain the filled target image I " in cavity using following formulaS:
I″S=output ⊙ mask '+I 'S⊙(1-mask′)。
The present invention is based on the DIBR system gap filling methods for generating confrontation network, collect a collection of reference picture sample first With corresponding depth image, training sample is generated, is trained, is used for confrontation network is generated using this training sample Then it is corresponding to generate target image to be filled using method identical with training sample for the generation model of DIBR system cavity filling Tensor, and generate prospect inhibit masking-out, processing obtain input tensor, be input to generation model, finally to output result It is further processed to obtain the filled target image in cavity.The present invention extracts low-level features and height using confrontation network is generated Grade semantic feature, proposes the consistency of highly filled hole region and remainder pixel.
Detailed description of the invention
Fig. 1 is that the present invention is based on the specific embodiment processes for the DIBR system gap filling method for generating confrontation network Figure;
Fig. 2 is the flow chart that training sample is obtained in the present invention;
Fig. 3 is training sample generation phase 3-D image transformation exemplary diagram;
Fig. 4 is the exemplary diagram that confrontation network structure is generated in the present embodiment;
Fig. 5 is to generate target image exemplary diagram to be filled;
Fig. 6 is the exemplary diagram of prospect filling cavity;
Fig. 7 is the flow chart for generating prospect in the present embodiment and inhibiting masking-out;
Fig. 8 is in the present embodiment using the exemplary diagram for generating the empty filling of model progress;
Fig. 9 is to fill exemplary diagram using the target image cavity that " 003 " frame is generated as reference picture in " Ballet " sequence;
Figure 10 is to fill example using the target image cavity that " 096 " frame is generated as reference picture in " Ballet " sequence Figure;
Figure 11 is that whether there is or not the target image cavities of prospect suppression mechanism to fill comparison diagram in the present embodiment.
Specific embodiment
A specific embodiment of the invention is described with reference to the accompanying drawing, preferably so as to those skilled in the art Understand the present invention.Requiring particular attention is that in the following description, when known function and the detailed description of design perhaps When can desalinate main contents of the invention, these descriptions will be ignored herein.
Embodiment
Fig. 1 is that the present invention is based on the specific embodiment processes for the DIBR system gap filling method for generating confrontation network Figure.As shown in Figure 1, the present invention is based on generate confrontation network DIBR system gap filling method the following steps are included:
S101: training sample is obtained:
Firstly the need of acquisition training sample in the present invention.Fig. 2 is the flow chart that training sample is obtained in the present invention.Such as Fig. 2 Shown, the specific steps that training sample is obtained in the present invention include:
S201: reference picture is collected:
Collect a collection of reference pictureAnd corresponding depth image Dt, t=1,2 ..., K, K expression number of reference pictures.
S202: 3-D image transformation:
By every width reference pictureWith depth image Dt3-D image transformation is carried out, corresponding target image is obtainedAnd Generation and target imageThe masking-out mask of identical sizet, masking-out masktFor marking empty pixel, if target image In pixel be empty pixel, then in masking-out masktIn corresponding pixel value be 1, if it is non-empty pixel, then exist Masking-out masktIn corresponding pixel value be 0.Due to target imageSize and reference pictureIt is identical, therefore masking-out maskt Size also with reference pictureIt is identical.
Fig. 3 is training sample generation phase 3-D image transformation exemplary diagram.As shown in figure 3, the hole region of target image It is marked as white, black region indicates non-hole region in masking-out.3-D image transformation is a kind of common technological means, Details are not described herein for detailed process.
S203: training sample is generated:
By every width reference pictureEach access matrix and corresponding masking-out masktBe separately converted to batch_size × Mt×Nt× 1 tensor, then successively splicing obtains batch_size × Mt×NtThe tensor of × (L+1) is as corresponding training Sample, wherein batch_size indicates batch size, Mt×NtIndicate reference pictureSize, L indicate reference pictureIt is logical Road number, in general reference pictureIncluding RGB triple channel, therefore L=3, and masking-out masktIt is in the present invention a channel.
Criticizing size batch_size is sample size needed for calculating gradient when generating confrontation network training, too small to will lead to Inefficiency can not restrain, and will lead to very much memory greatly can not support, after batch_size increases to a certain extent, decline Direction change very little, so batch_size is a critically important parameter, batch_size is set as 16 in the present embodiment.
For picture size Mt×NtFor, if will increase very much trained complexity greatly, keep memory consumption excessive, therefore Picture size when generating training sample can be set in practical applications according to the size of memory source, it is every in the present embodiment The size of width reference picture is 256 × 256.The size of each training sample is 16 × 256 × 256 × 4 so in the present embodiment.
S102: training obtains generating model:
The training sample generated using step S101 is trained to confrontation network is generated, and is obtained for DIBR system cavity The generation model of filling.
The specific structure for generating confrontation network, which can according to need, to be configured.Bibliography in the present embodiment " Yu, J., Et al., Free-Form Image Inpainting with Gated Convolution.2018. " devises a kind of generation Fight network structure.Fig. 4 is the exemplary diagram that confrontation network structure is generated in the present embodiment.As shown in figure 4, being generated in the present embodiment Fighting network includes generating model and discrimination model two parts, and generate model includes two stages again, and the first stage is rolled up by 15 Product core is constituted, and reconstruction loss (reconstruction loss) has been used to be trained, the filling result relatively obscured; Second stage is equally made of 15 convolution kernels, uses reconstruction loss, global and local WGAN-GP (Wasserstein GAN With gradient penalty) antagonism loss be trained, obtain final result.Discrimination model is by six convolution kernel structures At after the 6th convolution kernel, the size of characteristic pattern becomesWherein H indicates to generate the figure of training sample The height of picture, W indicates to generate the width of the image of training sample, at this time to element application SN-GAN each in characteristic pattern (Spectral Normalization GAN) loses to train differentiation network.Note that all convolution kernels in the network architecture It is all made of document " Yu, J., et al., Free-Form Image Inpainting with Gated Convolution.2018. the gate convolution (gated convolution) in ".Using the network architecture can relatively faster, More stably train the measured generation model of chymoplasm.
S103: target image to be filled is generated:
For through reference picture I 'RCarry out the target image I ' to cavity filling that image three-dimensional convertsS, generate with Target image I 'SThe non-empty matrix M ' of identical sizeSWith masking-out mask ', wherein non-cavity matrix M 'SFor distinguishing target figure As I 'SIn non-empty pixel and empty pixel, for target image I 'SIn pixel be empty pixel, then in non-empty Hole matrix M 'SIt is middle that corresponding pixel value is set to -1, whereas if being non-empty pixel, then in non-empty matrix M 'SIt is middle to incite somebody to action Corresponding pixel value is set to the depth value of the pixel;Masking-out Mask ' is for marking empty pixel, if target image I 'S In pixel be empty pixel, then be 1 by corresponding pixel value in masking-out Mask ', if it is non-empty pixel, then In masking-out Mask ', corresponding pixel value is set to 0 by.Clearly as non-cavity matrix M 'SWith the size and ginseng of masking-out mask ' Examine image I 'RIt is identical.Fig. 5 is to generate target image exemplary diagram to be filled.As shown in figure 5, the hole area of target image to be filled Domain is marked as white, and the gray area of non-cavity matrix indicates non-hole region, and black region indicates non-hole area in masking-out Domain.
S104: it generates prospect and inhibits masking-out:
In 3-D image transformation, empty generation is primarily due to the back in reference picture by foreground occlusion in target image Scenery body becomes visible in the target image.In this case, cavity should be filled by background information, and be unable to prospect of the application letter Breath filling.However it has been investigated that, the generation model that above-mentioned training obtains, since its context pays close attention to mechanism, in filling cavity When can pay close attention to all foreground and background information around cavity, lead to foreground information occur in the hole region that is filled of part, Generate semantic distortion.Fig. 6 is the exemplary diagram of prospect filling cavity.As shown in fig. 6, the part cavity on the right of female dancer is filled to be The dermatoid texture of class.In order to solve this problem, the invention proposes a kind of prospect suppression mechanisms, and prospect of the application information is inhibited to fill out Hole region is filled, needs to generate prospect and inhibits masking-out, generation method is as follows:
According to non-empty matrix M 'SIt carries out empty detection and the cavity detected is expanded, obtained matrix is made Inhibit masking-out mask ' for prospectFR
Fig. 7 is the flow diagram for generating prospect in the present embodiment and inhibiting masking-out.As shown in fig. 7, being generated in the present embodiment Details are as follows for the detailed process of prospect inhibition masking-out:
(1) empty edge is detected
To non-empty matrix M 'SFrom left to right, it progressively scans from top to bottom, records the location information at empty edge.To non- Empty matrix M 'SIn a point (x, y), if respective pixel value m (x, y)=- 1 (empty point), and m (x-1, y) ≠ -1, m (x+ 1, y)=1, then mark point (x-1, y) be empty left edge first non-empty point mL;If m (x, y)=- 1 (empty point), and M (x-1, y)=- 1, m (x+1, y) ≠ 1, then mark point (x+1, y) is first non-empty point m of empty right hand edgeR.From cavity Left edge is denoted as num_holes to point number in cavity continuous between right hand edge.
(2) transition is detected
In non-empty matrix M 'SIn, the parallax value of foreground point is greater than the parallax value of background dot.Therefore, if mL-mR≥ Sharp_th, sharp_th are transition value, then it is assumed that the cavity left edge region is prospect, and right hand edge region is background;If mR- mL>=sharp_th, then it is assumed that the cavity right hand edge region is prospect, and left edge region is background;If | mL-mR|≥sharp_th And mL>=bf_th, mR>=bf_th, bf_th are the threshold value for distinguishing foreground and background, then it is assumed that the cavity left and right edges region is Prospect;If | mL-mR| < sharp_th and mL< bf_th, mR< bf_th then thinks that the cavity left and right edges region is background.
(3) expansion cavity
After empty edge detection and transition detection, only when continuous cavity point number num_holes is greater than setting When threshold value len_hole, the operation in expansion cavity is just carried out.If empty left edge is prospect, empty right hand edge is background, then only Left edge cavity is expanded, the pixel number of expansion is preset parameter P;If empty right hand edge is prospect, empty left edge is Background, then only expansion right hand edge is empty;If empty the right and left is prospect, both sides expand P point respectively;If cavity left and right Both sides are background, then without expansive working.
In the present embodiment, the value of above-mentioned parameter sharp_th, bf_th, len_hole, P are respectively 4,110,35,3.
S105: it obtains and generates mode input:
Using same method in step S203 by target image I 'SEach access matrix and masking-out mask ' convert respectively For the tensor of batch_size × M ' × N ' × 1, wherein M ' × N ' expression target image I 'SSize, then successively splice To batch_size × M ' × N ' × (L+1) tensor R, input tensor input is then calculated according to the following formula:
Input=R ⊙ (1-mask ')
Wherein, ⊙ indicates that corresponding element is multiplied, i.e., respective pixel value is multiplied in image.1 indicates and mask ' identical dimensional Complete 1 tensor.Since R is batch_size × M ' × N ' × (L+1) tensor, mask ' is batch_size × M ' × N ' × 1 Tensor, therefore above formula is equivalent to and enables each component in R in fourth dimension be multiplied respectively with 1-mask ' carry out corresponding element Operation, obtains batch_size × M ' × N ' × (L+1) input tensor input.
Inhibit masking-out mask ' through prospectFRInput tensor input ' is obtained after processing:
Input '=input ⊙ (1-mask 'FR)
Similarly, above formula be equivalent in middle input each component in fourth dimension respectively with 1-mask 'FRCarry out pair The operation for answering element multiplication obtains batch_size × M ' × N ' × (L+1) input tensor input '.
S106: cavity filling:
Input tensor input ' is input in the generation model that step S102 is obtained, the image output exported, Then it handles to obtain the filled target image I " in cavity using following formulaS:
I″S=output ⊙ mask '+I 'S⊙(1-mask′)
Fig. 8 is in the present embodiment using the exemplary diagram for generating the empty filling of model progress.As shown in figure 8, through generating model Coding-decoded output output is relatively fuzzyyer.Therefore the present invention only takes the part that mask ' white area is corresponded in output (i.e. empty) and target image I 'SNon- hole region part be added to obtain final output result I "S, to ensure target image Clarity.
From the above description, it can be seen that when training generates confrontation network using reference picture and masking-out in the present invention, What input generated model when carrying out cavity filling is the input for inhibiting masking-out to generate according to target image, masking-out and prospect Amount.Therefore in practical applications, if it is needing to a sequence of video images through the transformed target image of image three-dimensional When sequence carries out cavity filling, training sample is generated using the image in original video image sequence, is carried out to confrontation network is generated After training obtains the generation model filled for DIBR system cavity, then it is empty to target image sequence progress using the generation model Hole filling.The training sample generated under this mode by original video image sequence ginseng related to target image sequence to be filled Number compares there are consistency and generates the generation model that training sample training obtains, cavity filling using other reference pictures Technical effect can be more preferable.
Technical effect in order to better illustrate the present invention carries out experimental verification using " Ballet " sequence image.Fig. 9 is Exemplary diagram is filled using the target image cavity that " 003 " frame is generated as reference picture in " Ballet " sequence.Figure 10 be with Fill exemplary diagram in the target image cavity that " 096 " frame is generated as reference picture in " Ballet " sequence.Such as Fig. 9 and Figure 10 institute Show, use cam5 view generation cam4 view herein, lists general image respectively and 4 comparison diagrams for being locally filled with front and back, Wherein white area is hole region.It can be seen that the present invention has very hole region part from the point of view of whole and part Good filling effect.
It is same to be had using " Ballet " sequence image in order to show the technical effect of prospect suppression mechanism in the present invention Experimental verification without prospect suppression mechanism.Figure 11 is that whether there is or not the fillings pair of the target image cavity of prospect suppression mechanism in the present embodiment Than figure.The apparent part of prospect inhibitory effect is marked using ellipse in Figure 11.As shown in figure 11, the present invention can press down well System generates model prospect of the application information and carrys out filling cavity, to obtain the target view of better authenticity.
Although the illustrative specific embodiment of the present invention is described above, in order to the technology of the art Personnel understand the present invention, it should be apparent that the present invention is not limited to the range of specific embodiment, to the common skill of the art For art personnel, if various change the attached claims limit and determine the spirit and scope of the present invention in, these Variation is it will be apparent that all utilize the innovation and creation of present inventive concept in the column of protection.

Claims (2)

1. a kind of based on the DIBR system gap filling method for generating confrontation network, which comprises the following steps:
S1: training sample is generated using following methods:
S1.1: a collection of reference picture is collectedAnd corresponding depth image Dt, t=1,2 ..., K, K expression number of reference pictures;
S1.2: by every width reference pictureWith depth image Dt3-D image transformation is carried out, corresponding target image is obtainedAnd Generation and target imageThe masking-out mask of identical sizet, masking-out masktFor marking empty pixel, if target image In pixel be empty pixel, then in masking-out masktIn corresponding pixel value be 1, if it is non-empty pixel, then exist Masking-out masktIn corresponding pixel value be 0;
S1.3: by every width reference pictureEach channel and corresponding masking-out masktIt is separately converted to batch_size × Mt× Nt× 1 tensor, then successively splicing obtains batch_size × Mt×NtThe tensor of × (L+1) as corresponding training sample, Wherein batch_size indicates batch size, Mt×NtIndicate reference pictureSize, L indicate reference picturePort number;
S2: the training sample generated using step S1 is trained to confrontation network is generated, and obtains filling out for DIBR system cavity The generation model filled;
S3: for through reference picture I 'RCarry out the target image I ' to cavity filling that image three-dimensional convertsS, generate and mesh Logo image I 'SThe non-empty matrix M ' of identical sizeSWith masking-out mask ', wherein non-cavity matrix M 'SFor distinguishing target image I′SIn non-empty pixel and empty pixel, for target image I 'SIn pixel be empty pixel, then in non-cavity Matrix M 'SIt is middle that corresponding element value is set to -1, whereas if being non-empty pixel, then in non-empty matrix M 'SIn will be right The element value answered is set to the depth value of the pixel;Masking-out mask ' is for marking empty pixel, if target image I 'SIn Pixel be empty pixel, then be 1 by corresponding element value in masking-out mask ', if it is non-empty pixel, then exist Masking-out mask ' corresponding element value by is set to 0;
S4: according to non-empty matrix M 'SEmpty detection is carried out simultaneously to expand the cavity detected, using obtained matrix as Prospect inhibits masking-out mask 'FR
S5: using same method in step S1.3 by target image I 'SEach access matrix and masking-out mask ' be separately converted to The tensor of batch_size × M ' × N ' × 1, wherein M ' × N ' expression target image I 'SSize, then successively splicing obtain Then input tensor input is calculated in batch_size × M ' × N ' × (L+1) tensor R according to the following formula:
Input=R ⊙ (1-mask ')
Wherein, ⊙ indicates that corresponding element is multiplied;
Inhibit masking-out mask ' through prospectFRInput tensor input ' is obtained after processing:
Input '=input ⊙ (1-mask 'FR);
S6: input tensor input ' is input in the generation model that step S2 is obtained, the image output exported, then It handles to obtain the filled target image I " in cavity using following formulaS:
I″S=output ⊙ mask '+I 'S⊙(1-mask′)。
2. DIBR system gap filling method according to claim 1, which is characterized in that needing to a video image It is raw using the image in original video image sequence when sequence carries out cavity filling through the transformed target image sequence of image three-dimensional It at training sample, is trained after obtaining the generation model for the filling of DIBR system cavity, then uses to confrontation network is generated The generation model carries out empty filling to target image sequence.
CN201811512179.3A 2018-12-11 2018-12-11 DIBR system cavity filling method based on generation countermeasure network Active CN109462747B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811512179.3A CN109462747B (en) 2018-12-11 2018-12-11 DIBR system cavity filling method based on generation countermeasure network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811512179.3A CN109462747B (en) 2018-12-11 2018-12-11 DIBR system cavity filling method based on generation countermeasure network

Publications (2)

Publication Number Publication Date
CN109462747A true CN109462747A (en) 2019-03-12
CN109462747B CN109462747B (en) 2020-06-26

Family

ID=65612970

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811512179.3A Active CN109462747B (en) 2018-12-11 2018-12-11 DIBR system cavity filling method based on generation countermeasure network

Country Status (1)

Country Link
CN (1) CN109462747B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110381304A (en) * 2019-07-23 2019-10-25 京东方科技集团股份有限公司 For repairing the model training method of empty picture, the restorative procedure of empty picture
CN110580687A (en) * 2019-08-05 2019-12-17 重庆大学 Data preprocessing method for improving filling quality of generated countermeasure network cavity
CN111091616A (en) * 2019-11-25 2020-05-01 艾瑞迈迪科技石家庄有限公司 Method and device for reconstructing three-dimensional ultrasonic image
CN111325693A (en) * 2020-02-24 2020-06-23 西安交通大学 Large-scale panoramic viewpoint synthesis method based on single-viewpoint RGB-D image
CN111968053A (en) * 2020-08-13 2020-11-20 南京邮电大学 Image restoration method based on gate-controlled convolution generation countermeasure network
CN111985608A (en) * 2019-05-23 2020-11-24 宏达国际电子股份有限公司 Method for training generation of confrontation network and method for generating image
CN112102284A (en) * 2020-09-14 2020-12-18 推想医疗科技股份有限公司 Marking method, training method and device of training sample of image segmentation model
CN112215151A (en) * 2020-10-13 2021-01-12 电子科技大学 Method for enhancing anti-interference capability of target detection system by using 3D (three-dimensional) antagonistic sample
CN112233127A (en) * 2020-10-15 2021-01-15 上海圭目机器人有限公司 Down-sampling method for curve splicing image

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101404777A (en) * 2008-11-06 2009-04-08 四川虹微技术有限公司 Drafting view synthesizing method based on depth image
CN102307312A (en) * 2011-08-31 2012-01-04 四川虹微技术有限公司 Method for performing hole filling on destination image generated by depth-image-based rendering (DIBR) technology
US20170188002A1 (en) * 2015-11-09 2017-06-29 The University Of Hong Kong Auxiliary data for artifacts - aware view synthesis

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101404777A (en) * 2008-11-06 2009-04-08 四川虹微技术有限公司 Drafting view synthesizing method based on depth image
CN102307312A (en) * 2011-08-31 2012-01-04 四川虹微技术有限公司 Method for performing hole filling on destination image generated by depth-image-based rendering (DIBR) technology
US20170188002A1 (en) * 2015-11-09 2017-06-29 The University Of Hong Kong Auxiliary data for artifacts - aware view synthesis

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
RAN LIU: "《Hole-filling Based on Disparity Map for DIBR》", 《KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS》 *

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111985608B (en) * 2019-05-23 2024-03-15 宏达国际电子股份有限公司 Method for training generation of countermeasure network and method for generating image
CN111985608A (en) * 2019-05-23 2020-11-24 宏达国际电子股份有限公司 Method for training generation of confrontation network and method for generating image
CN110381304A (en) * 2019-07-23 2019-10-25 京东方科技集团股份有限公司 For repairing the model training method of empty picture, the restorative procedure of empty picture
CN110580687B (en) * 2019-08-05 2021-02-02 重庆大学 Data preprocessing method for improving filling quality of generated countermeasure network cavity
CN110580687A (en) * 2019-08-05 2019-12-17 重庆大学 Data preprocessing method for improving filling quality of generated countermeasure network cavity
CN111091616A (en) * 2019-11-25 2020-05-01 艾瑞迈迪科技石家庄有限公司 Method and device for reconstructing three-dimensional ultrasonic image
CN111091616B (en) * 2019-11-25 2024-01-05 艾瑞迈迪医疗科技(北京)有限公司 Reconstruction method and device of three-dimensional ultrasonic image
CN111325693A (en) * 2020-02-24 2020-06-23 西安交通大学 Large-scale panoramic viewpoint synthesis method based on single-viewpoint RGB-D image
CN111325693B (en) * 2020-02-24 2022-07-12 西安交通大学 Large-scale panoramic viewpoint synthesis method based on single viewpoint RGB-D image
CN111968053B (en) * 2020-08-13 2022-08-30 南京邮电大学 Image restoration method based on gate-controlled convolution generation countermeasure network
CN111968053A (en) * 2020-08-13 2020-11-20 南京邮电大学 Image restoration method based on gate-controlled convolution generation countermeasure network
CN112102284A (en) * 2020-09-14 2020-12-18 推想医疗科技股份有限公司 Marking method, training method and device of training sample of image segmentation model
CN112215151A (en) * 2020-10-13 2021-01-12 电子科技大学 Method for enhancing anti-interference capability of target detection system by using 3D (three-dimensional) antagonistic sample
CN112215151B (en) * 2020-10-13 2022-10-25 电子科技大学 Method for enhancing anti-interference capability of target detection system by using 3D (three-dimensional) countermeasure sample
CN112233127A (en) * 2020-10-15 2021-01-15 上海圭目机器人有限公司 Down-sampling method for curve splicing image

Also Published As

Publication number Publication date
CN109462747B (en) 2020-06-26

Similar Documents

Publication Publication Date Title
CN109462747A (en) Based on the DIBR system gap filling method for generating confrontation network
CN106504190B (en) A kind of three-dimensional video-frequency generation method based on 3D convolutional neural networks
US9137512B2 (en) Method and apparatus for estimating depth, and method and apparatus for converting 2D video to 3D video
KR101625830B1 (en) Method and device for generating a depth map
JP6094863B2 (en) Image processing apparatus, image processing method, program, integrated circuit
US20140146139A1 (en) Depth or disparity map upscaling
AU2016397878B2 (en) Display method and system for converting two-dimensional image into multi-viewpoint image
CN112543317B (en) Method for converting high-resolution monocular 2D video into binocular 3D video
WO2011033673A1 (en) Image processing apparatus
CN104112275B (en) A kind of method and device for generating viewpoint
US20110229024A1 (en) Devices and Methods for Processing Images Using Scale Space
CN104252700B (en) A kind of histogram equalization method of infrared image
CN108449596B (en) 3D stereoscopic image quality evaluation method integrating aesthetics and comfort
WO2012036176A1 (en) Reducing viewing discomfort
CN106530336B (en) Stereo matching method based on color information and graph cut theory
CN102098528A (en) Method and device for converting planar image into stereoscopic image
CN103905813A (en) DIBR hole filling method based on background extraction and partition recovery
KR20130008555A (en) Method for visualizing three-dimensional images on a 3d display device and 3d display device
US20160180514A1 (en) Image processing method and electronic device thereof
CN102496138A (en) Method for converting two-dimensional images into three-dimensional images
Litvinov et al. Incremental solid modeling from sparse structure-from-motion data with improved visual artifacts removal
Lu et al. Foreground-object-protected depth map smoothing for DIBR
Yue et al. Improved seam carving for stereo image resizing
Chen et al. An effective video stitching method
Bharathi et al. 2D-to-3D Conversion of Images Using Edge Information

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20210616

Address after: 20 / F, building 5, Aerospace City Center Plaza, 666 East Chang'an Street, Xi'an national civil aerospace industry base, Shaanxi 710000

Patentee after: Shaanxi Hengsheng Bolong JunXin Technology Co.,Ltd.

Address before: 1 / F, no.111-14, Xiaojiahe street, high tech Zone, Chengdu, Sichuan 610041

Patentee before: CHENGDU MEILYU SCIENCE & TECHNOLOGY Co.,Ltd.

Patentee before: Chongqing University

TR01 Transfer of patent right