CN110602476A - Hole filling method of Gaussian mixture model based on depth information assistance - Google Patents

Hole filling method of Gaussian mixture model based on depth information assistance Download PDF

Info

Publication number
CN110602476A
CN110602476A CN201910729173.XA CN201910729173A CN110602476A CN 110602476 A CN110602476 A CN 110602476A CN 201910729173 A CN201910729173 A CN 201910729173A CN 110602476 A CN110602476 A CN 110602476A
Authority
CN
China
Prior art keywords
texture
background
depth
frame
interval
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910729173.XA
Other languages
Chinese (zh)
Other versions
CN110602476B (en
Inventor
高攀
朱恬恬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Aeronautics and Astronautics
Original Assignee
Nanjing University of Aeronautics and Astronautics
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Aeronautics and Astronautics filed Critical Nanjing University of Aeronautics and Astronautics
Priority to CN201910729173.XA priority Critical patent/CN110602476B/en
Publication of CN110602476A publication Critical patent/CN110602476A/en
Application granted granted Critical
Publication of CN110602476B publication Critical patent/CN110602476B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/40Image enhancement or restoration using histogram techniques
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/122Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • G06T2207/10021Stereoscopic video; Stereoscopic image sequence
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N2013/0074Stereoscopic image analysis
    • H04N2013/0081Depth or disparity estimation from stereoscopic image signals

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开一种基于深度信息辅助的高斯混合模型的空洞填补方法,该方法包括:(1)获取包含多个视图的图像序列数据,每个视图的图像序列数据均含有若干帧的纹理帧和深度帧;(2)采用混合高斯模型算法确定所述图像序列数据中的GMM纹理背景;(3)将纹理帧和深度帧按时序都分为若干个区间,并对每个区间内的深度帧进行直方图均衡化处理;(4)对每个区间中的纹理帧和深度帧进行前景深度相关算法的计算,对应得到FDC区间纹理背景和FDC区间深度背景;(5)将每个区间得到的FDC区间纹理背景,按照FDC区间深度背景中含有的深度信息结合起来最终得到FDC纹理背景。本发明方法可以肉眼观察到更好的空洞填充效果,并且可以在往复运动序列中得到显着的客观增益。

The invention discloses a hole filling method based on a Gaussian mixture model assisted by depth information. The method includes: (1) acquiring image sequence data including multiple views, and the image sequence data of each view contains several frames of texture frames and Depth frame; (2) using the mixed Gaussian model algorithm to determine the GMM texture background in the image sequence data; (3) dividing the texture frame and the depth frame into several intervals in time sequence, and analyzing the depth frame in each interval Perform histogram equalization processing; (4) calculate the foreground depth correlation algorithm for the texture frame and depth frame in each interval, and obtain the texture background of the FDC interval and the depth background of the FDC interval; (5) obtain the depth background of each interval The FDC interval texture background is combined with the depth information contained in the FDC interval depth background to finally obtain the FDC texture background. The method of the present invention can visually observe better cavity filling effect, and can obtain significant objective gain in the reciprocating motion sequence.

Description

一种基于深度信息辅助的高斯混合模型的空洞填补方法A Hole Filling Method Based on Gaussian Mixture Model Assisted by Depth Information

技术领域technical field

本发明涉及三维视频空洞填补技术领域,具体涉及一种基于深度信息辅助的高斯混合模型的空洞填补方法。The invention relates to the technical field of three-dimensional video hole filling, in particular to a hole filling method based on a Gaussian mixture model assisted by depth information.

背景技术Background technique

虚拟视点合成已成为三维视频(3-D)研究的核心部分,因为它能够避免在自由视点视频(FVV)中传输大量视点图像。最常用的视点合成技术是使用纹理图像及其相关深度图的基于深度图像的渲染(DIBR,Depth Image Based Rendering),虽然DIBR技术已经发展的较为成熟,然而虚拟视图中的空洞填补问题仍然是一个比较棘手的问题。Virtual view synthesis has become a central part of three-dimensional video (3-D) research because it enables the avoidance of transmitting a large number of view images in free view video (FVV). The most commonly used view synthesis technology is Depth Image Based Rendering (DIBR, Depth Image Based Rendering) using texture images and their associated depth maps. Although DIBR technology has been developed relatively maturely, the problem of hole filling in virtual views is still a problem. A trickier question.

一般来讲,产生空洞的情况主要有两种,一是不准确的深度值导致的虚拟视图中的不连续区域,二是在参考视图中被遮挡的背景对象可能在虚拟视图中变得可见,从而导致了虚拟视图中的大片空洞。图像修复是一种常用的填补空洞的技术,主要根据相邻像素之间的空间纹理相关性来确定空洞区域的像素值。然而,被遮挡区域与前景区域之间通常没有空域纹理相关性,在虚拟视图中存在大片因遮挡产生的空洞的情况下,采用普通图像修复的结果与实际真实视角图像之间的差异往往很大。Generally speaking, there are two main situations that cause holes. One is the discontinuous area in the virtual view caused by inaccurate depth values, and the other is that background objects that are occluded in the reference view may become visible in the virtual view. This results in large holes in the virtual view. Image inpainting is a common hole-filling technique, which mainly determines the pixel value of the hole area according to the spatial texture correlation between adjacent pixels. However, there is usually no spatial texture correlation between the occluded area and the foreground area. In the case of large holes in the virtual view due to occlusion, the difference between the results of ordinary image inpainting and the actual real-world view image is often large. .

为了得到更好的空洞填补结果,目前最流行的技术是GMM(混合高斯模型,Gaussian Mixture Model)算法。GMM可根据时间相关性信息生成稳定的场景背景,其中每个像素由多个高斯模型来建模和表示,这些高斯模型在出现新的样本数据时以一定学习率进行迭代更新。因为在虚拟视图中无法恢复的遮挡区域通常是背景,所以使用GMM生成的背景来填充被遮挡的区域是一种很有效的方法。但是,对于进行周期旋转或者往复运动的前景区域,由于前景在多个时间帧内的同样位置重复出现,GMM通常会错误地将它们视为背景,并填充到空洞区域中。这样得到的合成图像将会与真实相机在该位置拍摄的图像有较大的差异,进而影响3D视频的观看效果。In order to obtain better hole-filling results, the most popular technology at present is the GMM (Gaussian Mixture Model, Gaussian Mixture Model) algorithm. GMM can generate a stable scene background based on temporal correlation information, in which each pixel is modeled and represented by multiple Gaussian models, and these Gaussian models are iteratively updated at a certain learning rate when new sample data appears. Because the occluded areas that cannot be recovered in the virtual view are usually the background, it is an effective method to use the background generated by GMM to fill the occluded areas. However, for the foreground areas that undergo periodic rotation or reciprocating motion, because the foreground repeats at the same position in multiple time frames, GMM usually mistakenly regards them as background and fills them into the hollow area. The synthetic image obtained in this way will be quite different from the image captured by the real camera at the position, thereby affecting the viewing effect of the 3D video.

发明内容Contents of the invention

发明目的:为了克服现有技术的不足,本发明提供一种基于深度信息辅助的高斯混合模型的空洞填补方法,该方法可以解决现有GMM模型中使用造成的空洞填补效果不佳问题。Purpose of the invention: In order to overcome the deficiencies of the prior art, the present invention provides a hole filling method based on a Gaussian mixture model assisted by depth information, which can solve the problem of poor hole filling effect caused by the use of the existing GMM model.

技术方案:本发明所述的基于深度信息辅助的高斯混合模型的空洞填补方法,该方法包括:Technical solution: The hole filling method based on the Gaussian mixture model assisted by depth information according to the present invention, the method includes:

(1)使用包含多个视图的图像序列数据,每个视图的图像序列数据均含有若干帧的纹理帧和深度帧;(1) Using image sequence data comprising multiple views, the image sequence data of each view contains several frames of texture frames and depth frames;

(2)采用混合高斯模型算法确定所述图像序列数据中的GMM纹理背景;(2) adopting the mixed Gaussian model algorithm to determine the GMM texture background in the image sequence data;

(3)将纹理帧和深度帧按时序都分为若干个区间,并对每个区间内的深度帧进行直方图均衡化处理;(3) Divide the texture frame and the depth frame into several intervals according to the time sequence, and perform histogram equalization processing on the depth frame in each interval;

(4)对每个区间中的纹理帧进行前景深度相关算法的计算,对应得到FDC区间纹理背景;对每个区间中的二值化后的深度帧进行前景深度相关算法的计算,对应得到FDC区间深度背景;(4) Calculate the foreground depth correlation algorithm for the texture frame in each interval, and obtain the FDC interval texture background correspondingly; perform the calculation of the foreground depth correlation algorithm for the binarized depth frame in each interval, and obtain the FDC correspondingly Interval depth background;

(5)将每个区间得到的FDC区间纹理背景,按照FDC区间深度背景中含有的深度信息结合起来最终得到FDC纹理背景;(5) The FDC interval texture background obtained in each interval is combined with the depth information contained in the FDC interval depth background to finally obtain the FDC texture background;

进一步地,包括:Further, include:

所述步骤(2),采用混合高斯模型算法确定所述图像序列数据中的GMM纹理背景,具体包括:Described step (2), adopts mixed Gaussian model algorithm to determine the GMM texture background in described image sequence data, specifically comprises:

(21)将图像序列数据中的所有纹理帧和深度帧用于GMM流程中,所述纹理帧用于构建高斯模型,深度帧用于决定模型更新参数,对于在不同帧对应的单个像素点,包括以下步骤:(21) All texture frames and depth frames in the image sequence data are used in the GMM process, the texture frames are used to build a Gaussian model, and the depth frames are used to determine model update parameters. For a single pixel corresponding to a different frame, Include the following steps:

(211)第t帧纹理帧的第k个高斯模型记为ηk,t,模型均值记为μk,t,标准差记为σk,t,权重记为ωk,t,1≤k≤M,M为在每个像素点上构建的高斯模型总数;(211) The k-th Gaussian model of the t-th texture frame is recorded as η k,t , the model mean value is recorded as μ k,t , the standard deviation is recorded as σ k,t , and the weight is recorded as ω k,t , 1≤k ≤M, M is the total number of Gaussian models constructed on each pixel;

(212)计算下一帧t+1当前像素点的像素值Xt与每个高斯模型均值的差|Xtk,t-1|,若|Xtk,t-1|<2.5σk,t,则判定下一帧的当前像素点的像素值属于所述高斯模型,并对该模型的学习率进行更新;(212) Calculate the difference between the pixel value X t of the current pixel point in the next frame t+1 and the mean value of each Gaussian model |X tk,t-1 |, if |X tk,t-1 | <2.5σ k,t , it is determined that the pixel value of the current pixel point in the next frame belongs to the Gaussian model, and the learning rate of the model is updated;

否则,若|Xtk,t-1|≥2.5σk,t,则判定下一帧的当前像素点的像素值不属于M个高斯模型中的任意一个,且删除M个模型中权重与方差比最小的一个高斯模型,并构建新高斯模型替换;Otherwise, if |X tk,t-1 |≥2.5σ k,t , it is determined that the pixel value of the current pixel point in the next frame does not belong to any of the M Gaussian models, and the M models are deleted A Gaussian model with the smallest weight-to-variance ratio, and a new Gaussian model is constructed to replace it;

(22)将步骤(21)中得到的所有含有一个ωk,t=1的模型,其他M-1模型的权重ωk,t=0的像素点,确定为只含有一个高斯模型的像素点;(22) All the models obtained in step (21) containing a ω k, t = 1, the pixels of the weight ω k, t = 0 of other M-1 models are determined as the pixels containing only one Gaussian model ;

(23)新建膜图,采用像素值0记录只含有一个高斯模型所在的位置,确定其为背景位置,采用像素值255记录下含有多个高斯模型像素点所在的位置,从而确定GMM纹理背景。(23) Create a new film map, use pixel value 0 to record the position containing only one Gaussian model, determine it as the background position, and use pixel value 255 to record the position containing multiple Gaussian model pixels, thereby determining the GMM texture background.

进一步地,包括:Further, include:

所述步骤(212)中,对该模型的学习率进行更新,学习率更新公式为:In the step (212), the learning rate of the model is updated, and the learning rate update formula is:

其中,α为初始学习率,ε为序列中的经验值,是由深度帧观察得出的能区分前景背景区域的深度像素值,dt(x,y)为第t帧深度帧上位于(x,y)位置的像素值,αt为更新后的学习率。Among them, α is the initial learning rate, ε is the experience value in the sequence, which is the depth pixel value that can distinguish the foreground and background regions obtained from the observation of the depth frame, d t (x, y) is the depth frame of the tth frame located at ( x, y) pixel value, α t is the updated learning rate.

进一步地,包括:Further, include:

所述步骤(4)中,对每个区间中的纹理帧进行前景深度相关算法的计算,对应得到FDC区间纹理背景,具体包括:In described step (4), carry out the calculation of foreground depth correlation algorithm to the texture frame in each interval, correspondingly obtain FDC interval texture background, specifically include:

(41)用第一帧生成初始FDC纹理背景,生成方式为采用第一帧中被二值深度图区分为背景的区域填补空白图像,也就是得到第一帧背景区域对应的第一帧纹理图中的背景纹理区域;(41) Use the first frame to generate the initial FDC texture background, and the generation method is to fill the blank image with the area that is divided into the background by the binary depth map in the first frame, that is, to obtain the first frame texture map corresponding to the background area of the first frame The background texture area in ;

(42)用第二帧的二值深度图中的第二帧变为背景的区域对应的第二帧纹理图,更新FDC纹理背景中的未填充区域;(42) Update the unfilled area in the FDC texture background with the second frame texture map corresponding to the region where the second frame in the binary depth map of the second frame becomes the background;

(43)对之后的该区间内的所有纹理帧执行与第二帧相同的过程,进而得到该区间的FDC区间纹理背景。(43) Execute the same process as the second frame for all texture frames in the subsequent interval, and then obtain the FDC interval texture background of this interval.

进一步地,包括:Further, include:

所述步骤(4)中,对每个区间中的二值化后的深度帧进行前景深度相关算法的计算,对应得到FDC区间深度背景。In the step (4), the calculation of the foreground depth correlation algorithm is performed on the binarized depth frames in each interval, and the depth background of the FDC interval is correspondingly obtained.

进一步地,包括:Further, include:

所述步骤(5)中,将每个区间得到的FDC区间纹理背景,按照FDC区间深度背景中含有的深度信息结合起来最终得到FDC纹理背景,包括:In the step (5), the FDC interval texture background obtained by each interval is combined with the depth information contained in the FDC interval depth background to finally obtain the FDC texture background, including:

对单个像素点(x,y),采用具有最小深度值的区间深度背景对应的纹理背景来作为最终的纹理背景,其表示为,若有ψdk(x,y)=min(ψdi(x,y)),则ψtf(x,y)=ψtk(x,y),其中,ψtf(x,y)为在点(x,y)最终的FDC纹理背景,ψdi(x,y)为在点(x,y)第i个区间对应的FDC区间深度背景,ψtk(x,y)为在点(x,y)第k个区间对应的FDC区间纹理背景。k为得到的具有最小深度值的ψdi(x,y)的序号i,i∈[1,K],其中下标符号中的t和d用来区分纹理和深度图帧。For a single pixel point (x, y), the texture background corresponding to the interval depth background with the minimum depth value is used as the final texture background, which is expressed as, if ψ dk (x, y)=min(ψ di (x ,y)), then ψ tf (x,y)=ψ tk (x,y), where ψ tf (x,y) is the final FDC texture background at point (x,y), ψ di (x, y) is the FDC interval depth background corresponding to the i-th interval of point (x, y), and ψ tk (x, y) is the FDC interval texture background corresponding to the k-th interval of point (x, y). k is the serial number i of ψ di (x,y) obtained with the minimum depth value, i∈[1,K], where t and d in the subscript are used to distinguish texture and depth image frames.

进一步地,包括:Further, include:

所述步骤(6)中,自适应的采用GMM纹理背景和FDC纹理背景进行所有图像帧数的填充,包括:In described step (6), adaptively adopt GMM texture background and FDC texture background to carry out the filling of all image frame numbers, including:

对含有一个高斯模型的像素点,用GMM纹理背景对应位置的像素点填充;否则,采用FDC纹理背景对应位置的像素点进行填充For pixels containing a Gaussian model, fill them with the pixels corresponding to the GMM texture background; otherwise, fill them with the pixels corresponding to the FDC texture background

有益效果:本发明与现有技术相比,其显著优点是:1、本发明从生成可靠背景的角度出发,为了使得传统的高斯混合模型能够在周期运动的视频序列上具有较好背景生成效果,本发明提出采用深度信息调整学习率的策略;2、本发明为了得到更准确的前景背景分类结果,将序列中的所有帧进行区间划分,在每个区间都生成FDC区间纹理背景,并在结果中将这些背景有机结合起来;3、本发明还根据每个像素的高斯模型个数,将GMM纹理背景结果和FDC纹理背景结果自适应结合起来,该方法进一步完善了背景,适用于虚拟视图中由于遮挡产生的空洞的填补,并产生的更好的填补效果。Beneficial effects: Compared with the prior art, the present invention has the following significant advantages: 1. From the perspective of generating a reliable background, in order to enable the traditional Gaussian mixture model to have a better background generation effect on video sequences with periodic motion , the present invention proposes a strategy of adjusting the learning rate using depth information; 2. In order to obtain more accurate foreground and background classification results, the present invention divides all frames in the sequence into intervals, generates an FDC interval texture background in each interval, and These backgrounds are organically combined in the result; 3. The present invention also combines the GMM texture background result and the FDC texture background result adaptively according to the number of Gaussian models of each pixel. This method further improves the background and is suitable for virtual views Filling the voids caused by occlusion in the medium, and producing a better filling effect.

附图说明Description of drawings

图1为原FDC方法中在区分前景背景时区分结果不准确的样例;Figure 1 is an example of inaccurate results when distinguishing between foreground and background in the original FDC method;

图2为本方法空洞填补的总算法流程框架;Fig. 2 is the overall algorithm flow frame of this method hole filling;

图3为改进的FDC过程;Fig. 3 is the improved FDC process;

图4为改进的FDC过程中原FDC过程;Fig. 4 is the original FDC process in the improved FDC process;

图5为实验示例主观结果展示:5a为Ballet序列相机0到相机1第29帧的虚拟视点合成结果;5b为Breakdancers序列相机5到相机6第6帧的虚拟视点合成结果;5c为Ballet序列相机0到相机1第29帧的空洞填补结果;5d为Breakdancers序列相机5到相机6第6帧的空洞填补结果。Figure 5 shows the subjective results of the experimental example: 5a is the virtual viewpoint synthesis result of the Ballet sequence camera 0 to camera 1 frame 29; 5b is the virtual viewpoint synthesis result of the Breakdancers sequence camera 5 to camera 6 frame 6; 5c is the Ballet sequence camera Hole filling results of frame 29 from camera 0 to camera 1; 5d is the hole filling result of frame 6 from camera 5 to camera 6 of the Breakdancers sequence.

具体实施方式Detailed ways

下面将结合本发明实施例中的附图,对本发明实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本发明一部分实施例,并不是全部的实施例。基于本发明中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本发明保护的范围。The following will clearly and completely describe the technical solutions in the embodiments of the present invention in conjunction with the accompanying drawings in the embodiments of the present invention. Obviously, the described embodiments are only some of the embodiments of the present invention, not all of them. Based on the embodiments of the present invention, all other embodiments obtained by persons of ordinary skill in the art without making creative efforts belong to the protection scope of the present invention.

本发明是依据在参考视图不可见而在虚拟视图可见的遮挡区域通常为背景区域的原理,在FDC的基础上添加滑动窗口方法,并根据高斯模型个数来结合GMM纹理背景和FDC纹理背景,从而得到可靠的背景结果,以达到有效填补空洞的目的。The present invention is based on the principle that the occlusion area that is invisible in the reference view but visible in the virtual view is usually the background area, adds a sliding window method on the basis of FDC, and combines the GMM texture background and the FDC texture background according to the number of Gaussian models. So as to obtain reliable background results, in order to achieve the purpose of effectively filling holes.

在用所得背景来填充空洞的方法中,前景背景的准确区分至关重要。原FDC方法尽管在往复运动序列中能产生良好的主观和客观结果,但它仍然存在将前景错误分类为背景的问题,如图1所示,椭圆框区域为前景错分为背景的区域,并且前景区域一旦被错分为背景,那么该区域将一直存在在背景结果中。这就使得在按时域来逐渐获取FDC背景的过程中,第一帧的划分结果至关重要。In methods that fill holes with the resulting background, accurate distinction between foreground and background is crucial. Although the original FDC method can produce good subjective and objective results in the reciprocating motion sequence, it still has the problem of misclassifying the foreground as the background. As shown in Figure 1, the elliptical frame area is the area where the foreground is misclassified as the background, and Once the foreground area is misclassified as background, this area will always exist in the background result. This makes the segmentation result of the first frame very important in the process of gradually acquiring the FDC background in the time domain.

为了避免这种情况,本发明采用滑动窗口方式,将序列中的纹理帧和深度帧按时序划分为多个区间,对每个区间执行FDC方法,这样每个区间获取一个FDC区间纹理背景结果。由多个FDC区间纹理背景结果按深度信息结合起来决定最终FDC纹理背景,可以从而避免在单区间(即未划分区间)的情形下,出现前景误分为背景而无法挽回的情况。In order to avoid this situation, the present invention adopts a sliding window method, divides the texture frame and depth frame in the sequence into multiple intervals in time sequence, and executes the FDC method for each interval, so that each interval obtains an FDC interval texture background result. The final FDC texture background is determined by combining the texture background results of multiple FDC intervals according to the depth information, which can avoid the irreversible situation that the foreground is mistakenly divided into the background in the case of a single interval (that is, an undivided interval).

如图2所示,本发明公开了一种基于深度信息辅助的高斯混合模型的空洞填补方法,首先,为了避免将前景像素错误分类为背景像素,使用深度信息来调整GMM中的学习率,使得前景像素在模型中的比重减小,背景像素在模型中的比重增大。此外,提出了改进的前景深度相关(FDC)算法,其通过跟踪时域上的前景深度的变化来生成背景帧。与现有算法相比,该算法使用滑动窗口来获得多个背景参考帧,并将这些参考帧利用深度信息融合起来以生成更准确的背景帧。最后,自适应地选择GMM和FDC中的纹理背景像素来填充空洞。采用本发明方法,可以肉眼观察到更好的空洞填充效果,并且可以在往复运动序列中得到显着的客观增益。具体包括:As shown in Figure 2, the present invention discloses a hole filling method based on a Gaussian mixture model assisted by depth information. First, in order to avoid misclassifying foreground pixels as background pixels, the depth information is used to adjust the learning rate in the GMM, so that The proportion of foreground pixels in the model decreases, and the proportion of background pixels in the model increases. Furthermore, an improved Foreground Depth Correlation (FDC) algorithm is proposed, which generates background frames by tracking changes in foreground depth over time. Compared with existing algorithms, this algorithm uses a sliding window to obtain multiple background reference frames, and fuses these reference frames with depth information to generate a more accurate background frame. Finally, textured background pixels in GMM and FDC are adaptively selected to fill holes. By adopting the method of the present invention, better void filling effects can be observed with the naked eye, and significant objective gain can be obtained in the reciprocating motion sequence. Specifically include:

(1)采集包含多个视图的图像序列数据,每个视图的图像序列数据均含有若干帧的纹理帧和深度帧;每个视图是不同的,且纹理帧和深度帧数相同,本发明的空洞可采用下面方法得到:第一视图中的纹理帧和深度帧通过虚拟视点合成算法来相应获取第二视图相同帧数的虚拟纹理帧,由于前景的遮挡效应,这些虚拟纹理帧在前景背景分界处均存在大片空洞,影响了观看效果。(1) Gather image sequence data comprising a plurality of views, and the image sequence data of each view all contain texture frames and depth frames of several frames; each view is different, and the number of texture frames and depth frames is the same, the present invention The hole can be obtained by the following method: the texture frame and the depth frame in the first view use the virtual view synthesis algorithm to obtain the virtual texture frame of the same number of frames in the second view. Due to the occlusion effect of the foreground, these virtual texture frames are separated by the foreground There are large voids everywhere, which affects the viewing effect.

(2)采用混合高斯模型算法确定所述图像序列数据中的GMM纹理背景;(2) adopting the mixed Gaussian model algorithm to determine the GMM texture background in the image sequence data;

提供了参考纹理背景的获取方法。首先将序列中的所有纹理帧和深度帧用于GMM流程中,纹理帧用于构建高斯模型,深度帧用于决定模型更新参数。图像中的每个像素由M个高斯模型表示(一般为3-5个模型),并不断根据后续帧的像素值变化进行更新。Provides a method for obtaining the reference texture background. First, all the texture frames and depth frames in the sequence are used in the GMM process, the texture frames are used to build the Gaussian model, and the depth frames are used to determine the model update parameters. Each pixel in the image is represented by M Gaussian models (generally 3-5 models), which are constantly updated according to the pixel value changes in subsequent frames.

对于单个像素,第t帧的第k个高斯模型记为ηk,t,模型均值记为μk,t,标准差记为σk,t,权重记为ωk,t,其中,序列中的第一帧纹理帧用于构建每个像素的初始化高斯模型;For a single pixel, the kth Gaussian model of the tth frame is recorded as η k,t , the model mean value is recorded as μ k,t , the standard deviation is recorded as σ k,t , and the weight is recorded as ω k,t , where, The first texture frame in the sequence is used to build an initialized Gaussian model for each pixel;

计算下一帧的当前像素的像素值与每个高斯模型均值的差|Xtk,t-1|。如果|Xtk,t-1|<2.5σk,t,则判定下一帧的当前像素的像素值属于该高斯模型,并根据模型更新参数对该模型进行更新,模型更新参数按如下公式决定:Calculate the difference |X tk,t-1 | between the pixel value of the current pixel in the next frame and the mean value of each Gaussian model. If |X tk,t-1 |<2.5σ k,t , it is determined that the pixel value of the current pixel in the next frame belongs to the Gaussian model, and the model is updated according to the model update parameters, which are set according to Determined by the following formula:

其中,α为学习率,是一个常数,ε为序列中的经验值,是由深度帧观察得出的能区分前景背景区域的深度像素值,dt(x,y)为第t帧深度图上位于(x,y)位置的像素值,αt(x,y)为更新后的学习率,对每一帧每一个像素通常都不相同。Among them, α is the learning rate, which is a constant, ε is the experience value in the sequence, which is the depth pixel value that can distinguish the foreground and background regions obtained from the observation of the depth frame, d t (x, y) is the depth map of the tth frame α t (x, y) is the updated learning rate, which is usually different for each pixel of each frame.

对每个像素来说,模型更新的过程如下所示:For each pixel, the model update process is as follows:

μk,t←(1-αtk,t-1+αXt μ k,t ←(1-α tk,t-1 +αX t

ωk,t←(1-αtk,t-1ω k,t ←(1-α tk,t-1

要是下一帧的当前像素的像素值不属于其任意一高斯模型,则删除权重与方差比ωk,tk,t最小的一个模型并构建一个新模型。这样的方式可以生成更好的背景,使得填充效率更高。If the pixel value of the current pixel in the next frame does not belong to any of its Gaussian models, delete the model with the smallest weight-to-variance ratio ω k,tk,t and build a new model. This way can generate a better background, making the filling more efficient.

在GMM过程执行完毕后,有些像素含有一个权重为1的模型,其他M-1模型的权重为0,也就是只有一个模型。这代表在该序列的所有纹理帧中,同一位置的像素值没有发生变化,对于这些像素确定其为背景,并将应用于GMM纹理背景ψg中。After the GMM process is executed, some pixels contain a model with a weight of 1, and other M-1 models have a weight of 0, that is, there is only one model. This means that in all texture frames of the sequence, the pixel values at the same position do not change, and these pixels are determined to be the background and will be applied to the GMM texture background ψ g .

新建膜图,采用像素值0记录只含有一个高斯模型所在的位置,确定其为背景位置,采用像素值255记录下含有多个高斯模型像素点所在的位置,从而确定GMM纹理背景。Create a new membrane map, use pixel value 0 to record the location of only one Gaussian model, and determine it as the background location, use pixel value 255 to record the location of multiple Gaussian model pixels, so as to determine the GMM texture background.

(3)将纹理帧和深度帧按时序都分为若干个区间,并对每个区间内的深度帧进行直方图均衡化处理;(3) Divide the texture frame and the depth frame into several intervals according to the time sequence, and perform histogram equalization processing on the depth frame in each interval;

将序列中的所有纹理帧和深度帧按时序分为K个区间,使用K-Means方法对区间中的每个深度图进行处理,以获得二值化的深度图。其中,k=2。像素值0表示背景,值255表示前景。Divide all the texture frames and depth frames in the sequence into K intervals in time sequence, and use the K-Means method to process each depth map in the interval to obtain a binarized depth map. Among them, k=2. A pixel value of 0 represents the background and a value of 255 represents the foreground.

(4)对每个区间中的纹理帧进行前景深度相关算法的计算,对应得到FDC区间纹理背景;对每个区间中的二值化后的深度帧进行前景深度相关算法的计算,对应得到FDC区间深度背景;(4) Calculate the foreground depth correlation algorithm for the texture frame in each interval, and obtain the FDC interval texture background correspondingly; perform the calculation of the foreground depth correlation algorithm for the binarized depth frame in each interval, and obtain the FDC correspondingly Interval depth background;

参阅图4所示,对每个区间中的纹理帧进行前景深度相关算法的计算,对应得到FDC区间纹理背景,具体包括:Referring to Figure 4, the calculation of the foreground depth correlation algorithm is performed on the texture frame in each interval, and the corresponding FDC interval texture background is obtained, specifically including:

(41)用第一帧生成初始FDC纹理背景,生成方式为采用第一帧中被二值深度图区分为背景的区域填补空白图像,也就是得到第一帧背景区域对应的第一帧纹理图中的背景纹理区域;(41) Use the first frame to generate the initial FDC texture background, and the generation method is to fill the blank image with the area that is divided into the background by the binary depth map in the first frame, that is, to obtain the first frame texture map corresponding to the background area of the first frame The background texture area in ;

(42)用第二帧的二值深度图中的第二帧变为背景的区域对应的第二帧纹理图,更新FDC纹理背景中的未填充区域;(42) Update the unfilled area in the FDC texture background with the second frame texture map corresponding to the region where the second frame in the binary depth map of the second frame becomes the background;

(43)对之后的该区间内的所有纹理帧执行与第二帧相同的过程,进而得到该区间的FDC区间纹理背景。(43) Execute the same process as the second frame for all texture frames in the subsequent interval, and then obtain the FDC interval texture background of this interval.

一旦发现在之前所有帧中都被划为前景的区域,在某一帧被划分为背景,那么该纹理帧中的这个区域将用于恢复FDC区间纹理背景。Once it is found that an area classified as foreground in all previous frames is classified as background in a certain frame, then this area in the texture frame will be used to restore the FDC interval texture background.

得到FDC区间深度背景和FDC区间纹理背景的技术细节是一样的,只是方法作用的对象不同,前者对象为纹理帧,后者对象为二值化后的深度帧。The technical details of obtaining the depth background of the FDC interval and the texture background of the FDC interval are the same, but the object of the method is different. The object of the former is a texture frame, and the object of the latter is a binarized depth frame.

(5)将每个区间得到的FDC区间纹理背景,按照FDC区间深度背景中含有的深度信息结合起来最终得到FDC纹理背景;(5) The FDC interval texture background obtained in each interval is combined with the depth information contained in the FDC interval depth background to finally obtain the FDC texture background;

参阅图3所示,将获得的K张区间背景深度图进行对比,对每个位置对应的K个深度像素点,取深度值最小的像素点所在深度图对应的纹理背景图中,对应位置的纹理像素点来填充背景;即对单个像素点(x,y),我们采用具有最小深度值的区间深度背景对应的纹理背景来作为最终的纹理背景,其表示为,若有ψdk(x,y)=min(ψdi(x,y)),则ψtf(x,y)=ψtk(x,y),其中,ψtf(x,y)为在点(x,y)最终的FDC纹理背景,ψdi(x,y)为在点(x,y)第i个区间对应的FDC区间深度背景,ψtk(x,y)为在点(x,y)第k个区间对应的FDC区间纹理背景。k为得到的具有最小深度值的ψdi(x,y)的序号i,i∈[1,K],其中下标符号中的t和d用来区分纹理和深度图帧。Referring to Figure 3, compare the obtained K interval background depth maps, and for the K depth pixels corresponding to each position, take the texture background map corresponding to the depth map where the pixel point with the smallest depth value is located, and the corresponding position texture pixels to fill the background; that is, for a single pixel point (x, y), we use the texture background corresponding to the interval depth background with the minimum depth value as the final texture background, which is expressed as, if ψ dk (x, y)=min(ψ di (x,y)), then ψ tf (x,y)=ψ tk (x,y), where ψ tf (x,y) is the final FDC texture background, ψ di (x, y) is the FDC interval depth background corresponding to the i-th interval of point (x, y), and ψ tk (x, y) is corresponding to the k-th interval of point (x, y) FDC interval texture background. k is the serial number i,i∈[1,K] of ψ di (x,y) with the minimum depth value obtained, where t and d in the subscript are used to distinguish texture and depth image frames.

本发明由多个FDC区间背景结果按深度信息结合起来决定最终FDC方法背景,可以从而避免在单区间(即未划分区间)的情形下,出现前景误分为背景而无法挽回的情况。The present invention combines the background results of multiple FDC intervals according to the depth information to determine the final FDC method background, thereby avoiding the irreversible situation that the foreground is mistakenly divided into the background in the case of a single interval (that is, an undivided interval).

(6)自适应的采用GMM纹理背景和FDC纹理背景进行所有图像帧数的填充,完成参考纹理背景图,进而对虚拟视图中虚拟纹理帧的空洞区域分别用对应区域背景填充。(6) Adaptively use the GMM texture background and the FDC texture background to fill all the image frames, complete the reference texture background image, and then fill the hollow areas of the virtual texture frames in the virtual view with the corresponding area backgrounds.

对含有一个高斯模型的像素点,用GMM纹理背景对应位置的像素点填充;否则,采用FDC纹理背景对应位置的像素点进行填充,即:最后将GMM纹理背景和FDC纹理背景结合起来,将结合背景记为ψB。对于在GMM中,只有一个模型的像素点,ψB(x,y)=ψg(x,y),否则ψB(x,y)=ψtf(x,y)。For pixels containing a Gaussian model, fill them with the pixels corresponding to the GMM texture background; otherwise, fill them with the pixels corresponding to the FDC texture background, that is: finally combine the GMM texture background and the FDC texture background, and combine The background is denoted as ψ B . For the pixels of only one model in GMM, ψ B (x, y) = ψ g (x, y), otherwise ψ B (x, y) = ψ tf (x, y).

为了验证本发明的有效性,下面结合一个具体的实施方式对本发明作进一步的详细说明。为便于说明,且不失一般性,做如下假定:In order to verify the effectiveness of the present invention, the present invention will be further described in detail below in conjunction with a specific embodiment. For the convenience of explanation and without loss of generality, the following assumptions are made:

本发明提出的方法拟用Microsoft数据集中的Ballet测试序列进行测试,分辨率为1024×768,含有8个不同视图的纹理和深度数据,均为100帧。同时附有每个相机的内矩阵和外矩阵参数(相机0到相机7从右到左排列),以及序列中的最大最小真实深度值。The method proposed by the present invention intends to use the Ballet test sequence in the Microsoft data set for testing, with a resolution of 1024×768, containing 8 different views of texture and depth data, all of which are 100 frames. At the same time, the inner matrix and outer matrix parameters of each camera are attached (camera 0 to camera 7 are arranged from right to left), as well as the maximum and minimum real depth values in the sequence.

本实施例中使用相机0视图通过虚拟视点合成算法以及空洞填补技术来恢复相机1视图。首先使用相机0视图中的100帧纹理帧和深度帧通过虚拟视点合成算法来相应获取相机1视图的100帧虚拟纹理帧。由于前景的遮挡效应,这些虚拟纹理帧在前景背景分界处均存在大片空洞,如图5a和b中的人背后的灰色区域所示,图5c和d为本发明的方法分别进行的空洞填充。为了让虚拟视图有更佳的观看效果,本发明采用获取用GMM纹理背景和FDC纹理背景自适应选择的方式来填补这些空洞。In this embodiment, the view of camera 0 is used to restore the view of camera 1 through a virtual view synthesis algorithm and a hole filling technology. First, use the 100-frame texture frame and depth frame in the camera 0 view to obtain the 100-frame virtual texture frame of the camera 1 view through the virtual view synthesis algorithm. Due to the occlusion effect of the foreground, these virtual texture frames all have large holes at the boundary between the foreground and the background, as shown in the gray area behind the person in Figure 5a and b, and Figure 5c and d are the hole filling performed by the method of the present invention respectively. In order to make the virtual view have a better viewing effect, the present invention fills these holes by adopting the adaptive selection method of acquiring the GMM texture background and the FDC texture background.

为获取更佳的GMM纹理背景,将加入深度信息来调整GMM参数更新过程中的学习率。通过观察Microsoft数据集中的Ballet测试序列中的深度图数据,得到用于区别前景背景的经验深度值ε=60,结合每一帧的深度数据,通过公式来调整学习率。将每一帧的数据用于调整高斯模型参数,利用最终的模型参数得到GMM纹理背景。In order to obtain a better GMM texture background, depth information will be added to adjust the learning rate during the GMM parameter update process. By observing the depth map data in the Ballet test sequence in the Microsoft dataset, the empirical depth value ε=60 used to distinguish the foreground and background is obtained, combined with the depth data of each frame, through the formula to adjust the learning rate. The data of each frame is used to adjust the Gaussian model parameters, and the final model parameters are used to obtain the GMM texture background.

为了最终与FDC纹理背景的结合,在获取最终模型参数时,新建一张膜图,通过像素值0和255记录下只含有一个高斯模型和含有多个高斯模型像素所在的位置。In order to finally combine with the FDC texture background, when obtaining the final model parameters, create a new film image, and record the positions of pixels containing only one Gaussian model and multiple Gaussian models through pixel values 0 and 255.

为了得到FDC纹理背景,提前对深度帧进行直方图均衡化处理,将100帧纹理帧和处理后的深度帧分区间进行FDC操作。由于Ballet序列中单个视图有100帧数据,且用该序列进行FDC操作时,测得在30帧左右时所得背景不再发生变化,因此将100帧数据分为3个区间进行FDC操作。将得到的3个区间纹理背景图通过对应的深度背景图来进行有机结合获得最终的FDC纹理背景。In order to obtain the FDC texture background, the histogram equalization processing is performed on the depth frame in advance, and the FDC operation is performed on the 100-frame texture frame and the processed depth frame. Since a single view in the Ballet sequence has 100 frames of data, and when the sequence is used for FDC operation, it is measured that the background does not change at about 30 frames, so the 100 frames of data are divided into 3 intervals for FDC operation. The obtained three interval texture background images are organically combined with the corresponding depth background images to obtain the final FDC texture background.

单个区间的FDC操作如图4所示,对区间内的每一帧深度图通过K-Means方法得到二值深度图。首先初始化一张待填充的背景图,通过第一帧二值深度图确定第一帧纹理图的背景区域,将该纹理区域填入背景,然后通过第二帧二值深度图确定第二帧纹理图的背景区域,将该纹理区域填入还未填充的背景区域,对接下来的每一帧,重复该步骤,逐步完整背景。The FDC operation of a single interval is shown in Figure 4, and the binary depth map is obtained by the K-Means method for each frame depth map in the interval. First initialize a background image to be filled, determine the background area of the first frame texture image through the first frame binary depth image, fill the texture area into the background, and then determine the second frame texture through the second frame binary depth image For the background area of the picture, fill the texture area into the unfilled background area, and repeat this step for each next frame to gradually complete the background.

根据GMM过程得到的膜图,对于最终的背景图,在膜图像素值为0的位置填入GMM纹理背景像素值,否则,填入FDC纹理背景像素。获取最终的背景图后,用虚拟视点合成算法来合成虚拟视点的背景,并在100帧虚拟纹理帧的空洞区域分别用对应区域背景填充。According to the membrane image obtained by the GMM process, for the final background image, fill in the GMM texture background pixel value at the position where the membrane image pixel value is 0, otherwise, fill in the FDC texture background pixel. After obtaining the final background image, use the virtual viewpoint synthesis algorithm to synthesize the background of the virtual viewpoint, and fill the hollow areas of the 100-frame virtual texture frame with the background of the corresponding area.

本领域内的技术人员应明白,本发明的实施例可提供为方法、系统、或计算机程序产品。因此,本发明可采用完全硬件实施例、完全软件实施例、或结合软件和硬件方面的实施例的形式。而且,本发明可采用在一个或多个其中包含有计算机可用程序代码的计算机可用存储介质(包括但不限于磁盘存储器、CD-ROM、光学存储器等)上实施的计算机程序产品的形式。Those skilled in the art should understand that the embodiments of the present invention may be provided as methods, systems, or computer program products. Accordingly, the present invention can take the form of an entirely hardware embodiment, an entirely software embodiment, or an embodiment combining software and hardware aspects. Furthermore, the present invention may take the form of a computer program product embodied on one or more computer-usable storage media (including but not limited to disk storage, CD-ROM, optical storage, etc.) having computer-usable program code embodied therein.

本发明是参照根据本发明实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It should be understood that each procedure and/or block in the flowchart and/or block diagram, and a combination of procedures and/or blocks in the flowchart and/or block diagram can be realized by computer program instructions. These computer program instructions may be provided to a general purpose computer, special purpose computer, embedded processor, or processor of other programmable data processing equipment to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing equipment produce a An apparatus for realizing the functions specified in one or more procedures of the flowchart and/or one or more blocks of the block diagram.

这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to operate in a specific manner, such that the instructions stored in the computer-readable memory produce an article of manufacture comprising instruction means, the instructions The device realizes the function specified in one or more procedures of the flowchart and/or one or more blocks of the block diagram.

这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded onto a computer or other programmable data processing device, causing a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process, thereby The instructions provide steps for implementing the functions specified in the flow chart or blocks of the flowchart and/or the block or blocks of the block diagrams.

尽管已描述了本发明的优选实施例,但本领域内的技术人员一旦得知了基本创造性概念,则可对这些实施例作出另外的变更和修改。所以,所附权利要求意欲解释为包括优选实施例以及落入本发明范围的所有变更和修改。While preferred embodiments of the invention have been described, additional changes and modifications to these embodiments can be made by those skilled in the art once the basic inventive concept is appreciated. Therefore, it is intended that the appended claims be construed to cover the preferred embodiment as well as all changes and modifications which fall within the scope of the invention.

显然,本领域的技术人员可以对本发明实施例进行各种改动和变型而不脱离本发明实施例的精神和范围。这样,倘若本发明实施例的这些修改和变型属于本发明权利要求及其等同技术的范围之内,则本发明也意图包含这些改动和变型在内。Apparently, those skilled in the art can make various changes and modifications to the embodiments of the present invention without departing from the spirit and scope of the embodiments of the present invention. In this way, if the modifications and variations of the embodiments of the present invention fall within the scope of the claims of the present invention and equivalent technologies, the present invention also intends to include these modifications and variations.

Claims (6)

1. A hole filling method based on a depth information assisted Gaussian mixture model is characterized by comprising the following steps:
(1) acquiring image sequence data comprising a plurality of views, wherein the image sequence data of each view comprises texture frames and depth frames of a plurality of frames;
(2) determining a GMM texture background in the image sequence data by adopting a Gaussian mixture model algorithm;
(3) dividing the texture frame and the depth frame into a plurality of intervals according to time sequence, and carrying out histogram equalization processing on the depth frame in each interval;
(4) calculating a foreground depth correlation algorithm for the texture frame in each interval to correspondingly obtain an FDC interval texture background; calculating a foreground depth correlation algorithm for the depth frame after binarization in each interval to correspondingly obtain an FDC interval depth background;
(5) combining the FDC interval texture background obtained from each interval according to the depth information contained in the FDC interval depth background to finally obtain the FDC texture background;
(6) and adaptively generating a reference texture background image by adopting the GMM texture background and the FDC texture background, and further filling the cavity area of the virtual texture frame in the virtual view with the corresponding area background.
2. The method for filling up a hole based on the depth information assisted gaussian mixture model according to claim 1, wherein in the step (2), the GMM texture background in the image sequence data is determined by using a gaussian mixture model algorithm, which specifically comprises:
(21) all texture frames and depth frames in the image sequence data are used in a GMM process, the texture frames are used for constructing a Gaussian model, the depth frames are used for determining model updating parameters, and the method comprises the following steps for single pixel points corresponding to different frames:
(211) the kth Gaussian model of the texture frame of the tth frame is marked as etak,tModel mean is recorded as μk,tAnd the standard deviation is recorded as σk,tAnd the weight is recorded as ωk,tK is more than or equal to 1 and less than or equal to M, and M is the total number of Gaussian models constructed on each pixel point;
(212) calculating the pixel value X of the current pixel point of the next frame t +1tDifference | X from the mean of each Gaussian modeltk,t-1If Xtk,t-1|<2.5σk,tThen, the pixel value of the current pixel point of the next frame is determinedBelonging to the Gaussian model and updating the learning rate of the model;
otherwise, if | Xtk,t-1|≥2.5σk,tIf so, judging that the pixel value of the current pixel point of the next frame does not belong to any one of the M Gaussian models, deleting the Gaussian model with the minimum weight-variance ratio in the M models, and constructing a new Gaussian model for replacement;
(22) all the compounds obtained in step (21) contain one omegak,tWeight ω of other M-1 models, model 1k,tDetermining the pixel point which is 0 as a pixel point which only contains one Gaussian model;
(23) and (3) newly building a membrane map, recording the position of only one Gaussian model by adopting a pixel value of 0, determining the position as a background position, and recording the position of a plurality of Gaussian model pixels by adopting a pixel value of 255, thereby determining the GMM texture background.
3. The method for filling holes in a gaussian mixture model based on depth information assistance according to claim 2, wherein in step (212), the learning rate of the model is updated according to the formula:
where α is the initial learning rate, ε is the empirical value in the sequence, and is the depth pixel value that can distinguish the foreground and background regions observed by the depth frame, dt(x, y) is a pixel value at a (x, y) position on the t-th frame depth frame, αtIs the updated learning rate.
4. The method for filling up a hole based on the depth information assisted gaussian mixture model according to claim 1, wherein in the step (4), the calculation of the foreground depth correlation algorithm is performed on the texture frame in each interval, and the FDC interval texture background is correspondingly obtained, which specifically includes:
(41) generating an initial FDC texture background by using a first frame in a manner of filling a blank image by adopting a region which is distinguished as the background by a binary depth map in the first frame, namely obtaining a background texture region in a first frame texture map corresponding to the background region of the first frame;
(42) updating an unfilled region in the FDC texture background by using a second frame texture map corresponding to a region in which the second frame in the binary depth map of the second frame becomes the background;
(43) and executing the same process as the second frame on all the texture frames in the subsequent interval so as to obtain the FDC interval texture background of the interval.
5. The method for filling up a hole based on the depth information assisted gaussian mixture model as claimed in claim 4, wherein in the step (5), the FDC interval texture background obtained from each interval is combined according to the depth information contained in the FDC interval depth background to finally obtain the FDC texture background, and the method comprises:
for a single pixel point (x, y), the texture background corresponding to the interval depth background with the minimum depth value is used as the final texture background, which is expressed as psi if there isdk(x,y)=min(ψdi(x, y)), thentf(x,y)=ψtk(x, y) whereintf(x, y) is the final FDC texture background at pixel point (x, y), ψdi(x, y) is the depth background of FDC interval corresponding to the ith interval of pixel point (x, y), psitk(x, y) is the texture background of the FDC interval corresponding to the kth interval of the pixel point (x, y); k is the resulting psi with the smallest depth valuediThe number i, i ∈ [1, K ] of (x, y)](ii) a Where t and d in the subscript are used to distinguish between texture and depth map frames.
6. The method for filling holes in a gaussian mixture model based on depth information assistance as claimed in claim 2, wherein in the step (6), the adaptive filling of all image frame numbers by using the GMM texture background and the FDC texture background comprises:
filling pixel points containing a Gaussian model with pixel points at positions corresponding to GMM texture backgrounds; otherwise, filling the pixel points at the corresponding positions of the FDC texture background.
CN201910729173.XA 2019-08-08 2019-08-08 A Hole Filling Method Based on Depth Information Aided Gaussian Mixture Models Active CN110602476B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910729173.XA CN110602476B (en) 2019-08-08 2019-08-08 A Hole Filling Method Based on Depth Information Aided Gaussian Mixture Models

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910729173.XA CN110602476B (en) 2019-08-08 2019-08-08 A Hole Filling Method Based on Depth Information Aided Gaussian Mixture Models

Publications (2)

Publication Number Publication Date
CN110602476A true CN110602476A (en) 2019-12-20
CN110602476B CN110602476B (en) 2021-08-06

Family

ID=68853790

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910729173.XA Active CN110602476B (en) 2019-08-08 2019-08-08 A Hole Filling Method Based on Depth Information Aided Gaussian Mixture Models

Country Status (1)

Country Link
CN (1) CN110602476B (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113487660A (en) * 2021-06-16 2021-10-08 普联国际有限公司 Depth information fused moving target detection method, device, medium and equipment
WO2022022548A1 (en) * 2020-07-31 2022-02-03 阿里巴巴集团控股有限公司 Free viewpoint video reconstruction and playing processing method, device, and storage medium
JP2023524326A (en) * 2021-04-08 2023-06-12 グーグル エルエルシー Neural blending for novel view synthesis

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0330114A (en) * 1989-06-27 1991-02-08 Toshiba Corp Servo equipment and servo method for data recording and reproducing device
US7277118B2 (en) * 1999-08-09 2007-10-02 Fuji Xerox Co., Ltd. Method and system for compensating for parallax in multiple camera systems
CN101640809A (en) * 2009-08-17 2010-02-03 浙江大学 Depth extraction method of merging motion information and geometric information
US20120106791A1 (en) * 2010-10-27 2012-05-03 Samsung Techwin Co., Ltd. Image processing apparatus and method thereof
CN103384343A (en) * 2013-07-02 2013-11-06 南京大学 Image cavity filling method and device thereof
CN104268851A (en) * 2014-09-05 2015-01-07 浙江捷尚视觉科技股份有限公司 ATM self-service business hall behavior analysis method based on depth information
CN108040243A (en) * 2017-12-04 2018-05-15 南京航空航天大学 Multispectral 3-D visual endoscope device and image interfusion method
CN108681999A (en) * 2018-05-22 2018-10-19 浙江理工大学 SAR image target shape generation method based on depth convolutional neural networks model
CN108828481A (en) * 2018-04-24 2018-11-16 朱高杰 A kind of magnetic resonance reconstruction method based on deep learning and data consistency
CN108960080A (en) * 2018-06-14 2018-12-07 浙江工业大学 Based on Initiative Defense image to the face identification method of attack resistance
CN109934195A (en) * 2019-03-21 2019-06-25 东北大学 An anti-spoofing 3D face recognition method based on information fusion

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH0330114A (en) * 1989-06-27 1991-02-08 Toshiba Corp Servo equipment and servo method for data recording and reproducing device
US7277118B2 (en) * 1999-08-09 2007-10-02 Fuji Xerox Co., Ltd. Method and system for compensating for parallax in multiple camera systems
CN101640809A (en) * 2009-08-17 2010-02-03 浙江大学 Depth extraction method of merging motion information and geometric information
US20120106791A1 (en) * 2010-10-27 2012-05-03 Samsung Techwin Co., Ltd. Image processing apparatus and method thereof
CN103384343A (en) * 2013-07-02 2013-11-06 南京大学 Image cavity filling method and device thereof
CN104268851A (en) * 2014-09-05 2015-01-07 浙江捷尚视觉科技股份有限公司 ATM self-service business hall behavior analysis method based on depth information
CN108040243A (en) * 2017-12-04 2018-05-15 南京航空航天大学 Multispectral 3-D visual endoscope device and image interfusion method
CN108828481A (en) * 2018-04-24 2018-11-16 朱高杰 A kind of magnetic resonance reconstruction method based on deep learning and data consistency
CN108681999A (en) * 2018-05-22 2018-10-19 浙江理工大学 SAR image target shape generation method based on depth convolutional neural networks model
CN108960080A (en) * 2018-06-14 2018-12-07 浙江工业大学 Based on Initiative Defense image to the face identification method of attack resistance
CN109934195A (en) * 2019-03-21 2019-06-25 东北大学 An anti-spoofing 3D face recognition method based on information fusion

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
NA-EUN YANG: "Depth Hole Filling Using the Depth Distribution of Neighboring Regions of Depth Holes in the Kinect SENSOR", 《IEEE》 *
曾嘉: "基于图像稀疏性与多尺度神经网络的图像修复算法研究", 《万方》 *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022022548A1 (en) * 2020-07-31 2022-02-03 阿里巴巴集团控股有限公司 Free viewpoint video reconstruction and playing processing method, device, and storage medium
JP2023524326A (en) * 2021-04-08 2023-06-12 グーグル エルエルシー Neural blending for novel view synthesis
JP7519390B2 (en) 2021-04-08 2024-07-19 グーグル エルエルシー Neural Blending for Novel View Synthesis
CN113487660A (en) * 2021-06-16 2021-10-08 普联国际有限公司 Depth information fused moving target detection method, device, medium and equipment

Also Published As

Publication number Publication date
CN110602476B (en) 2021-08-06

Similar Documents

Publication Publication Date Title
US11017586B2 (en) 3D motion effect from a 2D image
US9414048B2 (en) Automatic 2D-to-stereoscopic video conversion
CN110084304B (en) Target detection method based on synthetic data set
CN107578435B (en) A kind of picture depth prediction technique and device
CN108124489B (en) Information processing method, apparatus, cloud processing device and computer program product
JP2012529691A (en) 3D image generation
WO2019071976A1 (en) Panoramic image saliency detection method based on regional growth and eye movement model
CN110602476B (en) A Hole Filling Method Based on Depth Information Aided Gaussian Mixture Models
CN106651853A (en) Establishment method for 3D saliency model based on prior knowledge and depth weight
Lee et al. Silhouette segmentation in multiple views
CN103268604B (en) Binocular video depth map acquiring method
Ward et al. Depth director: A system for adding depth to movies
Jain et al. Enhanced stable view synthesis
US20230080639A1 (en) Techniques for re-aging faces in images and video frames
Woodford et al. On New View Synthesis Using Multiview Stereo.
Mukherjee et al. A hybrid algorithm for disparity calculation from sparse disparity estimates based on stereo vision
Pini et al. Learning to generate facial depth maps
Kong et al. Stereo Matching via Learning Multiple Experts Behaviors.
US20230196659A1 (en) Computer implemented method and system for classifying an input image for new view synthesis in a 3d visual effect, and non-transitory computer readable storage medium
CN113034675B (en) Scene model construction method, intelligent terminal and computer readable storage medium
Saval-Calvo et al. Evaluation of sampling method effects in 3D non-rigid registration
CN108305269B (en) Image segmentation method and system for binocular image
CN115841546A (en) Scene structure associated subway station multi-view vector simulation rendering method and system
KR102181832B1 (en) Apparatus and method for 4d image reconstruction
Vikas et al. Fuzzy C-means clustering and Criminisi algorithm based shadow removal scheme for side scan sonar images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant