CN104484665A - Image foreground extracting method based on Gaussian variation model - Google Patents

Image foreground extracting method based on Gaussian variation model Download PDF

Info

Publication number
CN104484665A
CN104484665A CN201410740318.3A CN201410740318A CN104484665A CN 104484665 A CN104484665 A CN 104484665A CN 201410740318 A CN201410740318 A CN 201410740318A CN 104484665 A CN104484665 A CN 104484665A
Authority
CN
China
Prior art keywords
point
gauss
region
foreground
deteriorated
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201410740318.3A
Other languages
Chinese (zh)
Inventor
袁玉波
刘赟
戴光辉
陈志华
张静
应方立
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
East China University of Science and Technology
Original Assignee
East China University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by East China University of Science and Technology filed Critical East China University of Science and Technology
Priority to CN201410740318.3A priority Critical patent/CN104484665A/en
Publication of CN104484665A publication Critical patent/CN104484665A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20172Image enhancement details

Abstract

The invention discloses a novel method for automatic foreground extraction. Under the general conditions, the foreground extraction is interactive, and a foreground region needs an interactive labeling process. According to the method, an image is subjected to Gaussian variation processing to obtain a Gaussian variation point, then, the Gaussian variation point is subjected to marginalization processing by combining a Ncut technology, and marginal information of the image is obtained; finally, the center focus of the image is selected by a certain method, in addition, a key point is subjected to weighted growth in a direction towards the center focus, and a final foreground positioning point is obtained. The foreground region of a target image can be determined according to the proportion of the foreground positioning point in different regions. The whole process is fully automatic, and any manual interaction is not needed. Lots of experiments prove that the method is very effective, and in addition, more precise experiment results can be obtained.

Description

A kind of display foreground extracting method of the model that is deteriorated based on Gauss
Technical field
The present invention relates generally to image processing techniques, is specifically related to a kind of foreground extracting method be deteriorated based on Gauss.
Background technology
In the research of Computer Image Processing, the task of foreground extraction is by separated for interested object and background from single image or sequence image, for follow-up process.This is a flourishing long time study on classics problem, is also one of most basic research topic of Computer Image Processing.Interesting target generally refers to the moving object in the visual field, needs by analytical sequence Image Acquisition; Or the object of particular type, as face, vehicle etc., can by analyzing single width or sequence image extraction.Effective extraction of foreground area is extremely important for the task that target classification, identification and behavior understanding etc. are on the middle and senior level, because follow-up processing procedure only considers the pixel corresponding to foreground area in image usually, and the research of Video coding, retrieval, man-machine interaction, the problem such as motion-captured will be promoted greatly.But, owing to often there is many uncontrollable factors in actual acquisition environment, as the impact of the problems such as camera motion, illumination variation, shade, complex background, make Objective extraction fast and accurately become a difficult job.
Current foreground extracting method roughly can be divided into two classes: one is interactively foreground extraction, and another kind is automatic foreground extraction.Although these two kinds of methods are all for foreground extraction, the emphasis of these two kinds of methods is not identical.Interactively foreground extracting method lays particular emphasis on and takes foreground area exactly, does accurate process to the edge details of prospect.Interactively method needs artificial marking in target area or fringe region.But for the data of magnanimity on current internet, we obviously cannot carry out artificial mark, so we go to explore automatic foreground extraction.Automatic foreground extraction lays particular emphasis on the foreground area in positioning image, does not consider its details, and this process in interactively method by manually having marked.Automatic foreground extraction has a very wide range of applications prospect, such as image enhaucament, target identification, image index, the aspects such as content-based retrieval.
In these class methods of automatic foreground extraction, many experts have been had to explore.2011, Sungheum proposed a kind of method automatically extracting foreground object from multiple viewpoint.2012, the people such as Zhang Yupeng proposed the automatic prospect background cutting techniques of the degree of depth based on coded aperture.2013, the people such as Xie Chang-ting proposed an automatic trimap generation technique.Although said method achieves good effect, the accuracy and precision identified all has much room for improvement.
Summary of the invention
The object of the invention is to propose a kind of foreground extracting method be deteriorated based on Gauss, to improve the accuracy and precision of extraction prospect.In the invention, this implementation procedure is all completed automatically by computing machine, and user only needs to input target image, just can allow computer automatic analysis, final acquisition display foreground region.
Technical scheme of the present invention is as follows:
Step 1, uses Ncut technology to carry out region segmentation to target image, obtains the region segmentation figure of target image.
Step 2, carries out Edge contrast to former target image, obtains sharpening image, carries out Gauss and to be deteriorated the foreground extraction of model, obtain Gauss and be deteriorated a little in rgb space to sharpening image.
Step 3, is deteriorated a little filters the Gauss obtained in step 2, and acquisition Gauss is deteriorated key point p (x, y).
First our filter method is exactly first do filtration treatment to noise spot.Meanwhile, image is done to the process on de-black limit, then created a 3*3 filtrator and reduce candidate point.By the filtration of this step, will there is relative equilibrium in the distribution of candidate point.
Step 4, combines the result of step 1 and step 3, obtains the key point K being positioned at Ncut zone boundary (x, y).
The marginal information of image zones of different is found in the combination of the result that we are deteriorated by Gauss and Ncut, deletes the key point outside edge.After this step filtration treatment, we effectively can navigate to the edge of image.
Due to the performance issue of Ncut segmentation, the borderline region obtained is also unstable.To this, we take the mode of iteration, and the marginal information of Ncut segmentation under reservation different scale, the boundary information obtained like this can more stablize practicality.
Step 5, obtains the gonglion R of former target image f.
This focus is positioned at foreground area under normal circumstances, centered by the geometric center of figure, according to nine grids model, a figure is equally divided into nine regions.Add up the number that the key point remained through a upper link drops on each region.By the comparison of the key point number to these nine regions, image focal point is positioned at the center comprising the maximum region of key point number in these nine regions.Initial focus is positioned at image geometry center.
Step 6, the gonglion weighted direction key point obtained in step 4 obtained in step 5 increases, and obtains final foreground area anchor point.
After step 5 obtains the focus of image, the focus direction making key point to image increases by we.Each image has four borders.Through analysis to great amount of images, find that the foreground area of most of image all can not be distributed on the border of image, if a panel region comprises many marginal informations simultaneously, so this panel region is that the possibility of foreground area is minimum.Therefore, add certain weight when we do increase process to each key point, the quantity on the border that weights and this region comprise is inversely proportional to.
Step 7, combines the foreground area anchor point of the region segmentation figure of the target image of step 1 gained and step 6 gained, obtains final foreground area.
Retain the cut zone that the foreground area anchor point number comprised is greater than threshold value, remaining region then thinks that background parts is cast out.
Accompanying drawing explanation
Fig. 1 the present invention is based on Gauss to be deteriorated the process flow diagram of display foreground extracting method of model;
Fig. 2 is specific embodiment of the invention procedure chart;
Fig. 3 is step 3 to the procedure chart of processing procedure of the key point obtained that is deteriorated to Gauss of step 6;
Fig. 4 is final sum additive method prospect comparison diagram.
Embodiment
Step 1, uses Ncut technology to carry out region segmentation to target image, obtains the region segmentation figure of target image.
Step 2, carries out Edge contrast to former target image, obtains sharpening image, carries out Gauss and to be deteriorated the foreground extraction of model, obtain Gauss and be deteriorated a little in rgb space to sharpening image.
We do Gaussian convolution to each pixel in image, and as the following formula shown in (1), G is Gaussian function, and I is the pixel in image, and we allow the pixel in image and Gaussian function do convolution.L is the matrix obtained after convolution.σ is the yardstick of Gaussian function, and we are 1.6 for the value of σ.
L(x,y,σ)=G(x,y,σ)*I(x,y) (1)
In formula (2), we obtain Gauss and to be deteriorated difference original image being done to Gaussian convolution exactly under different scale of D, expression.
D(x,y,σ)=L(x,y,kσ)-L(x,y,σ) (2)
We do Gauss's variation to each pixel in image under four kinds of different yardsticks, and result obtains this variation value of three floor heights.For each pixel, we are Gauss's variation value in its middle layer and eight points around it, and these eight point bilevel Gauss's variation values compare, if he is the extreme value of these 27 points, we are then deteriorated this point a little as Gauss.
Step 3, is deteriorated a little filters the Gauss obtained in step 2, and acquisition Gauss is deteriorated key point p (x, y).
Each Gauss is deteriorated a p (x, y)value be this point of 1 expression be key point, value is this point of 0 expression is not key point,
p ( x , y ) = 0 , S ( x , y ) > 1 1 , S ( x , y ) ≤ 1 - - - ( 3 )
Wherein S (x, y)represent at p (x, y)the number that Gauss in the eight connectivity region at place is deteriorated a little.The be deteriorated point selected of Gauss is colouring information based on image, and change large pixel for color value and surrounding, the value that its Gauss is deteriorated also can be large.
Step 4, combines the result of step 1 and step 3, obtains the key point K being positioned at Ncut zone boundary (x, y).
For the key point K obtained in each step 3 (x, y), search in its eight connectivity region, if the institute in its eight connectivity region a little all with this point in same region, then remove this key point, otherwise, be then the key point of zone boundary this point location
K (x, y)be that 1 this point retains, be 0 and this point is removed from key point.
Step 5, obtains the gonglion R of former target image f.
Target image is divided into the region R of nine pieces of formed objects by the model of nine grids i, the step 4 calculated in each block region obtains the quantity of key point, and the region that the Key number finding out step 4 acquisition is maximum, the central point in this region is decided to be the gonglion of target image
Initial focus is defined as the geometric center point R of target image.
Step 6, the gonglion weighted direction key point obtained in step 4 obtained in step 5 increases, and obtains final foreground area anchor point;
The weights W of key point p pfor:
Key point is with W pweight increase to gonglion direction.
Step 7, combines the foreground area anchor point of the region segmentation figure of the target image of step 1 gained and step 6 gained, obtains final foreground area.Experiment net result figure as shown in Figure 4.
Retain the cut zone that the foreground area anchor point number comprised is greater than threshold value, remaining region then thinks that background parts is cast out.Threshold value be chosen for 0.2, his implication is the ratio that foreground area anchor point quantity that this region comprises accounts for total foreground area anchor point quantity.

Claims (7)

1. be deteriorated based on Gauss the display foreground extracting method of model, it is characterized in that: when extracting display foreground, carry out following steps:
Step 1, uses Ncut technology to carry out region segmentation to target image, obtains the region segmentation figure of target image;
Step 2, carries out Edge contrast to former target image, obtains sharpening image.In rgb space, carry out Gauss to sharpening image to be deteriorated the foreground extraction of model, obtain Gauss and be deteriorated a little;
Step 3, is deteriorated a little filters the Gauss obtained in step 2, and acquisition Gauss is deteriorated key point p (x, y);
Step 4, combines the result of step 1 and step 3, obtains the key point K being positioned at Ncut zone boundary (x, y);
Step 5, obtains the gonglion R of former target image f;
Step 6, the gonglion weighted direction key point obtained in step 4 obtained in step 5 increases, and obtains final foreground area anchor point;
Step 7, combines the foreground area anchor point of the region segmentation figure of the target image of step 1 gained and step 6 gained, obtains final foreground area.
2. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, is characterized in that: being deteriorated to the Gauss obtained in step 2 described in step 3 is a little carried out filter method and be, to be deteriorated a p for each Gauss (x, y)value be this point of 1 expression be key point, value is this point of 0 expression is not key point,
p ( x , y ) = 0 , S ( x , y ) > 1 1 , S ( x , y ) ≤ 1 - - - ( 1 )
Wherein S (x, y)represent at p (x, y)the number that Gauss in the eight connectivity region at place is deteriorated a little.
3. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, is characterized in that: the border key point K described in step 4 (x, y)acquisition methods be, for the key point obtained in each step 3, search in its eight connectivity region, if points all in its eight connectivity region all with this point in same region, then remove this point, otherwise, be then the key point of zone boundary this point location
K (x, y)be that 1 this point retains, be 0 and this point is removed from key point.
4. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, is characterized in that: the acquisition target image gonglion R described in step 5 fmethod be target image is divided into the region R of nine pieces of formed objects by the model of nine grids i, the step 4 calculated in each block region obtains the quantity of key point, and the region that the Key number finding out step 4 acquisition is maximum, the central point in this region is decided to be the gonglion of target image
Initial focus is defined as the geometric center point R of target image.
5. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, is characterized in that: the method that the gonglion weighted direction that the key point described in step 6 obtains in step 5 increases is, the weights W of key point p pfor:
Key point is with W pweight increase to gonglion direction.
6. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, it is characterized in that: the region segmentation figure of the target image of step 1 gained described in step 7 and the foreground area anchor point associated methods of step 6 gained are, retain the cut zone that the foreground area anchor point number comprised is greater than threshold value, remaining region then thinks that background parts is cast out.
7. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 6, it is characterized in that: the region segmentation figure of the target image of step 1 gained described in step 7 and the foreground area anchor point of step 6 gained in conjunction with time, threshold value be chosen for 0.2, his implication is the ratio that foreground area anchor point quantity that this region comprises accounts for total foreground area anchor point quantity.
CN201410740318.3A 2014-12-05 2014-12-05 Image foreground extracting method based on Gaussian variation model Pending CN104484665A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410740318.3A CN104484665A (en) 2014-12-05 2014-12-05 Image foreground extracting method based on Gaussian variation model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410740318.3A CN104484665A (en) 2014-12-05 2014-12-05 Image foreground extracting method based on Gaussian variation model

Publications (1)

Publication Number Publication Date
CN104484665A true CN104484665A (en) 2015-04-01

Family

ID=52759206

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410740318.3A Pending CN104484665A (en) 2014-12-05 2014-12-05 Image foreground extracting method based on Gaussian variation model

Country Status (1)

Country Link
CN (1) CN104484665A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105227831A (en) * 2015-09-06 2016-01-06 Tcl集团股份有限公司 A kind of method and system of self adaptation zoom

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1410942A (en) * 2001-09-21 2003-04-16 夏普公司 Image processing apparatus
US20090129679A1 (en) * 2007-11-16 2009-05-21 Canon Kabushiki Kaisha Image processing apparatus, image processing method, and computer-readable medium

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1410942A (en) * 2001-09-21 2003-04-16 夏普公司 Image processing apparatus
US20090129679A1 (en) * 2007-11-16 2009-05-21 Canon Kabushiki Kaisha Image processing apparatus, image processing method, and computer-readable medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
YUBO YUAN等: "Automatic Foreground Extraction Based on Difference of Gaussian", 《THE SCIENTIFIC WORLD JOURNAL》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105227831A (en) * 2015-09-06 2016-01-06 Tcl集团股份有限公司 A kind of method and system of self adaptation zoom
CN105227831B (en) * 2015-09-06 2019-08-06 Tcl集团股份有限公司 A kind of method and system of adaptive zoom

Similar Documents

Publication Publication Date Title
CN107169487B (en) Salient object detection method based on superpixel segmentation and depth feature positioning
Chen et al. 3d object proposals using stereo imagery for accurate object class detection
Tatem et al. Super-resolution target identification from remotely sensed images using a Hopfield neural network
CN107862698A (en) Light field foreground segmentation method and device based on K mean cluster
CN109285162A (en) A kind of image, semantic dividing method based on regional area conditional random field models
US20150310305A1 (en) Learning painting styles for painterly rendering
CN104134234A (en) Full-automatic three-dimensional scene construction method based on single image
US20160132498A1 (en) Color Sketch Image Searching
CN110991489B (en) Marking method, device and system for driving data
Chen et al. Optimal segmentation of a high-resolution remote-sensing image guided by area and boundary
Yang et al. MGC-VSLAM: A meshing-based and geometric constraint VSLAM for dynamic indoor environments
CN106709456B (en) Unmanned aerial vehicle target tracking frame initialization method based on computer vision
Chen et al. Photogrammetric point cloud segmentation and object information extraction for creating virtual environments and simulations
CN105513105A (en) Image background blurring method based on saliency map
CN107808524B (en) Road intersection vehicle detection method based on unmanned aerial vehicle
CN103577875A (en) CAD (computer-aided design) people counting method based on FAST (features from accelerated segment test)
CN105550688A (en) Method and device for point cloud data classification
CN110147841A (en) The fine grit classification method for being detected and being divided based on Weakly supervised and unsupervised component
CN112686902A (en) Two-stage calculation method for brain glioma identification and segmentation in nuclear magnetic resonance image
CN113052755A (en) High-resolution image intelligent matting method based on deep learning
CN104112279A (en) Target detection method and device
CN103955942A (en) SVM-based depth map extraction method of 2D image
CN107103608A (en) A kind of conspicuousness detection method based on region candidate samples selection
Zou et al. Boundary-aware CNN for semantic segmentation
CN102163285A (en) Cross-domain video semantic concept detection method based on active learning

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20150401

WD01 Invention patent application deemed withdrawn after publication