CN104484665A - Image foreground extracting method based on Gaussian variation model - Google Patents
Image foreground extracting method based on Gaussian variation model Download PDFInfo
- Publication number
- CN104484665A CN104484665A CN201410740318.3A CN201410740318A CN104484665A CN 104484665 A CN104484665 A CN 104484665A CN 201410740318 A CN201410740318 A CN 201410740318A CN 104484665 A CN104484665 A CN 104484665A
- Authority
- CN
- China
- Prior art keywords
- point
- gauss
- region
- foreground
- deteriorated
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20172—Image enhancement details
Abstract
The invention discloses a novel method for automatic foreground extraction. Under the general conditions, the foreground extraction is interactive, and a foreground region needs an interactive labeling process. According to the method, an image is subjected to Gaussian variation processing to obtain a Gaussian variation point, then, the Gaussian variation point is subjected to marginalization processing by combining a Ncut technology, and marginal information of the image is obtained; finally, the center focus of the image is selected by a certain method, in addition, a key point is subjected to weighted growth in a direction towards the center focus, and a final foreground positioning point is obtained. The foreground region of a target image can be determined according to the proportion of the foreground positioning point in different regions. The whole process is fully automatic, and any manual interaction is not needed. Lots of experiments prove that the method is very effective, and in addition, more precise experiment results can be obtained.
Description
Technical field
The present invention relates generally to image processing techniques, is specifically related to a kind of foreground extracting method be deteriorated based on Gauss.
Background technology
In the research of Computer Image Processing, the task of foreground extraction is by separated for interested object and background from single image or sequence image, for follow-up process.This is a flourishing long time study on classics problem, is also one of most basic research topic of Computer Image Processing.Interesting target generally refers to the moving object in the visual field, needs by analytical sequence Image Acquisition; Or the object of particular type, as face, vehicle etc., can by analyzing single width or sequence image extraction.Effective extraction of foreground area is extremely important for the task that target classification, identification and behavior understanding etc. are on the middle and senior level, because follow-up processing procedure only considers the pixel corresponding to foreground area in image usually, and the research of Video coding, retrieval, man-machine interaction, the problem such as motion-captured will be promoted greatly.But, owing to often there is many uncontrollable factors in actual acquisition environment, as the impact of the problems such as camera motion, illumination variation, shade, complex background, make Objective extraction fast and accurately become a difficult job.
Current foreground extracting method roughly can be divided into two classes: one is interactively foreground extraction, and another kind is automatic foreground extraction.Although these two kinds of methods are all for foreground extraction, the emphasis of these two kinds of methods is not identical.Interactively foreground extracting method lays particular emphasis on and takes foreground area exactly, does accurate process to the edge details of prospect.Interactively method needs artificial marking in target area or fringe region.But for the data of magnanimity on current internet, we obviously cannot carry out artificial mark, so we go to explore automatic foreground extraction.Automatic foreground extraction lays particular emphasis on the foreground area in positioning image, does not consider its details, and this process in interactively method by manually having marked.Automatic foreground extraction has a very wide range of applications prospect, such as image enhaucament, target identification, image index, the aspects such as content-based retrieval.
In these class methods of automatic foreground extraction, many experts have been had to explore.2011, Sungheum proposed a kind of method automatically extracting foreground object from multiple viewpoint.2012, the people such as Zhang Yupeng proposed the automatic prospect background cutting techniques of the degree of depth based on coded aperture.2013, the people such as Xie Chang-ting proposed an automatic trimap generation technique.Although said method achieves good effect, the accuracy and precision identified all has much room for improvement.
Summary of the invention
The object of the invention is to propose a kind of foreground extracting method be deteriorated based on Gauss, to improve the accuracy and precision of extraction prospect.In the invention, this implementation procedure is all completed automatically by computing machine, and user only needs to input target image, just can allow computer automatic analysis, final acquisition display foreground region.
Technical scheme of the present invention is as follows:
Step 1, uses Ncut technology to carry out region segmentation to target image, obtains the region segmentation figure of target image.
Step 2, carries out Edge contrast to former target image, obtains sharpening image, carries out Gauss and to be deteriorated the foreground extraction of model, obtain Gauss and be deteriorated a little in rgb space to sharpening image.
Step 3, is deteriorated a little filters the Gauss obtained in step 2, and acquisition Gauss is deteriorated key point p
(x, y).
First our filter method is exactly first do filtration treatment to noise spot.Meanwhile, image is done to the process on de-black limit, then created a 3*3 filtrator and reduce candidate point.By the filtration of this step, will there is relative equilibrium in the distribution of candidate point.
Step 4, combines the result of step 1 and step 3, obtains the key point K being positioned at Ncut zone boundary
(x, y).
The marginal information of image zones of different is found in the combination of the result that we are deteriorated by Gauss and Ncut, deletes the key point outside edge.After this step filtration treatment, we effectively can navigate to the edge of image.
Due to the performance issue of Ncut segmentation, the borderline region obtained is also unstable.To this, we take the mode of iteration, and the marginal information of Ncut segmentation under reservation different scale, the boundary information obtained like this can more stablize practicality.
Step 5, obtains the gonglion R of former target image
f.
This focus is positioned at foreground area under normal circumstances, centered by the geometric center of figure, according to nine grids model, a figure is equally divided into nine regions.Add up the number that the key point remained through a upper link drops on each region.By the comparison of the key point number to these nine regions, image focal point is positioned at the center comprising the maximum region of key point number in these nine regions.Initial focus is positioned at image geometry center.
Step 6, the gonglion weighted direction key point obtained in step 4 obtained in step 5 increases, and obtains final foreground area anchor point.
After step 5 obtains the focus of image, the focus direction making key point to image increases by we.Each image has four borders.Through analysis to great amount of images, find that the foreground area of most of image all can not be distributed on the border of image, if a panel region comprises many marginal informations simultaneously, so this panel region is that the possibility of foreground area is minimum.Therefore, add certain weight when we do increase process to each key point, the quantity on the border that weights and this region comprise is inversely proportional to.
Step 7, combines the foreground area anchor point of the region segmentation figure of the target image of step 1 gained and step 6 gained, obtains final foreground area.
Retain the cut zone that the foreground area anchor point number comprised is greater than threshold value, remaining region then thinks that background parts is cast out.
Accompanying drawing explanation
Fig. 1 the present invention is based on Gauss to be deteriorated the process flow diagram of display foreground extracting method of model;
Fig. 2 is specific embodiment of the invention procedure chart;
Fig. 3 is step 3 to the procedure chart of processing procedure of the key point obtained that is deteriorated to Gauss of step 6;
Fig. 4 is final sum additive method prospect comparison diagram.
Embodiment
Step 1, uses Ncut technology to carry out region segmentation to target image, obtains the region segmentation figure of target image.
Step 2, carries out Edge contrast to former target image, obtains sharpening image, carries out Gauss and to be deteriorated the foreground extraction of model, obtain Gauss and be deteriorated a little in rgb space to sharpening image.
We do Gaussian convolution to each pixel in image, and as the following formula shown in (1), G is Gaussian function, and I is the pixel in image, and we allow the pixel in image and Gaussian function do convolution.L is the matrix obtained after convolution.σ is the yardstick of Gaussian function, and we are 1.6 for the value of σ.
L(x,y,σ)=G(x,y,σ)*I(x,y) (1)
In formula (2), we obtain Gauss and to be deteriorated difference original image being done to Gaussian convolution exactly under different scale of D, expression.
D(x,y,σ)=L(x,y,kσ)-L(x,y,σ) (2)
We do Gauss's variation to each pixel in image under four kinds of different yardsticks, and result obtains this variation value of three floor heights.For each pixel, we are Gauss's variation value in its middle layer and eight points around it, and these eight point bilevel Gauss's variation values compare, if he is the extreme value of these 27 points, we are then deteriorated this point a little as Gauss.
Step 3, is deteriorated a little filters the Gauss obtained in step 2, and acquisition Gauss is deteriorated key point p
(x, y).
Each Gauss is deteriorated a p
(x, y)value be this point of 1 expression be key point, value is this point of 0 expression is not key point,
Wherein S
(x, y)represent at p
(x, y)the number that Gauss in the eight connectivity region at place is deteriorated a little.The be deteriorated point selected of Gauss is colouring information based on image, and change large pixel for color value and surrounding, the value that its Gauss is deteriorated also can be large.
Step 4, combines the result of step 1 and step 3, obtains the key point K being positioned at Ncut zone boundary
(x, y).
For the key point K obtained in each step 3
(x, y), search in its eight connectivity region, if the institute in its eight connectivity region a little all with this point in same region, then remove this key point, otherwise, be then the key point of zone boundary this point location
K
(x, y)be that 1 this point retains, be 0 and this point is removed from key point.
Step 5, obtains the gonglion R of former target image
f.
Target image is divided into the region R of nine pieces of formed objects by the model of nine grids
i, the step 4 calculated in each block region obtains the quantity of key point, and the region that the Key number finding out step 4 acquisition is maximum, the central point in this region is decided to be the gonglion of target image
Initial focus is defined as the geometric center point R of target image.
Step 6, the gonglion weighted direction key point obtained in step 4 obtained in step 5 increases, and obtains final foreground area anchor point;
The weights W of key point p
pfor:
Key point is with W
pweight increase to gonglion direction.
Step 7, combines the foreground area anchor point of the region segmentation figure of the target image of step 1 gained and step 6 gained, obtains final foreground area.Experiment net result figure as shown in Figure 4.
Retain the cut zone that the foreground area anchor point number comprised is greater than threshold value, remaining region then thinks that background parts is cast out.Threshold value be chosen for 0.2, his implication is the ratio that foreground area anchor point quantity that this region comprises accounts for total foreground area anchor point quantity.
Claims (7)
1. be deteriorated based on Gauss the display foreground extracting method of model, it is characterized in that: when extracting display foreground, carry out following steps:
Step 1, uses Ncut technology to carry out region segmentation to target image, obtains the region segmentation figure of target image;
Step 2, carries out Edge contrast to former target image, obtains sharpening image.In rgb space, carry out Gauss to sharpening image to be deteriorated the foreground extraction of model, obtain Gauss and be deteriorated a little;
Step 3, is deteriorated a little filters the Gauss obtained in step 2, and acquisition Gauss is deteriorated key point p
(x, y);
Step 4, combines the result of step 1 and step 3, obtains the key point K being positioned at Ncut zone boundary
(x, y);
Step 5, obtains the gonglion R of former target image
f;
Step 6, the gonglion weighted direction key point obtained in step 4 obtained in step 5 increases, and obtains final foreground area anchor point;
Step 7, combines the foreground area anchor point of the region segmentation figure of the target image of step 1 gained and step 6 gained, obtains final foreground area.
2. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, is characterized in that: being deteriorated to the Gauss obtained in step 2 described in step 3 is a little carried out filter method and be, to be deteriorated a p for each Gauss
(x, y)value be this point of 1 expression be key point, value is this point of 0 expression is not key point,
Wherein S
(x, y)represent at p
(x, y)the number that Gauss in the eight connectivity region at place is deteriorated a little.
3. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, is characterized in that: the border key point K described in step 4
(x, y)acquisition methods be, for the key point obtained in each step 3, search in its eight connectivity region, if points all in its eight connectivity region all with this point in same region, then remove this point, otherwise, be then the key point of zone boundary this point location
K
(x, y)be that 1 this point retains, be 0 and this point is removed from key point.
4. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, is characterized in that: the acquisition target image gonglion R described in step 5
fmethod be target image is divided into the region R of nine pieces of formed objects by the model of nine grids
i, the step 4 calculated in each block region obtains the quantity of key point, and the region that the Key number finding out step 4 acquisition is maximum, the central point in this region is decided to be the gonglion of target image
Initial focus is defined as the geometric center point R of target image.
5. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, is characterized in that: the method that the gonglion weighted direction that the key point described in step 6 obtains in step 5 increases is, the weights W of key point p
pfor:
Key point is with W
pweight increase to gonglion direction.
6. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 1, it is characterized in that: the region segmentation figure of the target image of step 1 gained described in step 7 and the foreground area anchor point associated methods of step 6 gained are, retain the cut zone that the foreground area anchor point number comprised is greater than threshold value, remaining region then thinks that background parts is cast out.
7. the display foreground extracting method of the model that is deteriorated based on Gauss according to claim 6, it is characterized in that: the region segmentation figure of the target image of step 1 gained described in step 7 and the foreground area anchor point of step 6 gained in conjunction with time, threshold value be chosen for 0.2, his implication is the ratio that foreground area anchor point quantity that this region comprises accounts for total foreground area anchor point quantity.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410740318.3A CN104484665A (en) | 2014-12-05 | 2014-12-05 | Image foreground extracting method based on Gaussian variation model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410740318.3A CN104484665A (en) | 2014-12-05 | 2014-12-05 | Image foreground extracting method based on Gaussian variation model |
Publications (1)
Publication Number | Publication Date |
---|---|
CN104484665A true CN104484665A (en) | 2015-04-01 |
Family
ID=52759206
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410740318.3A Pending CN104484665A (en) | 2014-12-05 | 2014-12-05 | Image foreground extracting method based on Gaussian variation model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104484665A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105227831A (en) * | 2015-09-06 | 2016-01-06 | Tcl集团股份有限公司 | A kind of method and system of self adaptation zoom |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1410942A (en) * | 2001-09-21 | 2003-04-16 | 夏普公司 | Image processing apparatus |
US20090129679A1 (en) * | 2007-11-16 | 2009-05-21 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and computer-readable medium |
-
2014
- 2014-12-05 CN CN201410740318.3A patent/CN104484665A/en active Pending
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1410942A (en) * | 2001-09-21 | 2003-04-16 | 夏普公司 | Image processing apparatus |
US20090129679A1 (en) * | 2007-11-16 | 2009-05-21 | Canon Kabushiki Kaisha | Image processing apparatus, image processing method, and computer-readable medium |
Non-Patent Citations (1)
Title |
---|
YUBO YUAN等: "Automatic Foreground Extraction Based on Difference of Gaussian", 《THE SCIENTIFIC WORLD JOURNAL》 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105227831A (en) * | 2015-09-06 | 2016-01-06 | Tcl集团股份有限公司 | A kind of method and system of self adaptation zoom |
CN105227831B (en) * | 2015-09-06 | 2019-08-06 | Tcl集团股份有限公司 | A kind of method and system of adaptive zoom |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107169487B (en) | Salient object detection method based on superpixel segmentation and depth feature positioning | |
Chen et al. | 3d object proposals using stereo imagery for accurate object class detection | |
Tatem et al. | Super-resolution target identification from remotely sensed images using a Hopfield neural network | |
CN107862698A (en) | Light field foreground segmentation method and device based on K mean cluster | |
CN109285162A (en) | A kind of image, semantic dividing method based on regional area conditional random field models | |
US20150310305A1 (en) | Learning painting styles for painterly rendering | |
CN104134234A (en) | Full-automatic three-dimensional scene construction method based on single image | |
US20160132498A1 (en) | Color Sketch Image Searching | |
CN110991489B (en) | Marking method, device and system for driving data | |
Chen et al. | Optimal segmentation of a high-resolution remote-sensing image guided by area and boundary | |
Yang et al. | MGC-VSLAM: A meshing-based and geometric constraint VSLAM for dynamic indoor environments | |
CN106709456B (en) | Unmanned aerial vehicle target tracking frame initialization method based on computer vision | |
Chen et al. | Photogrammetric point cloud segmentation and object information extraction for creating virtual environments and simulations | |
CN105513105A (en) | Image background blurring method based on saliency map | |
CN107808524B (en) | Road intersection vehicle detection method based on unmanned aerial vehicle | |
CN103577875A (en) | CAD (computer-aided design) people counting method based on FAST (features from accelerated segment test) | |
CN105550688A (en) | Method and device for point cloud data classification | |
CN110147841A (en) | The fine grit classification method for being detected and being divided based on Weakly supervised and unsupervised component | |
CN112686902A (en) | Two-stage calculation method for brain glioma identification and segmentation in nuclear magnetic resonance image | |
CN113052755A (en) | High-resolution image intelligent matting method based on deep learning | |
CN104112279A (en) | Target detection method and device | |
CN103955942A (en) | SVM-based depth map extraction method of 2D image | |
CN107103608A (en) | A kind of conspicuousness detection method based on region candidate samples selection | |
Zou et al. | Boundary-aware CNN for semantic segmentation | |
CN102163285A (en) | Cross-domain video semantic concept detection method based on active learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20150401 |
|
WD01 | Invention patent application deemed withdrawn after publication |