CN105513070A - RGB-D salient object detection method based on foreground and background optimization - Google Patents

RGB-D salient object detection method based on foreground and background optimization Download PDF

Info

Publication number
CN105513070A
CN105513070A CN201510897635.0A CN201510897635A CN105513070A CN 105513070 A CN105513070 A CN 105513070A CN 201510897635 A CN201510897635 A CN 201510897635A CN 105513070 A CN105513070 A CN 105513070A
Authority
CN
China
Prior art keywords
pixel
super
significantly
prospect
rgb
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510897635.0A
Other languages
Chinese (zh)
Inventor
周圆
陈阳
崔波
霍树伟
侯春萍
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201510897635.0A priority Critical patent/CN105513070A/en
Publication of CN105513070A publication Critical patent/CN105513070A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection

Abstract

The invention discloses an RGB-D salient object detection method based on foreground and background optimization. The method comprises the following steps: initial foreground modeling is performed based on low-level feature contrast, and a superpixel-level initial salient figure is obtained; a middle-level aggregation processing is performed on the superpixel-level initial salient figure, and a middle-level salient figure is obtained; a high-level prior is introduced in the middle-level salient figure to improve the detection effect, and a foreground probability is generated; edge connectivity mixing depth information is calculated, and the edge connectivity is converted into a background probability; the foreground probability and the background probability are optimized, and a objective function is obtained; the objective function is solved, a optimal salient figure is obtained, and the detection of a salient object is realized. According to the invention, a optimization framework based on foreground and background measurement and the depth information of a scene is fully utilized by the invention, a high recall rate can be obtained, and the accuracy is high; the method can accurately position the salient object in different scenes and different sizes of objects and can also obtain nearly equal salience values in the target object.

Description

A kind of RGB-D obvious object detection method optimized based on prospect background
Technical field
The present invention relates to the detection field of computer vision, particularly relate to a kind of RGB-D (colored and depth image) obvious object merging depth information and detect.
Background technology
At computer vision field, detecting from natural scene and being partitioned into obvious object is an active problem, also creates a lot of significant application.Current most obvious object detection method utilizes colouring information and multiple prior imformation to obtain good result.Although scene depth plays very important role in the vision system of the mankind, it obvious object detect in effect also do not excavated by the degree of depth.The current appearance due to a series of depth camera, makes to obtain depth information of scene more convenient.
It is a kind of well clue that scene depth detects for conspicuousness.Compared with detecting with 2D conspicuousness, the work of minority is only had to be detect this task for the conspicuousness based on RGB-D data.The people such as Desingh [1]propose a kind of method of compute depth conspicuousness, then by Support vector regression, degree of depth significance and 2D significance are merged.But this method carries out significance calculating to colouring information and depth information respectively, have ignored these two kinds of information correlativity each other.The people such as Ciptadi [2]the 3D profile of explicit extraction scene, shape and color characteristic simply, significance is defined as the diversity between these features, but the result that the method draws does not carry out depth optimization, Detection results is not very desirable.The people such as Peng [3]construct first database being exclusively used in obvious object and detecting, and propose a kind of multistage conspicuousness detection algorithm.But its performance is still limited.
Up to now, there is not yet the RGB-D obvious object detection algorithm carried out about based on prospect background optimize in the paper published and document at home and abroad.
Summary of the invention
The invention provides a kind of RGB-D obvious object detection method optimized based on prospect background, the present invention makes full use of the depth information of image, introduces the framework optimized based on prospect background, to realize the accurate detection of obvious object, described below:
Based on the RGB-D obvious object detection method that prospect background is optimized, described obvious object detection method comprises the following steps:
Carry out initial prospect modeling based on low-level image feature contrast, obtain the initial significantly figure of super-pixel rank; Middle layers of polymer process is carried out to the initial significantly figure of described super-pixel rank, obtains middle rank significantly figure;
Middle rank is significantly schemed to introduce high-rise priori and promotes Detection results further, generate prospect probability;
Calculate the contour connection degree merging depth information, described contour connection degree is converted into background probability;
Described prospect probability and described background probability are optimized, obtain objective function;
Described objective function is solved, obtains optimum significantly figure, realize the detection to obvious object.
Wherein, described obvious object detection method also comprises:
To a given input RGB-D image, by over-segmentation algorithm, RGB-D Iamge Segmentation is become super-pixel.
Wherein, describedly carry out initial prospect modeling based on low-level image feature contrast, the step obtaining the initial significantly figure of super-pixel rank is specially:
Each super-pixel proper vector is represented; Define multiple contextual information;
By proper vector, contextual information and Weighted Gauss cuclear density to each super-pixel calculating probability density, corresponding probability density value is assigned to corresponding super-pixel, gets the initial significantly figure of super-pixel rank.
Further, described contextual information comprises:
1) nearest with current super-pixel n lindividual adjacent super-pixel;
2) comprise except current super-pixel, the every other super-pixel in image;
3) super-pixel on four angles of image is comprised.
Wherein, the described initial significantly figure to described super-pixel rank carries out middle layers of polymer process, obtains the remarkable figure of middle rank and is specially:
Weighting function is obtained by the weight parameter after estimation and side-play amount;
Carry out spanning tree process by weighting function in conjunction with the initial significantly figure of spanning tree algorithm to described super-pixel rank, obtain middle rank significantly figure.
Wherein, described by carrying out spanning tree process to the initial significantly figure of described super-pixel rank, the step obtaining the remarkable figure of middle rank is specially:
Generate a conspicuousness subset, make it to comprise m the super-pixel with higher saliency value;
The set of iterative tree is carried out to each conspicuousness seed in conspicuousness subset, selects the fillet with weight limit, and add in spanning tree, export spanning tree;
Calculate the frequency that each super-pixel occurs in spanning tree, in this, as saliency value, generate middle rank significantly figure.
Describedly significantly scheme to introduce high-rise priori to middle rank and promote Detection results further, the step generating prospect probability is specially:
By significantly to scheme described middle rank and Gauss model carries out priori and is multiplied, obtain senior remarkable figure, in described senior remarkable figure, the saliency value of each super-pixel is the prospect probability obtained by bottom, middle level, high-rise prospect modeling.
The beneficial effect of technical scheme provided by the invention is: this method takes full advantage of the depth information of Optimization Framework and the scene of measuring based on prospect background, and can obtain very high recall rate, degree of accuracy is also very high.For different scene type and dimension of object, this method can accurately locate remarkable object, and provides significance value almost equal in target object.Background and marking area discrimination less time the method still can play a role well.
In addition, this method has carried out objective examination in disclosed NLPR image data base, adopt accuracy rate (Precision), recall rate (Recall), curve lower area area (AUC) and F-measure tetra-indexs to carry out the effect of quantitative measurement RGB-D obvious object detection, fully demonstrate the accuracy of this method.
Accompanying drawing explanation
Fig. 1 is the process flow diagram of the RGB-D obvious object detection method based on prospect background optimization;
Fig. 2 is that this method compares schematic diagram with the PR curve with reference to algorithm;
Fig. 3 is that this method compares schematic diagram with the ROC curve with reference to algorithm;
Fig. 4 is that this method compares schematic diagram with AUC and F-measure with reference to algorithm;
Fig. 5 is that the qualitative results of RGB-D image compares schematic diagram.
Embodiment
For making the object, technical solutions and advantages of the present invention clearly, below embodiment of the present invention is described further in detail.
The feature that this method detects from obvious object and based on prospect background Optimization Framework, how research carries out effective prospect tolerance and background is measured.In prospect tolerance, this method takes a kind of multistage prospect measure, considers that the low layer of image, middle level and high-rise characteristic extract the most reliable prospect part effectively.In background tolerance, this method carries out modeling by proposing a kind of contour connection degree merging the degree of depth to image background.Prospect and background tolerance bring the majorized function of an appropriate design into the most at last, are significantly schemed by effective optimization.
Embodiment 1
Based on the RGB-D obvious object detection method that prospect background is optimized, see Fig. 1, this obvious object detection method comprises the following steps: 101: carry out initial prospect modeling based on low-level image feature contrast, obtains the initial significantly figure of super-pixel rank; Middle layers of polymer process is carried out to the initial significantly figure of super-pixel rank, obtains middle rank significantly figure;
102: middle rank is significantly schemed to introduce high-rise priori and promotes Detection results further, generate prospect probability;
103: calculate the contour connection degree merging depth information, contour connection degree is converted into background probability;
104: prospect probability and described background probability are optimized, obtain objective function;
105: objective function is solved, obtain optimum significantly figure, realize the detection to obvious object.
Wherein, this obvious object detection method also comprises:
To a given input RGB-D image, by over-segmentation algorithm, RGB-D Iamge Segmentation is become super-pixel.
Wherein, carry out initial prospect modeling in step 101 based on low-level image feature contrast, the step obtaining the initial significantly figure of super-pixel rank is specially:
Each super-pixel proper vector is represented; Define multiple contextual information;
By proper vector, contextual information and Weighted Gauss cuclear density to each super-pixel calculating probability density, corresponding probability density value is assigned to corresponding super-pixel, gets the initial significantly figure of super-pixel rank.
Wherein, contextual information comprises:
1) nearest with current super-pixel n lindividual adjacent super-pixel;
2) comprise except current super-pixel, the every other super-pixel in image;
3) super-pixel on four angles of image is comprised.
Wherein, in step 101, middle layers of polymer process is carried out to the initial significantly figure of super-pixel rank, obtain middle rank significantly figure be specially:
Weighting function is obtained by the weight parameter after estimation and side-play amount;
Carry out spanning tree process by weighting function in conjunction with the initial significantly figure of spanning tree algorithm to super-pixel rank, obtain middle rank significantly figure.
Further, by carrying out spanning tree process to the initial significantly figure of super-pixel rank, the step obtaining the remarkable figure of middle rank is specially:
Generate a conspicuousness subset, make it to comprise m the super-pixel with higher saliency value;
The set of iterative tree is carried out to each conspicuousness seed in conspicuousness subset, selects the fillet with weight limit, and add in spanning tree, export spanning tree;
Calculate the frequency that each super-pixel occurs in spanning tree, in this, as saliency value, generate middle rank significantly figure.
Wherein, middle rank significantly schemed to introduce high-rise priori promote Detection results further in step 102, the step generating prospect probability is specially:
By significantly to scheme middle rank and Gauss model carries out priori and is multiplied, obtain senior remarkable figure, in senior remarkable figure, the saliency value of each super-pixel is the prospect probability obtained by bottom, middle level, high-rise prospect modeling.
In sum, the embodiment of the present invention makes full use of the depth information of image by above-mentioned steps 101-step 105, introduces the framework optimized based on prospect background, achieves the accurate detection of obvious object.
Embodiment 2
Below in conjunction with Fig. 1, concrete computing formula, the scheme in embodiment 1 is introduced, described below:
201: to a given input RGB-D image, by over-segmentation algorithm, RGB-D Iamge Segmentation is become super-pixel;
This step is specially: each pixel in image is expressed as sextuple proper vector, [L, a, b] for pixel is at the color characteristic of CIElab (a kind of color mode that International Commission on Illumination announced in 1976) color space, wherein Lab pattern is made up of three passages, first passage is lightness, i.e. " L ".The color of " a " passage is from redness to dark green; " b " passage is from blueness to yellow then.The volume coordinate that [x, y, z] is pixel; Then the color distance d between two pixels (i, j) is defined cwith space length d s; Finally obtain the distance metric d between pixel.
Wherein, d c = ( L i - L j ) 2 + ( a i - a j ) 2 + ( b i - b j ) 2 d s = ( x i - x j ) 2 + ( y i - y j ) 2 + ( z i - z j ) 2 utilize this distance metric formula, RGB-D Iamge Segmentation can be become super-pixel.
Wherein, L iit is the lightness of i-th pixel; L jfor the lightness of a jth pixel; a iit is a channel value of i-th pixel; a jfor a channel value of a jth pixel; b iit is the b channel value of i-th pixel; b jfor the b channel value of a jth pixel; x iit is the abscissa value of i-th pixel; x jfor the abscissa value of a jth pixel; y iit is the ordinate value of i-th pixel; y jfor the ordinate value of a jth pixel; z iit is the depth value of i-th pixel; z jfor the depth value of a jth pixel; M is regular shape degree coefficient; K is the total number of image pixel; N is the super-pixel number needing to generate.
202: after RGB-D Iamge Segmentation is become super-pixel, carry out initial prospect modeling based on low-level image feature contrast, obtain the initial significantly figure of super-pixel rank;
For each super-pixel P, its proper vector is defined as f=[c, l, r, d], and wherein c represents in CIELab color space, the average color of all pixels in P; L represents the center of super-pixel P on imaging plane; R represents the size in super-pixel region, i.e. the number of pixel in super-pixel; D represents the mean depth of all pixels in super-pixel P.
Define multiple contextual information ψ k, k ∈ L, G, B}, wherein represent the set of local context, comprise and n that this super-pixel is nearest lindividual adjacent super-pixel. represent global context set, contain the every other super-pixel except this super-pixel in image. represent background set of context, contain the super-pixel on four angles of image.
Use feature defined above and multiple contextual information, carry out estimated probability by the method for a kind of Weighted Gauss cuclear density modeling.Weighted Gauss nuclear density model is defined as:
p ( P | ψ k ) = p ^ ( f | F k ) = 1 n k Σ j = 1 n k α j k e - | | c - c j k | | 2 2 ( σ c k ) 2 e - | | l - l j k | | 2 2 ( σ l k ) 2 e - | | d - d j k | | 2 2 ( σ d k ) 2
Wherein, p (P| ψ k) be the potential probability density of super-pixel P under given set of context; be and super-pixel set ψ kcorresponding character representation set; for super-pixel P characteristic of correspondence vector f=[c, l, r, d] is at given character representation set F kunder potential probability density; n kfor the super-pixel number adjacent with super-pixel P; for the average color of a jth super-pixel; for the center of a jth super-pixel on imaging plane; d j kfor the mean depth of all pixels in a jth super-pixel; for the Gaussian function coefficient of variation of color item; for the Gaussian function coefficient of variation of location entries; for the Gaussian function coefficient of variation of degree of depth item; be a kind of weight coefficient, be defined as target super-pixel P and context super-pixel size ratio.
By above-mentioned Weighted Gauss cuclear density formula, its probability density is calculated to each super-pixel, corresponding probability density value is assigned to corresponding super-pixel, just can obtain the initial significantly figure of super-pixel rank.
203: middle layers of polymer process is carried out to the initial significantly figure of super-pixel rank, obtains middle rank significantly figure.
Another G=(ν, ε, ρ) represents the upper weighting connection layout formed of super-pixel segmentation on RGB-D image, and the summit ν of figure represents each super-pixel, and limit (P, Q) ∈ ε connects adjacent super-pixel P and Q.Weighting function ρ: ε → [0,1] is that the limit connecting super-pixel P and Q assigns weight, and the method uses logistic function [4]for ρ modeling is as follows:
ρ P,Q=σ(w Tx P,Q+b),σ(h)=(1+exp(-h)) -1
Wherein, h represents w tx p,Q+ b, x p,Qbe a proper vector, contain the validity feature that similarity and the compatibility of super-pixel P and Q are measured.σ is sigmoid function [4].Weight parameter w and side-play amount b comes from training data learning.In order to reach this target of training, first each super-pixel is divided into foreground area or background area by the method.Partitioning standards is: for a super-pixel, if belong to the ratio of conspicuousness object more than 80% in its pixel of comprising, is so just divided into prospect part, otherwise is divided into background parts.So for all super-pixel in entire image, the method is gone to find and may be included in conspicuousness object, and is its distributing labels, thinks that it is positive sample (y p,Q=1), otherwise, for it distributes the label (y of negative sample p,Q=0).
By maximizing the estimation obtaining weight parameter w and side-play amount b with minor function:
{ w , b } = argmax w , b Σ ∀ ( P , Q ) ∈ ϵ y P , Q logρ P , Q + ( 1 - y P , Q ) l o g ( 1 - ρ P , Q )
Wherein, { w, b} are the vector of weight parameter w and side-play amount b composition; y p,Qfor the label value distributed; ρ p,Qfor the weighting function that the limit connecting super-pixel P and Q is distributed.
Namely weighting function ρ is obtained, by weighting function ρ in conjunction with spanning tree algorithm spanning tree generating model by the weight parameter w after estimation and side-play amount b.
Given weighting connected-graph model, the initial significantly figure of this method to super-pixel rank carries out middle layers of polymer process, obtains middle rank significantly figure.From a conspicuousness seed, the limit with larger weight is absorbed, forms a part spanning tree.The key step of carrying out middle layers of polymer process for initial significantly figure is as follows:
1, a conspicuousness subset { s is generated 1..., s m, make it to comprise m the super-pixel with higher saliency value.This can arrange threshold value by the initial significantly figure being calculated to gained the first stage, gets and exceedes threshold portion to realize.
2, for each conspicuousness seed s in conspicuousness subset i, repeat following steps: use s iinitialization spanning tree T 1 (i); By based on Prim algorithm [5]the polymerization of iterative tree, select that there is weight limit ρ p,Qfillet (P, Q) ∈ ε, and add spanning tree T to 1 (i).After this algorithm meets end condition, export this spanning tree T (i).
3, calculate the frequency that each super-pixel occurs in these spanning trees, in this, as their saliency value, generate and significantly scheme.
In a first step, the method needs the set generating conspicuousness seed.The method can by arranging threshold value T, and the super-pixel in initial significantly figure is divided into two groups: have one group of higher saliency value and have one group of lower saliency value.In the group with higher saliency value, super-pixel constitutes the set of conspicuousness seed.In the iterative process of spanning tree algorithm, an effective end condition is absolutely necessary.It can ensure that the spanning tree that the method obtains can either comprise whole obvious object as far as possible, can effectively avoid again the super-pixel of non-interior of articles to comprise to come in.In order to reach this purpose, the method devises one and stops function, and it comprises two ingredients: two super-pixel that (1) reference edge (P, Q) connects do not belong to the probability 1-ρ of same object p,Q.(2) reflect conspicuousness contrast, calculate poor S (the P)-S (Q) of the saliency value of gained in the low layer stage.In fact, the method setting stops function is the mean value of the two
f P,Q=(1-ρ P,Q+S(P)-S(Q))/2
Wherein, f p,Qfor stopping function.Use this termination function, spanning tree algorithm will check end condition f in each iteration p,Q> f 0, thus determine whether limit (P, Q) is brought into spanning tree.F 0it is the threshold parameter set in the starting stage of algorithm.
According to above step, the initial significantly figure of this method to super-pixel rank carries out middle layers of polymer process, obtains middle rank significantly figure.
204: middle rank is significantly schemed to introduce high-rise priori and promotes Detection results further.
This priori is expressed as a Gauss model by the embodiment of the present invention:
G ( a ) = exp [ - ( ( x a - μ x ) 2 2 σ x 2 + ( y a - μ y ) 2 2 σ y 2 + ( z a - μ z ) 2 2 σ z 2 ) ]
Wherein, (x a, y a, z a) be the coordinate of pixel a on normalized image plane X-Y and depth range Z.(μ x, μ y, μ z) be mean value from all pixel coordinates inside the set of conspicuousness seed, i.e. the center of conspicuousness object.Consider the impact of article size, the method is by variance be set to (2o x, 2o y, 2o z), o x, o yand o zthe coordinate figure that conspicuousness subset is combined in X, Y and Z tri-coordinate axis respectively.
Final remarkable figure is significantly schemed to be multiplied with this Gauss model priori to obtain by middle rank.Wherein in remarkable figure, the saliency value of each super-pixel is the prospect probability obtained by bottom, middle level, high-rise prospect modeling.
205: calculate the contour connection degree merging depth information;
First connect all neighbouring super pixels (p, q), and distribute their color and degree of depth weight, set up undirected weighted graph.Color weight d color(p, q) and degree of depth weight d depth(p, q) is calculated by the Euclidean distance between the respective average color of two neighbouring super pixels and between the average normalized degree of depth respectively.Color calculates carries out at CIE-Lab color space.Geodesic line color distance tolerance between any two super-pixel [6]be defined as:
d c l r ( p , q ) = min Σ i = 1 n - 1 d c o l o r ( p i , q i + 1 )
Wherein, d clr(p, q) is the geodesic line color distance tolerance between any two super-pixel [6]; d color(p i, q i+1) be super-pixel p iand q i+1euclidean distance between respective average color.
Geodesic line depth distance measure definitions between two super-pixel is:
d d e p ( p , q ) = min Σ i = 1 n - 1 d d e p t h ( p i , q i + 1 )
Wherein, d depth(p i, q i+1) be super-pixel p iand q i+1euclidean distance between the respective average normalized degree of depth.
Then consider color and the degree of depth, what the embodiment of the present invention used following formula to represent each super-pixel " opens into region [7]":
A ( p ) = Σ i = 1 N e - d c l r 2 ( p , p i ) 2 σ c l r 2 - d d e p 2 ( p , p i ) 2 σ d e p 2 = Σ i = 1 N S ( p , p i )
Wherein, A (p) " opens into region for each super-pixel p [7]", above formula opens into region by calculating belonging to p; S (p, p i) span be (0,1], represent super-pixel p ito the contribution opening into region of p; d clr(p, q) is the geodesic line color distance tolerance between two super-pixel [6]; d dep(p, q) is the geodesic line depth distance tolerance between two super-pixel.
Work as p iwhen belonging to same region with p, then there is d clr(p, q)=0 and d dep(p, q)=0, thus make S (p, p i)=1, can guarantee p like this ican for p open into area contribution unit area.Work as p iwhen not belonging to a region with p, at least have a limit in shortest path between which, its color weight or degree of depth weight are very large, thus make d clr(p, q) and d depat least one is had much larger than 0, to guarantee p in (p, q) imuch areas can not be increased for p opens into region.σ clrand σ depit is variance parameter.In an experiment, for obtaining stabilization result, the two is set to 10 and 0.5 by the embodiment of the present invention respectively.Final utilization following formula calculates the contour connection degree merging the degree of depth:
C o n ( p ) = Σ i = 1 N S ( p , p i ) δ ( p i ∈ B n d ) A ( p )
Wherein, Con (p) the contour connection degree that is super-pixel p; Bnd is the set of a series of super-pixel compositions in image boundary.Be 1 for the value being in borderline super-pixel δ (), the value for other super-pixel δ () is 0.
206: above-mentioned contour connection degree is converted into background probability, introduce variable
According to the computing formula of contour connection degree in 205, when super-pixel p is positioned at image boundary, contour connection degree Con (p) value calculated is far longer than 1; When super-pixel p is positioned at picture centre, Con (p) value calculated is close to zero.
The embodiment of the present invention adopts following mapping function, and above-mentioned contour connection degree is converted into background probability
w i b g = 1 - e - Con 2 ( p i ) 2 σ C o n 2
Wherein, for super-pixel p ibackground probability; Con (p i) be super-pixel p icontour connection degree; σ conbe variance parameter, in an experiment it be set to 1.
Utilize above-mentioned conversion formula, can see, when super-pixel p is positioned at image boundary, contour connection degree Con (p) value calculated is far longer than 1, now background probability value close to 1; When super-pixel p is positioned at picture centre, Con (p) value calculated close to zero, background probability value close to 0.So just effectively can distinguish the prospect in image and background, calculate the background probability of super-pixel
207: the background probability that the prospect probability generate step 204 and step 206 generate is optimized.
The embodiment of the present invention builds an objective function, can give saliency value 1 for object area, for saliency value 0 is given in background area.By optimizing this objective function, optimum remarkable figure can be obtained.The saliency value of all N number of super-pixel is made to be define following objective function:
E = Σ i = 1 N w i b g s i 2 + Σ i = 1 N w i f g ( s i - 1 ) 2 + λ c Σ i , j w i , j c ( s i - s j ) 2 + λ d Σ i , j w i , j d ( s i - s j ) 2
Wherein, it is the prospect probability obtained by the modeling of multilayer prospect; for super-pixel p ibackground probability; λ cand λ dfor Lagrange multiplier, its value is 0.5.Four parts in above formula come from four different constraint conditions.S iit is the saliency value of i-th super-pixel; s jfor the saliency value of a jth super-pixel.
For first part the probability of background is belonged in super-pixel time larger, in order to make entirety is little as far as possible, can cause s ilittle as far as possible, so this super-pixel is assigned with the saliency value close to 0.Similarly, the second part be that it distributes the saliency value close to 1 when the probability that super-pixel belongs to prospect is larger.Latter two is smoothness condition, and will the super-pixel belonging to a region together be impelled to be endowed similar saliency value, the super-pixel not belonging to same region be endowed different saliency value, thus ensures flatness and the continuity of saliency value in image.Wherein, two weights with represent the color between two neighbouring super pixels and degree of depth weight respectively.Be expressed as:
w i , j c = e - d c o l o r 2 ( p i , p j ) 2 σ c l r 2 + μ , w i , j d = e - d d e p t h 2 ( p i , p j ) 2 σ d e p 2
Wherein, d depth(p i, p j) be super-pixel p iand p jeuclidean distance between the respective average normalized degree of depth; d color(p i, p j) be super-pixel p iand p jeuclidean distance between respective average color; μ=0.1 is a little canonical constant, is used for ensureing with the stability of result when being 0, and reduce issuable noise in prospect modeling and background modeling.
208: solve the objective function of definition in step 207, the embodiment of the present invention adopts lowest mean square solution.
Optimization aim can be expressed as:
S * = arg min S { E = Σ i = 1 N w i b g s i 2 + Σ i = 1 N w i f g ( s i - 1 ) 2 + Σ i , j ( λ c w i , j c + λ c w i , j d ) ( s i - s j ) 2 }
For the ease of calculating, definition Laplacian Matrix L=D-W, D is a diagonal matrix, and diagonal element is wherein for the element in weight matrix W.Write energy function as matrix form as follows:
E(S)=w bg·S 2+w fg·(S-I) 2+2SLS T
Wherein, w bgfor the background probability of super-pixel; w fgfor the prospect probability obtained by the modeling of multilayer prospect; S is the saliency value of super-pixel; T is transposition; L is the Laplacian Matrix of constructed figure.S is differentiated, obtains following derivation:
▿ S E ( S ) = ▿ S ( w b g · S 2 + w f g · ( S - I ) 2 + 2 SLS T ) = ▿ S ( w b g · S 2 + w f g · ( S 2 - 2 S + I ) + 2 SLS T ) = 2 w b g S + 2 w f g S - 2 w f g + 4 S L
Wherein, represent and S is differentiated.
Derivative is made to be 0 can to obtain: w fg=(w bg+ w fg+ 2L) S *
Wherein S *be exactly required lowest mean square solution, be finally expressed as: S *=(w fg) -1(w bg+ w fg+ 2L)
After obtaining the saliency value of each super-pixel, the saliency value of super-pixel is distributed to each pixel that this super-pixel comprises.Each pixel in this sampled images can obtain a saliency value.By the objective function in step 207, this method can give saliency value 1 for object area, for saliency value 0 is given in background area.By optimizing this objective function, optimum remarkable figure can be obtained.
In sum, the embodiment of the present invention makes full use of the depth information of image by above-mentioned steps 201-step 207, introduces the framework optimized based on prospect background, achieves the accurate detection of obvious object.
Embodiment 3
This method mainly adopts accuracy rate (Precision), and recall rate (Recall) and curve lower area area (AUC) and F-measure tetra-indexs carry out the effect that quantitative measurement RGB-D obvious object detects.
In statistics, ROC curve refers to Receiver operating curve (ReceiverOperatingcharacteristicCurve), and it features the performance of a binary classifier system when threshold value changes.In order to better weigh the quality of ROC Curves expression of results, the embodiment of the present invention additionally uses curve lower area area (AreaUnderCurve, AUC).AUC is exactly the ratio that PR curve lower right corner area accounts for whole rectangular coordinates axial plane area in simple terms.PR (Precision-Recall) curve is the another kind of standard evaluated binary classifier performance.Detect for conspicuousness, degree of accuracy is exactly under certain threshold value, first calculates the number of pixels of obvious object intersection in the remarkable figure of acquisition and the remarkable figure of true value, then with obtain obvious object number of pixels in remarkable figure and ask ratio.Recall ratio is then the number of pixels first calculating obvious object intersection in the remarkable figure of acquisition and the remarkable figure of true value, then asks ratio with obvious object number of pixels in the remarkable figure of true value.Generally along with the increase accuracy of recall ratio also will increase.PR curve also can be converted into numerical value by appropriate formula.
F-measure is a kind of effective method, is defined as the harmonic-mean of accuracy and recall rate: F β = ( 1 + β 2 ) p r e c i s i o n * r e c a l l β 2 * p r e c i s i o n + r e c a l l , Wherein β 2=0.3.
One, quantitative result analysis
This method is tested on disclosed database.This database is by people such as Peng [3]in NLPR image library, extracted 1000 width image constructions, and people be marked there is precision target border true value for the test of algorithm.Accompanying drawing 2, Fig. 3 shows the quantitative comparison of this method in PR (precision and recall rate) curve and ROC curve.This method is labeled as: `Opti'.In PR curve map, abscissa axis represents recall rate, and axis of ordinates represents accuracy rate.When recall rate is identical, the accuracy rate value of curve is higher, and to represent Detection results better.Can see, this method is better than other several methods in PR index.In ROC curve map, abscissa axis represents the ratio of obvious object number of pixels in the number being mistaken for obvious object pixel of acquisition and the remarkable figure of true value, and axis of ordinates represents the ratio being correctly judged to obvious object number of pixels in obvious object number of pixels and the remarkable figure of true value of acquisition.Equally, when abscissa axis is fixing, the higher expression Detection results of value of axis of ordinates is better.Can find out, this method is obviously better than multi-stage process, is in same level with based on the method learnt and based on the method for Bayesian Fusion.
In addition, this method also compares curve lower area area (AreaUnderCurve, AUC) and F-measure, refers to Fig. 4.Can see, this method is obviously better than MR_D [8]method, DSR_D [9]method and multi-stage process, with based on the method learnt and the methods and results based on Bayesian Fusion very close.This is because the 2D method of expansion separately process color and depth information, thus ignore complementarity very strong between outward appearance and the corresponding clue of the degree of depth.
Two, qualitative results analysis
Fig. 5 demonstrates the qualitative comparison of some RGB-D images.Observe from figure, for different scene type and dimension of object, this method can accurately locate remarkable object, and provides significance value almost equal in target object.Background and marking area discrimination less time the method still can play a role well.Fig. 5-a represents original RGB image, and Fig. 5-b represents original depth-map, and Fig. 5-c represents MR_D [8]method, Fig. 5-d represents DSR_D [9]method, Fig. 5-e represents multi-stage process, and Fig. 5-f represents the method based on study, and Fig. 5-g represents the result that this method obtains.Finally, the obvious object Detection results of this method is obviously better than MR_D [8]method, DSR_D [9]method and multi-stage process, obvious object is lighted equably.As can be seen from the 3rd width and last width test pattern, this method creates than based on the more accurate result of the method for study, has carried out effective suppression when foreground object is accurately detected to background area simultaneously.Even if under complex scene or foreground object and background difference less time, this method can be still that foreground object distributes higher saliency value effectively, Background suppression information simultaneously.
List of references
[1]K.Desingh,K.M.Krishna,C.V.Jawahar,andD.Rajan,Depthreallymatters:Improvingvisualsalientregiondetectionwithdepth,ProceedingsofBritishMachineVisionConference,2013.
[2]A.Ciptadi,T.Hermans,andJ.M.Rehg,AninDepthViewofSaliency,ProceedingsofBritishMachineVisionConference,2013.
[3]H.Peng,B.Li,W.Xiong,W.HuandR.Ji,RGBDSalientObjectDetection:ABenchmarkandAlgorithms,ProceedingsofEuropeanConferenceonComputerVision,2014.
[4]Weisstein,EricW.LogisticEquation.MathWorld.
[5]PrimRC.Shortestconnectionnetworksandsomegeneralizations.Bellsystemtechnicaljournal,1957,36(6):1389-1401.
[6]W.Zhu,S.Liang,Y.Wei,andJ.Sun,SaliencyOptimizationfromRobustBackgroundDetection,ProceedingsofIEEEConferenceonComputerVisionandPatternRecognition,2014.
[7]RutishauserU,WaltherD,KochC,etal.,Isbottom-upattentionusefulforobjectrecognition,ProceedingsofIEEEConferenceonComputerVisionandPatternRecognition,2004,37–44.
[8]C.Yang,L.Zhang,H.Lu,X.Ruan,andM.-H.Yang,Saliencydetectionviagraph-basedmanifoldranking,ProceedingsofIEEEConferenceonComputerVisionandPatternRecognition,2013,3166-3173.
[9]X.Li,H.Lu,L.Zhang,X.Ruan,andM.-H.Yang,Saliencydetectionviadenseandsparsereconstruction,ProceedingsofIEEEInternationalConferenceonComputerVision,2013,2976-2983.
It will be appreciated by those skilled in the art that accompanying drawing is the schematic diagram of a preferred embodiment, the invention described above embodiment sequence number, just to describing, does not represent the quality of embodiment.
The foregoing is only preferred embodiment of the present invention, not in order to limit the present invention, within the spirit and principles in the present invention all, any amendment done, equivalent replacement, improvement etc., all should be included within protection scope of the present invention.

Claims (7)

1., based on the RGB-D obvious object detection method that prospect background is optimized, it is characterized in that, described obvious object detection method comprises the following steps:
Carry out initial prospect modeling based on low-level image feature contrast, obtain the initial significantly figure of super-pixel rank; Middle layers of polymer process is carried out to the initial significantly figure of described super-pixel rank, obtains middle rank significantly figure;
Middle rank is significantly schemed to introduce high-rise priori and promotes Detection results further, generate prospect probability;
Calculate the contour connection degree merging depth information, described contour connection degree is converted into background probability;
Described prospect probability and described background probability are optimized, obtain objective function;
Described objective function is solved, obtains optimum significantly figure, realize the detection to obvious object.
2. a kind of RGB-D obvious object detection method optimized based on prospect background according to claim 1, it is characterized in that, described obvious object detection method also comprises:
To a given input RGB-D image, by over-segmentation algorithm, RGB-D Iamge Segmentation is become super-pixel.
3. a kind of RGB-D obvious object detection method optimized based on prospect background according to claim 1, is characterized in that, describedly carry out initial prospect modeling based on low-level image feature contrast, and the step obtaining the initial significantly figure of super-pixel rank is specially:
Each super-pixel proper vector is represented; Define multiple contextual information;
By proper vector, contextual information and Weighted Gauss cuclear density to each super-pixel calculating probability density, corresponding probability density value is assigned to corresponding super-pixel, gets the initial significantly figure of super-pixel rank.
4. a kind of RGB-D obvious object detection method optimized based on prospect background according to claim 3, it is characterized in that, described contextual information comprises:
1) nearest with current super-pixel n lindividual adjacent super-pixel;
2) comprise except current super-pixel, the every other super-pixel in image;
3) super-pixel on four angles of image is comprised.
5. a kind of RGB-D obvious object detection method optimized based on prospect background according to claim 1, is characterized in that, the described initial significantly figure to described super-pixel rank carries out middle layers of polymer process, obtains intermediate significantly figure and is specially:
Weighting function is obtained by the weight parameter after estimation and side-play amount;
Carry out spanning tree process by weighting function in conjunction with the initial significantly figure of spanning tree algorithm to described super-pixel rank, obtain middle rank significantly figure.
6. a kind of RGB-D obvious object detection method optimized based on prospect background according to claim 5, is characterized in that, described by carrying out spanning tree process to the initial significantly figure of described super-pixel rank, the step obtaining intermediate significantly figure is specially:
Generate a conspicuousness subset, make it to comprise m the super-pixel with higher saliency value;
The set of iterative tree is carried out to each conspicuousness seed in conspicuousness subset, selects the fillet with weight limit, and add in spanning tree, export spanning tree;
Calculate the frequency that each super-pixel occurs in spanning tree, in this, as saliency value, generate middle rank significantly figure.
7. a kind of RGB-D obvious object detection method optimized based on prospect background according to claim 1, is characterized in that, describedly significantly schemes to introduce high-rise priori to middle rank and promotes Detection results further, and the step generating prospect probability is specially:
By significantly to scheme described middle rank and Gauss model carries out priori and is multiplied, obtain senior remarkable figure, in described senior remarkable figure, the saliency value of each super-pixel is the prospect probability obtained by bottom, middle level, high-rise prospect modeling.
CN201510897635.0A 2015-12-07 2015-12-07 RGB-D salient object detection method based on foreground and background optimization Pending CN105513070A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510897635.0A CN105513070A (en) 2015-12-07 2015-12-07 RGB-D salient object detection method based on foreground and background optimization

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510897635.0A CN105513070A (en) 2015-12-07 2015-12-07 RGB-D salient object detection method based on foreground and background optimization

Publications (1)

Publication Number Publication Date
CN105513070A true CN105513070A (en) 2016-04-20

Family

ID=55721027

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510897635.0A Pending CN105513070A (en) 2015-12-07 2015-12-07 RGB-D salient object detection method based on foreground and background optimization

Country Status (1)

Country Link
CN (1) CN105513070A (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106327507A (en) * 2016-08-10 2017-01-11 南京航空航天大学 Color image significance detection method based on background and foreground information
CN106780582A (en) * 2016-12-16 2017-05-31 西安电子科技大学 Based on the image significance detection method that textural characteristics and color characteristic are merged
CN106815843A (en) * 2016-11-30 2017-06-09 江苏城乡建设职业学院 A kind of fruit object acquisition methods based on convex closure center priori and absorbing Marcov chain
CN106991669A (en) * 2017-03-14 2017-07-28 北京工业大学 A kind of conspicuousness detection method based on depth-selectiveness difference
CN107085848A (en) * 2017-04-20 2017-08-22 安徽大学 A kind of detection method of RGB D figure conspicuousnesses
CN107274419A (en) * 2017-07-10 2017-10-20 北京工业大学 A kind of deep learning conspicuousness detection method based on global priori and local context
CN107424161A (en) * 2017-04-25 2017-12-01 南京邮电大学 A kind of indoor scene image layout method of estimation by thick extremely essence
CN107766857A (en) * 2017-10-17 2018-03-06 天津大学 The vision significance detection algorithm propagated based on graph model structure with label
CN107895162A (en) * 2017-10-17 2018-04-10 天津大学 Saliency algorithm of target detection based on object priori
CN108122239A (en) * 2016-11-29 2018-06-05 Sap欧洲公司 Use the object detection in the image data of depth segmentation
CN108154488A (en) * 2017-12-27 2018-06-12 西北工业大学 A kind of image motion ambiguity removal method based on specific image block analysis
CN108154107A (en) * 2017-12-22 2018-06-12 北京航空航天大学 A kind of method of the scene type of determining remote sensing images ownership
CN108416347A (en) * 2018-01-04 2018-08-17 天津大学 Well-marked target detection algorithm based on boundary priori and iteration optimization
CN109409376A (en) * 2018-11-05 2019-03-01 昆山紫东智能科技有限公司 For the image partition method, terminal and storage medium of solid waste object
CN110096961A (en) * 2019-04-04 2019-08-06 北京工业大学 A kind of indoor scene semanteme marking method of super-pixel rank
CN111461150A (en) * 2019-01-18 2020-07-28 上海大学 Unsupervised learning object appearance algorithm

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
JING SUN 等: "SALIENCY DETECTION BASED ON INTEGRATION OF BOUNDARY AND SOFT-SEGMENTATION", 《ICIP 2012》 *
崔波: "基于RGB-D信息的显著物体检测", 《万方数据企业知识服务平台》 *

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106327507B (en) * 2016-08-10 2019-02-22 南京航空航天大学 A kind of color image conspicuousness detection method based on background and foreground information
CN106327507A (en) * 2016-08-10 2017-01-11 南京航空航天大学 Color image significance detection method based on background and foreground information
CN108122239A (en) * 2016-11-29 2018-06-05 Sap欧洲公司 Use the object detection in the image data of depth segmentation
CN108122239B (en) * 2016-11-29 2020-12-01 Sap欧洲公司 Object detection in image data using depth segmentation
CN106815843A (en) * 2016-11-30 2017-06-09 江苏城乡建设职业学院 A kind of fruit object acquisition methods based on convex closure center priori and absorbing Marcov chain
CN106780582A (en) * 2016-12-16 2017-05-31 西安电子科技大学 Based on the image significance detection method that textural characteristics and color characteristic are merged
CN106780582B (en) * 2016-12-16 2019-08-13 西安电子科技大学 The image significance detection method merged based on textural characteristics and color characteristic
CN106991669A (en) * 2017-03-14 2017-07-28 北京工业大学 A kind of conspicuousness detection method based on depth-selectiveness difference
CN106991669B (en) * 2017-03-14 2019-09-27 北京工业大学 A kind of conspicuousness detection method based on depth-selectiveness difference
CN107085848A (en) * 2017-04-20 2017-08-22 安徽大学 A kind of detection method of RGB D figure conspicuousnesses
CN107424161A (en) * 2017-04-25 2017-12-01 南京邮电大学 A kind of indoor scene image layout method of estimation by thick extremely essence
CN107424161B (en) * 2017-04-25 2021-03-02 南京邮电大学 Coarse-to-fine indoor scene image layout estimation method
CN107274419B (en) * 2017-07-10 2020-10-13 北京工业大学 Deep learning significance detection method based on global prior and local context
CN107274419A (en) * 2017-07-10 2017-10-20 北京工业大学 A kind of deep learning conspicuousness detection method based on global priori and local context
CN107766857B (en) * 2017-10-17 2021-08-03 天津大学 Visual saliency detection algorithm based on graph model construction and label propagation
CN107895162B (en) * 2017-10-17 2021-08-03 天津大学 Image saliency target detection algorithm based on object prior
CN107895162A (en) * 2017-10-17 2018-04-10 天津大学 Saliency algorithm of target detection based on object priori
CN107766857A (en) * 2017-10-17 2018-03-06 天津大学 The vision significance detection algorithm propagated based on graph model structure with label
CN108154107B (en) * 2017-12-22 2021-09-14 北京航空航天大学 Method for determining scene category to which remote sensing image belongs
CN108154107A (en) * 2017-12-22 2018-06-12 北京航空航天大学 A kind of method of the scene type of determining remote sensing images ownership
CN108154488B (en) * 2017-12-27 2019-10-22 西北工业大学 A kind of image motion ambiguity removal method based on specific image block analysis
CN108154488A (en) * 2017-12-27 2018-06-12 西北工业大学 A kind of image motion ambiguity removal method based on specific image block analysis
CN108416347A (en) * 2018-01-04 2018-08-17 天津大学 Well-marked target detection algorithm based on boundary priori and iteration optimization
CN109409376A (en) * 2018-11-05 2019-03-01 昆山紫东智能科技有限公司 For the image partition method, terminal and storage medium of solid waste object
CN109409376B (en) * 2018-11-05 2020-10-30 昆山紫东智能科技有限公司 Image segmentation method for solid waste object, computer terminal and storage medium
CN111461150A (en) * 2019-01-18 2020-07-28 上海大学 Unsupervised learning object appearance algorithm
CN111461150B (en) * 2019-01-18 2023-07-21 上海大学 Algorithm for unsupervised learning of object appearance
CN110096961B (en) * 2019-04-04 2021-03-02 北京工业大学 Indoor scene semantic annotation method at super-pixel level
CN110096961A (en) * 2019-04-04 2019-08-06 北京工业大学 A kind of indoor scene semanteme marking method of super-pixel rank

Similar Documents

Publication Publication Date Title
CN105513070A (en) RGB-D salient object detection method based on foreground and background optimization
Arteta et al. Interactive object counting
CN107330897B (en) Image segmentation method and system
CN105869178B (en) A kind of complex target dynamic scene non-formaldehyde finishing method based on the convex optimization of Multiscale combination feature
CN108537239B (en) Method for detecting image saliency target
CN105825502B (en) A kind of Weakly supervised method for analyzing image of the dictionary study based on conspicuousness guidance
CN103035013B (en) A kind of precise motion shadow detection method based on multi-feature fusion
Kim et al. Color–texture segmentation using unsupervised graph cuts
CN107909015A (en) Hyperspectral image classification method based on convolutional neural networks and empty spectrum information fusion
CN106228544A (en) A kind of significance detection method propagated based on rarefaction representation and label
CN109766936A (en) Image change detection method based on information transmitting and attention mechanism
CN108416347A (en) Well-marked target detection algorithm based on boundary priori and iteration optimization
CN103489005A (en) High-resolution remote sensing image classifying method based on fusion of multiple classifiers
CN107301430A (en) Broad sense Multivariable Fuzzy c means clustering algorithms
Li et al. A robust shape model for multi-view car alignment
CN104657980A (en) Improved multi-channel image partitioning algorithm based on Meanshift
CN107862702A (en) A kind of conspicuousness detection method of combination boundary connected and local contrast
CN104680151B (en) A kind of panchromatic remote sensing image variation detection method of high-resolution for taking snow covering influence into account
CN105701493A (en) Methods and systems for image matting and foreground estimation based on hierarchical graphs
CN105046673A (en) Self-learning based hyperspectral image and visible image fusion classification method
CN108388901B (en) Collaborative significant target detection method based on space-semantic channel
CN104599288A (en) Skin color template based feature tracking method and device
CN109816714A (en) A kind of point cloud object type recognition methods based on Three dimensional convolution neural network
CN115861738A (en) Category semantic information guided remote sensing target detection active sampling method
CN105787505A (en) Infrared image clustering segmentation method combining sparse coding and spatial constraints

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20160420

WD01 Invention patent application deemed withdrawn after publication