CN106570498B - Salient region detecting method and detection system - Google Patents

Salient region detecting method and detection system Download PDF

Info

Publication number
CN106570498B
CN106570498B CN201610889100.3A CN201610889100A CN106570498B CN 106570498 B CN106570498 B CN 106570498B CN 201610889100 A CN201610889100 A CN 201610889100A CN 106570498 B CN106570498 B CN 106570498B
Authority
CN
China
Prior art keywords
super
pixel
input picture
image
notable
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610889100.3A
Other languages
Chinese (zh)
Other versions
CN106570498A (en
Inventor
王鹏
罗永康
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Automation of Chinese Academy of Science
Original Assignee
Institute of Automation of Chinese Academy of Science
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Automation of Chinese Academy of Science filed Critical Institute of Automation of Chinese Academy of Science
Priority to CN201610889100.3A priority Critical patent/CN106570498B/en
Publication of CN106570498A publication Critical patent/CN106570498A/en
Application granted granted Critical
Publication of CN106570498B publication Critical patent/CN106570498B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/255Detecting or recognising potential candidate objects based on visual cues, e.g. shapes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Abstract

The present invention relates to a kind of salient region detecting method and detection systems.Wherein, this method includes carrying out super-pixel segmentation to input picture, obtains image superpixel processing unit;Then, it is based on image superpixel processing unit, measures the topological background confidence level of input picture;Then, the topological background confidence level based on input picture calculates the contrast in color of image and spatial position feature, obtains image primary saliency value;It recycles compactedness method of diffusion to correct image primary saliency value, obtains compactedness notable figure;And topological background confidence level and compactedness notable figure using input picture, saliency value assignment is carried out, image single scale notable figure is obtained;Finally, the method using Multiscale Fusion handles image single scale notable figure, scale notable figure is obtained, to obtain image obvious object region.By using above-mentioned technical proposal, solve how fast and effeciently abstract image marking area the technical issues of.

Description

Salient region detecting method and detection system
Technical field
The present embodiments relate to technical field of image processing, more particularly, to a kind of salient region detecting method and detection System.
Background technique
Human visual system has attention function, can rapidly lock key area, and only selection enters the visual field A portion information is for subsequent deeper information extraction and processing, to obtain efficient visual perception.Inspire in The conspicuousness detection model of human visual attention mechanism can quickly and efficiently extract useful information for computer system realization Solution route is provided.The cognitive theory and physiological models that view-based access control model pays attention to, the researcher of computer field propose very much Conspicuousness detection model.Vision significance detects the pretreatment process as visual task, has boundless application prospect, It can be applied to many different high-level complicated visual processes tasks, such as object detection, visual target tracking, object are known Not, compression of images and robot cognition etc..
In recent years, the obvious object detection method much based on frame background priori is constantly proposed that they are obtained Good detection effect.Such as:
Xiaohui Li,Huchuan Lu,Lihe Zhang,Xiang Ruan,and Ming-Hsuan Yang.Saliency detection via dense and sparse reconstruction.In IEEE International Conference on Computer Vision(ICCV),pages2976–2983,2013;
Wangjiang Zhu,Shuang Liang,YichenWei,and Jian Sun.Saliency optimization from robust background detection.In IEEE Conference on Computer Vision and Pattern Recognition(CVPR),pages 2814–2821,2014;
Chuan Yang,Lihe Zhang,Huchuan Lu,Xiang Ruan,and Ming-Hsuan Yang.Saliency detection via graph-based manifold ranking.In IEEE Conference on Computer Vision and Pattern Recognition(CVPR),pages 3166–3173,2013。
However, these methods are during inferring saliency value using the topological relation of image-region and frame region, it is past It is past to be more heavily weighted toward global conspicuousness.By it is existing based on the conspicuousness detection method of frame background priori for, will lead to may deposit In following problem: when obvious object region is similar to framing mask region, the local conspicuousness of obvious object is mistakenly pressed down System;When image background regions and framing mask region are extremely dissimilar, the conspicuousness of background area may be exaggerated by mistake, thus Generate the conspicuousness testing result of mistake.
In view of this, the present invention is specifically proposed.
Summary of the invention
How fast and effeciently the embodiment of the present invention provides a kind of salient region detecting method, significant to solve abstract image The technical issues of region.In addition, the embodiment of the present invention also provides a kind of marking area detection system.
To achieve the goals above, according to an aspect of the present invention, the following technical schemes are provided:
A kind of salient region detecting method, the method may include:
Super-pixel segmentation is carried out to input picture, obtains image superpixel processing unit;
Based on described image super-pixel processing unit, the topological background confidence level of the input picture is measured;
The topological background confidence level based on the input picture, calculates pair in color of image and spatial position feature Than degree, image primary saliency value is obtained;
Described image primary saliency value is corrected using compactedness method of diffusion, obtains compactedness notable figure;
Using the topological background confidence level of the input picture and the compactedness notable figure, saliency value tax is carried out Value, obtains image single scale notable figure;
Described image single scale notable figure is handled using the method for Multiscale Fusion, obtains scale notable figure, from And obtain image obvious object region.
Further, described to be based on described image super-pixel processing unit, the topological background for measuring the input picture is set Reliability can specifically include:
Based on described image super-pixel processing unit, the super-pixel for calculating the input picture extends area;
Super-pixel is calculated in the input picture along the length of the input picture frame;
Extend area and the length based on the super-pixel, calculates the frame Connected degree of the input picture super-pixel;
Based on the frame Connected degree, the topological background confidence level of the input picture is calculated.
Further, the topological background confidence level based on the input picture, calculates color of image and space Contrast on position feature obtains image primary saliency value, can specifically include:
The topological background confidence level based on the input picture, and described image color characteristic pair is calculated according to following formula Than degree:
Wherein, the p and piIndicate super-pixel, the i=1 ..., N;The N indicates institute in the input picture The super-pixel number for including;The biIndicate the topological background confidence level of the input picture;The dcolor(p,pi) table Show the super-pixel p and piEuclidean distance on CIE L*a*b* color space;The contrastcolor(p) it indicates Described image color characteristic contrast;
Wherein, the wpos(p,pi) determined by following formula:
Wherein, the dpos(p,pi) indicate the p and piThe Euclidean distance of center;The σposIndicate comparison Spend opereating specification adjustment factor;
The topological background confidence level based on the input picture, and the spatial position feature pair is calculated according to following formula Than degree:
Wherein, wcolor(p,pi) indicate the super-pixel p and piSimilitude on color characteristic;The contrastpos (p) the spatial position Characteristic Contrast degree is indicated;
Wherein,
Wherein, the σcolorIndicate the spatial position Characteristic Contrast degree to the adjusting system of the color characteristic sensitivity Number;The dcolor(p,pi) indicate the super-pixel p and piIn CIE L*a*b*Euclidean distance on color space;
Based on described image color characteristic contrast and the spatial position Characteristic Contrast degree, and according to described in following formula calculating Image primary saliency value:
CS (p)=contrastcolor(p)·exp(-k·contrastpos(p))
Wherein, the k indicates described image color characteristic and the spatial position feature to the super-pixel saliency value The weight coefficient of contribution.
Further, described to correct described image primary saliency value using compactedness method of diffusion, it is significant to obtain compactedness Figure, can specifically include:
Construct k-regular closed loop figure, wherein respectively connect side by incidence matrix to measure in the closed loop figure;
Base is determined using the manifold smoothness constraint in graph theory based on described image primary saliency value and the incidence matrix In the manifold ranking function of non-normalized Laplacian Matrix, to obtain the compactedness notable figure.
Further, the topological background confidence level and the compactedness notable figure using the input picture, Saliency value assignment is carried out, image single scale notable figure is obtained, can specifically include:
Existed using the topological background confidence level of the input picture with the compactedness notable figure and the super-pixel Euclidean distance on CIE L*a*b* color space carries out saliency value assignment, obtains the saliency value of the input picture super-pixel;
Assign the saliency value of the input picture super-pixel to all pictures in the input picture in the super-pixel Element, as the saliency value of those pixels, to obtain described image single scale notable figure.
Further, the incidence matrix is determined by following formula:
Wherein, the wijIndicate bonding strength;The n takes positive integer;The dcolor(pi,pj) indicate super-pixel piAnd pj In CIE L*a*b*Euclidean distance on color space;The σ indicates adjustment factor;The i and the j take 1 ... N, the N Indicate super-pixel number.
Further, it is described using the topological background confidence level of the input picture and the compactedness notable figure and The super-pixel is in CIE L*a*b*Euclidean distance on color space carries out saliency value assignment, obtains the super picture of the input picture The saliency value of element, can specifically include:
The saliency value of the input picture super-pixel is obtained using following objective function:
Wherein, the siIndicate super-pixel piSaliency value;The sjIndicate super-pixel pjSaliency value;The i, j= 1,...,N;The N indicates the number of super-pixel;The biIndicate the topological background confidence level of the input picture;It is described fiIndicate compactedness notable figure;The λ indicates regulatory factor;The dcolor(pi,pj) indicate the super-pixel piAnd pjIn CIE L*a*b*Euclidean distance on color space;The σcolorIndicate adjustment factor;The μsmIndicate the adjustment factor of inhibition noise.
To achieve the goals above, according to another aspect of the present invention, a kind of marking area detection system, institute are also provided The system of stating includes:
Divide module, for carrying out super-pixel segmentation to input picture, obtains image superpixel processing unit;
Metric module is connected with the segmentation module, for being based on described image super-pixel processing unit, measures described defeated Enter the topological background confidence level of image;
Computing module is connected with the metric module, for the topological background confidence level based on the input picture, The contrast in color of image and spatial position feature is calculated, image primary saliency value is obtained;
Correction module is connected with the computing module, for aobvious using compactedness method of diffusion amendment described image primary Work value obtains compactedness notable figure;
Assignment module is connected with the metric module and the correction module respectively, for utilizing the input picture The topology background confidence level and the compactedness notable figure carry out saliency value assignment, obtain image single scale notable figure;
Fusion Module is connected with the assignment module, for the method using Multiscale Fusion to described image single scale Notable figure is handled, and scale notable figure is obtained, to obtain image obvious object region.
The embodiment of the present invention provides a kind of salient region detecting method and detection system.Wherein, this method includes to input Image carries out super-pixel segmentation, obtains image superpixel processing unit;Then, it is based on image superpixel processing unit, is measured defeated Enter the topological background confidence level of image;Then, the topological background confidence level based on input picture calculates color of image and space bit The contrast in feature is set, image primary saliency value is obtained;It recycles compactedness method of diffusion to correct image primary saliency value, obtains To compactedness notable figure;And topological background confidence level and compactedness notable figure using input picture, carry out saliency value tax Value, obtains image single scale notable figure;Finally, the method using Multiscale Fusion handles image single scale notable figure, Scale notable figure is obtained, to obtain image obvious object region.By using above-mentioned technical proposal, take into account in the detection process Image overall conspicuousness and local conspicuousness, realize the beneficial effect of fast and effeciently abstract image marking area.
Detailed description of the invention
Fig. 1 is the flow diagram according to the salient region detecting method of the embodiment of the present invention;
Fig. 2 is the structural schematic diagram according to the marking area detection system of the embodiment of the present invention.
Specific embodiment
The preferred embodiment of the present invention described with reference to the accompanying drawings.It will be apparent to a skilled person that this A little embodiments are used only for explaining technical principle of the invention, it is not intended that limit the scope of the invention.The present embodiment Implemented under the premise of the technical scheme of the present invention, in conjunction with detailed embodiment and specific operating process, but this hair Bright protection scope is not limited to following embodiments.
The prior art is during inferring saliency value using the topological relation of image-region and frame region, often more partially Global conspicuousness is overweighted, because such methods utilize frame region approximate image global context, and frame region might not be located In the local neighbor of image-region, so as to lead to the conspicuousness for mistakenly inhibiting obvious object region or mistakenly overstate The conspicuousness in overall background region.
Basic thought of the embodiment of the present invention is: carrying out super-pixel segmentation to input picture, obtains image processing unit;It connects , image background confidence level is calculated according to frame Connected degree, and be based on background confidence calculations image saliency value;Then, to institute It obtains notable figure and carries out compactedness diffusion, correct saliency, obtain more acurrate more compact notable figure;Recycle background confidence Degree carries out saliency value with compactedness notable figure and spatial neighbors information and optimizes assignment;Finally, the method using Multiscale Fusion obtains Obtain multiple dimensioned notable figure.
For this purpose, the embodiment of the present invention provides a kind of salient region detecting method.As shown in Figure 1, this method can pass through step Rapid S100 is realized to step S150.Wherein:
S100: super-pixel segmentation is carried out to input picture, obtains image superpixel processing unit.
Specifically, this step can use superpixel segmentation method and carry out super-pixel segmentation to input picture, to obtain figure As super-pixel processing unit.
S110: it is based on image superpixel processing unit, measures the topological background confidence level of input picture.
Specifically, this step may include:
S111: being based on image superpixel processing unit, and the super-pixel of calculating input image extends area.
As an example, this step can according to the following formula calculating input image super-pixel extend area:
Wherein, p and piIndicate super-pixel, i=1 ..., N;N indicates super-pixel number included in input picture;σgeo Indicate adjustment factor;dgeo(p,pi) indicate super-pixel p and piIn CIE L*a*b*Geodesic distance on color space;SpanArea (p) indicate that the super-pixel of input picture extends area.
For sake of clarity, herein by dgeo(p,pi) it is expressed as dgeo(p, q) can be realized by following formula:
Wherein, dcolor(pi, pi+1) indicate super-pixel piAnd pi+1In CIE L*a*b*Euclidean distance on color space;N takes Positive integer;dgeo(p, q) indicates super-pixel p and q in CIE L*a*b*Geodesic distance on color space.
S112: in calculating input image super-pixel along input picture frame length.
As an example, this step can carry out in calculating input image super-pixel along the length of input picture frame according to the following formula Degree:
Wherein, piIndicate super-pixel, i=1 ..., N;N indicates super-pixel number included in input picture;δ(·) Indicate indicator function, when super-pixel is on input picture frame, otherwise value 1 is 0;B is indicated on input picture frame Super-pixel set;σgeoIndicate adjustment factor;dgeo(p,pi) indicate super-pixel p and piIn CIE L*a*b*Geodetic on color space Distance;LB (p) indicates that super-pixel is along the length of input picture frame in input picture.
S113: super-pixel in area and input picture is extended based on super-pixel and is calculated along the length of input picture frame The frame Connected degree of input picture super-pixel.
As an example, the embodiment of the present invention can be by following formula come the frame Connected degree of calculating input image super-pixel:
Wherein, LB (p) indicates that super-pixel is along the length of input picture frame in input picture;SpanArea (p) is indicated The super-pixel of input picture extends area;BC (p) indicates input picture frame Connected degree.
S114: frame Connected degree, the topological background confidence level of calculating input image are based on.
As an example, the embodiment of the present invention can be by following formula come the topological background confidence level of calculating input image:
Wherein, BC (pi) indicate input picture frame Connected degree;σbcIndicate the frame for adjusting input picture super-pixel Weighing factor of the Connected degree to the topological background confidence level of input picture;biIndicate the topological background confidence level of input picture.
S120: the topological background confidence level based on input picture calculates the comparison in color of image and spatial position feature Degree obtains image primary saliency value.
Specifically, this step may include:
S121: the topological background confidence level based on input picture, and color of image Characteristic Contrast degree is calculated according to following formula:
Wherein, p and piIndicate super-pixel, i=1 ..., N;N indicates super-pixel number included in input picture;biTable Show the topological background confidence level of input picture;dcolor(p,pi) indicate super-pixel p and piIn CIE L*a*b*Europe on color space Family name's distance;contrastcolor(p) color of image Characteristic Contrast degree is indicated.
Wherein, wpos(p,pi) determined by following formula:
Wherein, dpos(p,pi) indicate p and piThe Euclidean distance of center;σposIndicate that contrast operation range adjusts system Number.
S122: the topological background confidence level based on input picture, and spatial position Characteristic Contrast degree is calculated according to following formula:
Wherein, dpos(p,pi) indicate p and piThe Euclidean distance of center;biIndicate the topological background confidence of input picture Degree;wcolor(p,pi) indicate super-pixel p and piSimilitude on color characteristic;contrastpos(p) representation space position is special Levy contrast.
Wherein,
Wherein, σcolorAdjustment factor of the representation space position feature contrast to color characteristic sensitivity;dcolor(p, pi) indicate super-pixel p and piIn CIE L*a*b*Euclidean distance on color space.
S123: it is based on color of image Characteristic Contrast degree and spatial position Characteristic Contrast degree, and according at the beginning of following formula calculating image Grade saliency value:
CS (p)=contrastcolor(p)·exp(-k·contrastpos(p));
Wherein, k indicates color of image feature and spatial position feature to the weight coefficient of the contribution of super-pixel saliency value; contrastpos(p) representation space position feature contrast;contrastcolor(p) color of image Characteristic Contrast degree is indicated.
S130: image primary saliency value is corrected using compactedness method of diffusion, obtains compactedness notable figure.
Specifically, in some alternative embodiments, this step may further include:
S131: construction k-regular closed loop figure, wherein respectively connect side by incidence matrix to measure in the closed loop figure.
In some alternative embodiments, incidence matrix is determined by following formula:
Wherein, wijIndicate bonding strength;N takes positive integer;dcolor(pi, pj) indicate super-pixel piAnd pjIn CIE L*a*b*Face Euclidean distance in the colour space;σ indicates adjustment factor;I and j take 1 ..., and N, N indicate super-pixel number.
The weight that σ is used to control connection side can take 1/ σ in the specific implementation process2=10.
S132: being based on image primary saliency value and incidence matrix, using the manifold smoothness constraint in graph theory, determines based on non- The manifold ranking function for normalizing Laplacian Matrix, to obtain compactedness notable figure.
In practical applications, the diffusion process of the manifold ranking function based on non-normalized Laplacian Matrix can basis Following formula calculates:
f*=(D- α W)-1y;
Wherein, D indicates connection matrix, D=diag { d11,...,dnn, diijwij, wijIndicate bonding strength;Y= [y1,...,yn], yi=CS (pi);CS(pi) indicate image primary saliency value, piIndicate super-pixel;α indicates adjustment factor;i,j, N takes positive integer;f*Indicate compactedness notable figure,f*It can be resulting by manifold ranking diffusion process Numerical value can be used to indicate that the saliency value of each super-pixel in the saliency value of each super-pixel in image namely image is fi= fi *
Preferably, above-mentioned α takes 0.99.
By step S130, enhance model using compactedness diffusion process to the descriptive power of image local conspicuousness, So as to obtain more acurrate more compact notable figure.
S140: using the topological background confidence level and compactedness notable figure of input picture, saliency value assignment is carried out, figure is obtained As single scale notable figure.
Specifically, this step may include:
S141: using the topological background confidence level and compactedness notable figure and super-pixel of input picture in CIE L*a*b*Face Euclidean distance in the colour space carries out saliency value assignment, obtains the saliency value of input picture super-pixel.
In some alternative embodiments, this step can use following objective function and obtain input picture super-pixel Saliency value:
Wherein, siIndicate super-pixel piSaliency value;sjIndicate super-pixel pjSaliency value;I, j=1 ..., N;N is indicated The number of super-pixel;biIndicate the topological background confidence level of input picture;fiIndicate compactedness notable figure;λ indicates regulatory factor, Preferably, λ takes 1-5.
Wherein,
Wherein, dcolor(pi, pj) indicate super-pixel piAnd pjIn CIE L*a*b*Euclidean distance on color space;σcolorTable Show adjustment factor;μsmIndicate to inhibit the adjustment factor of noise to inhibit in scene for adjusting the optimization problem under complex scene Noise.Preferably, μsm=0.1.
As an example, being obtained with input picture super-pixel by carrying out the operations such as derivation to above-mentioned objective function Saliency value si
S142: assigning the saliency value of input picture super-pixel to all pixels in input picture in the super-pixel, as The saliency value of those pixels, to obtain image single scale notable figure.
S150: handling image single scale notable figure using the method for Multiscale Fusion, obtain scale notable figure, from And obtain image obvious object region.
Specifically, in some alternative embodiments, multiple dimensioned notable figure can be obtained according to the following formula:
Wherein, r=1,2,3,4 correspond to the super-pixel segmentation generated according to the setting of different super-pixel sizes;SrIt indicates The single scale notable figure generated under different scale super-pixel segmentation result;S indicates multiple dimensioned notable figure.
Above-described embodiment has taken into account image overall conspicuousness and local conspicuousness in the detection process, realizes quickly and effectively The beneficial effect of ground abstract image marking area.
Although each step is described in the way of above-mentioned precedence in above-described embodiment, this field Technical staff is appreciated that the effect in order to realize the present embodiment, executes between different steps not necessarily in such order, It (parallel) execution simultaneously or can be executed with reverse order, these simple variations all protection scope of the present invention it It is interior.
Based on technical concept identical with embodiment of the method, the embodiment of the present invention also provides a marking area detection system. As shown in Fig. 2, the system 20 may include: segmentation module 21, metric module 22, computing module 23, correction module 24, assignment mould Block 25 and Fusion Module 26.Wherein, segmentation module 21 is used to carry out super-pixel segmentation to input picture, obtains at image superpixel Manage unit.Metric module 22 is connected with segmentation module 21, for being based on image superpixel processing unit, measures opening up for input picture Flutter background confidence level.Computing module 23 is connected with metric module 22, for the topological background confidence level based on input picture, calculates Contrast on color of image and spatial position feature obtains image primary saliency value.Correction module 24 and 23 phase of computing module Even, for correcting image primary saliency value using compactedness method of diffusion, compactedness notable figure is obtained.Assignment module 25 respectively with Metric module 22 is connected with correction module 24, for utilizing the topological background confidence level and compactedness notable figure of input picture, into Row saliency value assignment obtains image single scale notable figure.Fusion Module 26 is connected with assignment module 25, for being melted using multiple dimensioned The method of conjunction handles image single scale notable figure, obtains scale notable figure, to obtain image obvious object region.
It will be understood by those skilled in the art that above-mentioned marking area detection system can also include some other known knot Structure, such as processor, controller, memory and bus etc., wherein memory include but is not limited to random access memory, flash memory, only Read memory, programmable read only memory, volatile memory, nonvolatile memory, serial storage, parallel storage or Register etc., processor include but is not limited to single core processor, multi-core processor, the processor based on X86-based, CPLD/ FPGA, DSP, arm processor, MIPS processor etc., bus may include data/address bus, address bus and control bus.In order to Embodiment of the disclosure is unnecessarily obscured, these well known structures are not shown in FIG. 2.
It should be noted that marking area detection system provided by the above embodiment and detection method are carrying out marking area When detection, only carried out with the division of above-mentioned each functional module or step for example, in practical applications, can according to need and Above-mentioned function distribution is completed by different functional modules, i.e., by the embodiment of the present invention module or step decompose again or Person's combination, to complete all or part of the functions described above.For module involved in the embodiment of the present invention or step Title, it is only for distinguish modules or step, be not intended as the improper restriction to the scope of the present invention.
Also it should be noted is that, in order to avoid having obscured the present invention because of unnecessary details, in the accompanying drawings only It shows the processing step closely related with technical solution of the present invention, therefore and is omitted and little other of relationship of the present invention Details.
So far, preferred embodiment shown in the drawings is had been combined to the purpose of the present invention, technical scheme and beneficial effects It has been further described, still, it will be readily appreciated by those skilled in the art that the obvious not office of protection scope of the present invention It is limited to these specific embodiments.Under the premise of without departing from the principle of the present invention, all within the spirits and principles of the present invention, Those skilled in the art can make equivalent change or replacement to the relevant technologies feature, the technology after these changes or replacement Scheme will fall within the scope of protection of the present invention.

Claims (8)

1. a kind of salient region detecting method, which is characterized in that the described method includes:
Super-pixel segmentation is carried out to input picture, obtains image superpixel processing unit;
Based on described image super-pixel processing unit, the topological background confidence level of the input picture is measured;
The topological background confidence level based on the input picture, calculates the comparison in color of image and spatial position feature Degree obtains image primary saliency value;
Described image primary saliency value is corrected using compactedness method of diffusion, obtains compactedness notable figure;
Using the topological background confidence level of the input picture and the compactedness notable figure, saliency value assignment is carried out, is obtained To image single scale notable figure;
Described image single scale notable figure is handled using the method for Multiscale Fusion, scale notable figure is obtained, to obtain Obtain image obvious object region.
2. the method according to claim 1, wherein described be based on described image super-pixel processing unit, measurement The topological background confidence level of the input picture, specifically includes:
Based on described image super-pixel processing unit, the super-pixel for calculating the input picture extends area;
Super-pixel is calculated in the input picture along the length of the input picture frame;
Extend area and the length based on the super-pixel, calculates the frame Connected degree of the input picture super-pixel;
Based on the frame Connected degree, the topological background confidence level of the input picture is calculated.
3. the method according to claim 1, wherein the topological background based on the input picture is set Reliability calculates the contrast in color of image and spatial position feature, obtains image primary saliency value, specifically includes:
The topological background confidence level based on the input picture, and the comparison of described image color characteristic is calculated according to following formula Degree:
Wherein, p and piIndicate super-pixel, i=1 ..., N;N indicates super-pixel number included in the input picture;biTable Show the topological background confidence level of the input picture;dcolor(p,pi) indicate the super-pixel p and piIn CIE L*a* b*Euclidean distance on color space;contrastcolor(p) described image color characteristic contrast is indicated;
Wherein, wpos(p,pi) determined by following formula:
Wherein, dpos(p,pi) indicate the p and piThe Euclidean distance of center;σposIndicate contrast operation range tune Save coefficient;
The topological background confidence level based on the input picture, and the spatial position Characteristic Contrast is calculated according to following formula Degree:
Wherein, wcolor(p,pi) indicate the super-pixel p and piSimilitude on color characteristic;contrastpos(p) institute is indicated State spatial position Characteristic Contrast degree;
Wherein,
Wherein, σcolorIndicate the spatial position Characteristic Contrast degree to the adjustment factor of the color characteristic sensitivity;
Described image is calculated based on described image color characteristic contrast and the spatial position Characteristic Contrast degree, and according to following formula Primary saliency value:
CS (p)=contrastcolor(p)·exp(-k·contrastpos(p))
Wherein, k indicates described image color characteristic and the spatial position feature to the power of the contribution of the super-pixel saliency value Weight coefficient.
4. the method according to claim 1, wherein at the beginning of the amendment described image using compactedness method of diffusion Grade saliency value, obtains compactedness notable figure, specifically includes:
Construct k-regular closed loop figure, wherein respectively connect side by incidence matrix to measure in the closed loop figure;
It is determined using the manifold smoothness constraint in graph theory based on non-based on described image primary saliency value and the incidence matrix The manifold ranking function for normalizing Laplacian Matrix, to obtain the compactedness notable figure.
5. the method according to claim 1, wherein described set using the topological background of the input picture Reliability and the compactedness notable figure carry out saliency value assignment, obtain image single scale notable figure, specifically include:
Using the topological background confidence level of the input picture and the compactedness notable figure and the super-pixel in CIE L* a*b*Euclidean distance on color space carries out saliency value assignment, obtains the saliency value of the input picture super-pixel;
It assigns the saliency value of the input picture super-pixel to all pixels in the input picture in the super-pixel, makees For the saliency value of those pixels, to obtain described image single scale notable figure.
6. according to the method described in claim 4, it is characterized in that, the incidence matrix is determined by following formula:
W=[wij]n×n,
Wherein, wijIndicate bonding strength;N takes positive integer;dcolor(pi, pj) indicate super-pixel piAnd pjIn CIE L*a*b*Color is empty Between on Euclidean distance;σ indicates adjustment factor;I and j take 1 ..., and N, N indicate super-pixel number.
7. according to the method described in claim 5, it is characterized in that, described set using the topological background of the input picture Reliability and the compactedness notable figure and the super-pixel are in CIE L*a*b*Euclidean distance on color space carries out saliency value Assignment obtains the saliency value of the input picture super-pixel, specifically includes:
The saliency value of the input picture super-pixel is obtained using following objective function:
Wherein, siIndicate super-pixel piSaliency value;sjIndicate super-pixel pjSaliency value;I, j=1 ..., N;N indicates super-pixel Number;biIndicate the topological background confidence level of the input picture;fiIndicate each super-pixel in compactedness notable figure Saliency value;λ indicates regulatory factor;dcolor(pi, pj) indicate the super-pixel piAnd pjIn CIE L*a*b*Europe on color space Family name's distance;σcolorIndicate adjustment factor;μsmIndicate the adjustment factor of inhibition noise.
8. a kind of marking area detection system, which is characterized in that the system comprises:
Divide module, for carrying out super-pixel segmentation to input picture, obtains image superpixel processing unit;
Metric module is connected with the segmentation module, for being based on described image super-pixel processing unit, measures the input figure The topological background confidence level of picture;
Computing module is connected with the metric module, for the topological background confidence level based on the input picture, calculates Contrast on color of image and spatial position feature obtains image primary saliency value;
Correction module is connected with the computing module, for correcting described image primary saliency value using compactedness method of diffusion, Obtain compactedness notable figure;
Assignment module is connected with the metric module and the correction module respectively, for using described in the input picture Topological background confidence level and the compactedness notable figure carry out saliency value assignment, obtain image single scale notable figure;
Fusion Module is connected with the assignment module, for significant to described image single scale using the method for Multiscale Fusion Figure is handled, and scale notable figure is obtained, to obtain image obvious object region.
CN201610889100.3A 2016-10-12 2016-10-12 Salient region detecting method and detection system Active CN106570498B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610889100.3A CN106570498B (en) 2016-10-12 2016-10-12 Salient region detecting method and detection system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610889100.3A CN106570498B (en) 2016-10-12 2016-10-12 Salient region detecting method and detection system

Publications (2)

Publication Number Publication Date
CN106570498A CN106570498A (en) 2017-04-19
CN106570498B true CN106570498B (en) 2019-06-14

Family

ID=58532817

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610889100.3A Active CN106570498B (en) 2016-10-12 2016-10-12 Salient region detecting method and detection system

Country Status (1)

Country Link
CN (1) CN106570498B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108805136B (en) * 2018-03-26 2021-11-05 中国地质大学(武汉) Significance detection method for water surface pollutant monitoring
CN109284753B (en) * 2018-08-30 2020-07-07 深圳大学 Positioning method and application of liquid transmission pipeline
CN109753957B (en) * 2018-12-07 2020-11-27 东软集团股份有限公司 Image significance detection method and device, storage medium and electronic equipment
CN110942095A (en) * 2019-11-27 2020-03-31 中国科学院自动化研究所 Method and system for detecting salient object area

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103745468A (en) * 2014-01-07 2014-04-23 上海交通大学 Significant object detecting method based on graph structure and boundary apriority
CN103914834A (en) * 2014-03-17 2014-07-09 上海交通大学 Significant object detection method based on foreground priori and background priori
CN105046701A (en) * 2015-07-08 2015-11-11 安徽大学 Image composition line-based multi-scale salient target detection method

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103745468A (en) * 2014-01-07 2014-04-23 上海交通大学 Significant object detecting method based on graph structure and boundary apriority
CN103914834A (en) * 2014-03-17 2014-07-09 上海交通大学 Significant object detection method based on foreground priori and background priori
CN105046701A (en) * 2015-07-08 2015-11-11 安徽大学 Image composition line-based multi-scale salient target detection method

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Salient region detection based on Local and Global Saliency;Peng Wang 等;《IEEE International Conference on Robotics & Automation (ICRA)》;20140929;全文 *
Sparse-Distinctive Saliency Detection;Yongkang Luo 等;《IEEE SIGNAL PROCESSING LETTERS》;20141218;全文 *

Also Published As

Publication number Publication date
CN106570498A (en) 2017-04-19

Similar Documents

Publication Publication Date Title
CN106570498B (en) Salient region detecting method and detection system
CN108717710B (en) Positioning method, device and system in indoor environment
US20180278823A1 (en) Auto-exposure technologies using odometry
CN108401457A (en) A kind of control method of exposure, device and unmanned plane
CN110807350A (en) System and method for visual SLAM for scan matching
EP2879080B1 (en) Image processing device and method, and computer readable medium
CN113674416B (en) Three-dimensional map construction method and device, electronic equipment and storage medium
CN108322724B (en) Image solid matching method and binocular vision equipment
CN110222686B (en) Object detection method, object detection device, computer equipment and storage medium
CN110147382A (en) Lane line update method, device, equipment, system and readable storage medium storing program for executing
US8416989B2 (en) Image processing apparatus, image capture apparatus, image processing method, and program
CN105144710A (en) Technologies for increasing the accuracy of depth camera images
CN108474658A (en) Ground Morphology observation method and system, unmanned plane landing method and unmanned plane
JP2020017173A (en) Moving entity
CN109948624A (en) Method, apparatus, electronic equipment and the computer storage medium of feature extraction
El Bouazzaoui et al. Enhancing rgb-d slam performances considering sensor specifications for indoor localization
US9958868B2 (en) Moving object controller, moving object control method, and integrated circuit
CN115655291A (en) Laser SLAM closed-loop mapping method and device, mobile robot, equipment and medium
CN114782496A (en) Object tracking method and device, storage medium and electronic device
CN113971687A (en) Object tracking method, device electronic equipment and storage medium
CN114972465A (en) Image target depth detection method and device, electronic equipment and storage medium
CN105323460B (en) Image processing equipment and its control method
US10447992B1 (en) Image processing method and system
JP2014203133A (en) Image processing device and image processing method
CN110288633A (en) Target tracking method, device, readable storage medium storing program for executing and electronic equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant