CN107680106A - A kind of conspicuousness object detection method based on Faster R CNN - Google Patents

A kind of conspicuousness object detection method based on Faster R CNN Download PDF

Info

Publication number
CN107680106A
CN107680106A CN201710974083.8A CN201710974083A CN107680106A CN 107680106 A CN107680106 A CN 107680106A CN 201710974083 A CN201710974083 A CN 201710974083A CN 107680106 A CN107680106 A CN 107680106A
Authority
CN
China
Prior art keywords
cnn
faster
conspicuousness
pixel
algorithm
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710974083.8A
Other languages
Chinese (zh)
Inventor
王超
李静
刘铭坚
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Aeronautics and Astronautics
Original Assignee
Nanjing University of Aeronautics and Astronautics
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Aeronautics and Astronautics filed Critical Nanjing University of Aeronautics and Astronautics
Priority to CN201710974083.8A priority Critical patent/CN107680106A/en
Publication of CN107680106A publication Critical patent/CN107680106A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/187Segmentation; Edge detection involving region growing; involving region merging; involving connected component labelling
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/194Segmentation; Edge detection involving foreground-background segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image

Abstract

The invention discloses a kind of conspicuousness object detection method based on Faster R CNN.This method carries out multi-scale division to image first, then possible conspicuousness target is outlined using Faster R CNN, establish like thing figure, connected afterwards by prospect and proportion of foreground is assigned in super-pixel, the proportion of conspicuousness Techniques of Optimum combination prospect and background is recycled to obtain round and smooth and smooth notable figure, finally carrying out fusion using multilayer cellular automata obtains final notable figure.Input picture is split on three yardsticks using super-pixel segmentation algorithm, super-pixel segmentation algorithm is according to low-level image features such as color, texture and brightness, adjacent similar pixel is polymerized to image-region of different sizes, the complexity of conspicuousness detection can be effectively reduced;Using the segmentation figure of different scale as one layer of cellular, the super-pixel segmentation figure of each different scale is merged using multilayer cellular automata, ensure that the uniformity of saliency testing result.

Description

A kind of conspicuousness object detection method based on Faster R-CNN
Technical field
The invention belongs to the conspicuousness detection field of image in computer vision, in particular to deep learning to be directed to certain kinds Conspicuousness detection method.
Background technology
In recent years, computer, internet and multimedia technology are fast-developing, and people can connect daily in work and life Touch substantial amounts of image, the information of video.It is that people receive information because image and video contain abundant and intuitively content Effective channel, it is one of important sources of information, for example, the detection of Online Video, Video chat, machine components, network direct broadcasting, intelligence Can monitor etc..Digital picture and video are all in exponential increase, just with manually being handled image or video, analyzed often With significant limitation.Computer vision is as a collection Intelligent Information Processing and the cross discipline of digital assay, simulation Human visual system carries out perception processing to image, obtains and the same or like result of artificial treatment so that computer can be with Analyzed as people and understand the true content expressed by image.Research shows that information of concern, which often focuses on, to be had In the target of vision significance., can be aobvious in input picture by only handling when needing amount of images to be processed more huge Work property target so that improve treatment effeciency in the case where not changing processing method.
The low-level image feature that vision system can identify is included with color, edge, texture etc..Carried out based on these features local Diversity between each region is calculated in contrast or global contrast, determines signal portion in image.Itti et al. is in vision On the basis of attention model KU, it is proposed that classical conspicuousness computation model, turn into a master die of existing algorithm comparison Type.The algorithm, for multi-scale image, phase is obtained by center-surrounding contrast's degree of low-level image feature according to human-eye visual characteristic The notable figure answered, and final notable figure is obtained by notable figure fusion.Harel et al. proposes that the conspicuousness detection based on graph theory is calculated Method GVBS, biological vision principle is combined with mathematical computations, Ma Erke is introduced among the notable figure generating process of Itti models Husband's chain, seek its balanced distribution with the calculating of pure mathematics and obtain notable figure.Cheng et al. is special by color using gauss hybrid models Levy similar pixel to gather for image-region, consider the color contrast and spatial distribution in each region, generated with probabilistic model Notable figure.Yang et al. calculates its color characteristic and space spy by dividing an image into multiple dimensioned figure layer, for each figure layer The contrast of sign, the notable figure for merging multiple figure layer generations obtain final notable figure.This method can ensure conspicuousness target Uniformity and integrality, but when conspicuousness target is smaller, conspicuousness target can be dissolved into background area as background.Use The conspicuousness detection method of different low-level image features often just for certain a kind of specific image significant effect, can not be applied to complicated field Multi-Target Image under scape.The low-level image feature that view-based access control model stimulates lacks the understanding to well-marked target essence, it is impossible to deeper Represent conspicuousness clarification of objective.It is such as similar to low-level image feature but be not belonging to same class for noise object present in image Target, well-marked target is often detected as by mistake.In recent years, the automatic study based on deep learning obtained depth characteristic The method of (or high-level characteristic) has begun to be applied in saliency detects.Li et al. passes through depth convolutional Neural net Network detects MDF, detection effect to learn to obtain the part and the global depth feature that obtain image superpixel region to carry out conspicuousness Fruit compared to conventional method have it is bright be obviously improved, but the speed of service is slower.Lee et al. is by extracting super-pixel block region and side Edge feature, it is sent into convolutional neural networks and learns to obtain notable confidence map.The region for seeking energy minimization using condition random field is entered Row conspicuousness detects.It is preferable to single well-marked target Detection results, but because feature selection issues are not suitable for multi-Target Image. Pan et al. carries out Pixel-level processing using a kind of method that gradient declines to original image, then carries out the depth spy of super-pixel level Sign extraction and saliency value calculate, boosting algorithm speed, but gradient descent method is poor to sparse matrix effect, causes whole detection Effect is inaccurate.Hu et al. obtains part and global characteristics by combining the priori of convolutional neural networks and area validation. Algorithm Detection results are preferable, but highly complex model have impact on algorithm operational efficiency.
The content of the invention
The present invention is a kind of conspicuousness object detection method based on Faster R-CNN, it is intended to solves existing conspicuousness inspection The Deep Semantics feature of image can not be extracted using low-level image feature by surveying in model, and the effect for causing conspicuousness to detect is not ideal The problem of.
The invention provides a kind of conspicuousness detection method based on depth characteristic, comprise the steps:
Step 1:Multiple dimensioned super-pixel segmentation.It is using SLIC super-pixel segmentations algorithm that input picture is enterprising in three yardsticks Row segmentation, SLIC super-pixel segmentation algorithms are according to low-level image features such as color, texture and brightness, and adjacent similar pixel is gathered Into image-region of different sizes, the complexity that conspicuousness calculates can be effectively reduced.It is to be based on color and distance similarity A kind of gradient descent algorithm of measurement, dividing number can control, and it is relatively uniform can to obtain the more regular size of shape Segmentation result.Split yardstick number and each yardstick under super-pixel quantity number, can all influence Detection results and operation is fast Degree.Too many segmentation number can increase computation complexity, and be likely to occur over-segmentation phenomenon.And super-pixel number is very little, then The accuracy of conspicuousness Detection results can be influenceed.Different segmentation yardsticks can influence the effect and the speed of service of subsequent detection, Therefore invention rule of thumb limits the number of pixel in each super-pixel and includes 100 pixels from each super-pixel unit, to 500 Between individual pixel, each yardstick is incremented by by 100 pixels.
Step 2:Obtain like thing figure., will by carrying out certain kinds target detection training to Faster R-CNN neutral nets Certain kinds target detection in image comes out, recycle its characteristic will likely Objective extraction come out.And by obtained target seemingly Physical property fraction extracts.Then, generation is started like thing figure.Faster R-CNN verification and measurement ratio is very high, but in extreme circumstances Certain kinds target can be can't detect, when Faster R-CNN can't detect target if using entire image as target at Reason.
Possibility that target has told us this window contains target like physical property fraction, using may target obtain picture Plain level like physical property fraction, the fraction is designated as the significant possibility of target, such as formula of the definition like physical property fraction of Pixel-level (1), wherein siFor either with or without the possible fraction of the target comprising pixel p, GiFor Gaussian window, x, y are the coordinates of pixel p.
Then Pixel-level is that target is divided in super-pixel region like physical property like summation of the physical property fraction in super-pixel region Number, definition such as formula (2), piIt is a pixel for belonging to super-pixel region R.In order to obtain super-pixel region, we use SLIC, in super-pixel segmentation, we have chosen multiple segmentation yardsticks to be detected.We set a suitable threshold afterwards Value is set as that the pixel quantity like thing figure is more than 1.5 times divided by figure like object image prime number amount to obtain like thing figure, threshold value Total amount of pixels.
Objectness (R)=∑i∈RPixObj(pi) (2)
Step 3:Calculating prospect connects.Threshold value utilizes like the part that thing figure is coarse acquisition prospect super-pixel " prospect connects " (the Foreground Connectivity) that Srivatsa proposes with Babu, this method are utilized according to super-pixel Connective significance value is assigned to the prospect estimated.Figure, neighbouring super-pixel section are established by the use of super-pixel as node There is side between point, the weight definition on side is the Euclidean distance of the average Lab colors of two nodes.Define super-pixel R prospect Connection such as formula (3).Here d (R, Rk) represent R to RkBetween beeline, δ () like thing figure if prospect is estimated as It is entered as the sum that 1, N is super-pixel.
It is more more more can guarantee that the value of molecule is lower and the value of denominator is higher to be estimated super-pixel similitude for prospect, so Son will make it that FG value is lower, which imply that having higher connectedness, so taking FG inverse as prospect weight.
Step 4:Conspicuousness optimizes.Directly our prospect weight and background are combined using existing optimization framework Weight, shown in the cost function such as formula (4) of minimum, tiRepresent to be assigned to p after minimizing costiEnd value,Represent With super-pixel piRelevant prospect weight,Represent and super-pixel piRelevant background weight,More high then piIt is intended to 1, More high then piIt is intended to 0, wijIt is smoothing factor.
Step 5:Multilayer cellular automata merges.In the multiple dimensioned notable figure obtained by above step, because division yardstick Difference to obtain notable figure also not quite identical.Simultaneously because conspicuousness calculating is in units of super-pixel, well-marked target Saliency value is also piecemeal, discontinuous, so multiple dimensioned notable figure is using there are still differ after the fusion of in general method Cause, in order that final notable figure is consistent as far as possible, consensus optimization is carried out using multilayer cellular automata.
Brief description of the drawings
Fig. 1 is the conspicuousness detection method flow chart in depth characteristic;
Fig. 2 is CNN structural representations;
Fig. 3 is the depth characteristic extraction Organization Chart in convolutional neural networks;
Fig. 4 is lamination C1 output characteristic figures;
Fig. 5 is Fast R-CNN frame diagrams;
Fig. 6 is PR curve map of the algorithms of different on certain kinds data set and MAE block diagrams;
Fig. 7 is PR curve map of the algorithms of different on HKU-IS data sets and MAE block diagrams;
Fig. 8 is PR curve map of the algorithms of different on MSRA-1000 data sets and MAE block diagrams.
Embodiment
To be easy to understand the technical means, the inventive features, the objects and the advantages of the present invention, with reference to Embodiment, the present invention is expanded on further.
The present invention carries out multi-scale division to image first, then outlines possible conspicuousness mesh using Faster R-CNN Mark, establish like thing figure, connected afterwards by prospect and proportion of foreground is assigned in super-pixel, recycle conspicuousness Techniques of Optimum Round and smooth and smooth notable figure is obtained with reference to the proportion of prospect and background, finally carrying out fusion using multilayer cellular automata obtains Obtain notable figure finally.Specifically include following steps:
Step 1:Multiple dimensioned super-pixel segmentation.It is using SLIC super-pixel segmentations algorithm that input picture is enterprising in three yardsticks Row segmentation, SLIC super-pixel segmentation algorithms are according to low-level image features such as color, texture and brightness, and adjacent similar pixel is gathered Into image-region of different sizes, the complexity that conspicuousness calculates can be effectively reduced.It is to be based on color and distance similarity A kind of gradient descent algorithm of measurement, dividing number can control, and it is relatively uniform can to obtain the more regular size of shape Segmentation result.Split yardstick number and each yardstick under super-pixel quantity number, can all influence Detection results and operation is fast Degree.Too many segmentation number can increase computation complexity, and be likely to occur over-segmentation phenomenon.And super-pixel number is very little, then The accuracy of conspicuousness Detection results can be influenceed.Different segmentation yardsticks can influence the effect and the speed of service of subsequent detection, Therefore invention rule of thumb limits the number of pixel in each super-pixel and includes 100 pixels from each super-pixel unit, to 500 Between individual pixel, each yardstick is incremented by by 100 pixels.Input picture Image, all complexity can be reduced using gray-scale map, RGB coloured images can also be used, if using RGB color image, now input picture has three, respectively RGB component. Input picture generally requires normalization.Handled by convolutional layer (C)-down-sampling layer (S), by the output of last layer and this layer Weight W does convolution and obtains each convolutional layer, and then down-sampling obtains each down-sampling layer, and their output is referred to as characteristic pattern.Will Image rasterization (X).Each pixel in the characteristic pattern of last layer output is taken out successively, lines up a vector.Using multilayer Perceptron (N&O) is finally handled, and features training grader typically uses Softmax, if two classification, can use logic Homing method.
Step 2:Obtain like thing figure.Faster R-CNN (convolutional neural networks based on region faster) are to be based on Fast R-CNN (the quickly convolutional neural networks based on region) improvement, and Fast R-CNN are then the R-CNN (volumes based on region Product neutral net) merged with SPPNet (spatial pyramid pond network).These four are all based on convolution god in image procossing Object detection method through network, the developmental sequence between them are R-CNN, SPPNet, Fast R-CNN, Faster successively R-CNN.R-CNN core concept is extracted using candidate region frame to replace traditional sliding window to detect, and then to each Feature is extracted in candidate region using CNN, connects the grader of stand-alone training afterwards to predict that the region includes interesting target Confidence level, this is transformed into the method for an image classification.It this method solve CNN orientation problem, but since it is desired that Each region candidate frame is handled, so there is largely compute repeatedly.R-CNN frame diagram is as shown in Figure 5.
SPPNet core concept is utilization space pyramid pond (spatial pyramid pooling, abbreviation SPP) Layer removes the limitation of network fixed size, spatial pyramid pond Hua Ceng ponds feature and produces the output of fixed size.Cause For in CNN full articulamentum be the restriction that has fixed input picture size, and the image that conventional part need not then be fixed is big It is small, so spatial pyramid pond layer is placed on behind last convolutional layer, before first full articulamentum.This method Strategy is improved close to 100 times than the speed of R-CNN target detection.
Fast R-CNN use for reference SPPNet thought, it is proposed that layer region interested (Region of Interest, Abbreviation RoI), the SPPNet of individual layer Internet is can be regarded as, different size of input can be mapped to fixation by RoI The characteristic vector of yardstick, softmax features trainings grader progress type identification and window regression algorithm is recycled to be positioned. The advantages of Fast R-CNN combination SPPNet and R-CNN, greatly reduces the time of target detection process needs.
By carrying out certain kinds target detection training to FasterR-CNN neutral nets, by the certain kinds target in image Detect, recycle its characteristic will likely Objective extraction come out.And obtained target is extracted like physical property fraction.So Afterwards, generation is started like thing figure.Faster R-CNN verification and measurement ratio is very high, but can can't detect specific classification in extreme circumstances Mark, when Faster R-CNN can't detect target if handled entire image as target.
Possibility that target has told us this window contains target like physical property fraction, using may target obtain picture Plain level like physical property fraction, the fraction is used for teaching that the possibility that this pixel is a target part, Pixel-level like thing The definition such as formula (1) of property fraction, wherein siFor either with or without the possible fraction of the target comprising pixel p, GiFor Gaussian window, x, y are pictures Plain p coordinate.
Then Pixel-level is that target is divided in super-pixel region like physical property like summation of the physical property fraction in super-pixel region Number, definition such as formula (2), piIt is a pixel for belonging to super-pixel region R.In order to obtain super-pixel region, we use SLIC, in super-pixel segmentation, we have chosen multiple segmentation yardsticks to be detected.We set a suitable threshold afterwards Value is set as that the pixel quantity like thing figure is more than 1.5 times divided by figure like object image prime number amount to obtain like thing figure, threshold value Total amount of pixels.
Objectness (R)=∑i∈RPixObj(pi) (2)
Step 3:Calculating prospect connects.Threshold value utilizes like the part that thing figure is coarse acquisition prospect super-pixel " prospect connects " (the Foreground Connectivity) that Srivatsa proposes with Babu, this method are utilized according to super-pixel Connective significance value is assigned to the prospect estimated.Figure, neighbouring super-pixel section are established by the use of super-pixel as node There is side between point, the weight definition on side is the Euclidean distance of the average Lab colors of two nodes.Define super-pixel R prospect Connection such as formula (3).Here d (R, Rk) represent R to RkBetween beeline, δ () like thing figure if prospect is estimated as It is entered as the sum that 1, N is super-pixel.
It is more more more can guarantee that the value of molecule is lower and the value of denominator is higher to be estimated super-pixel similitude for prospect, so Son will make it that FG value is lower, which imply that having higher connectedness, so taking FG inverse as prospect weight.
Step 4:Conspicuousness optimizes.Directly our prospect weight and background are combined using existing optimization framework Weight, shown in the cost function such as formula (4) of minimum, tiRepresent to be assigned to p after minimizing costiEnd value,Represent With super-pixel piRelevant prospect weight,Represent and super-pixel piRelevant background weight,More high then piIt is intended to 1, More high then piIt is intended to 0, wijIt is smoothing factor.
Step 5:Multilayer cellular automata merges.It is linear among currently used notable figure or Feature fusion to melt The methods of conjunction, the dot product fusion based on pixel, condition random field are merged, merged based on cellular automata.What Qin et al. was proposed Multilayer cellular automata (Multi-layer Cellular Automata, abbreviation MCA) fusion method is relatively preferable.This method By the notable figure that different algorithms obtains as one layer of cellular automata, and the advantage of algorithms of different is combined, updated by one kind Mechanism constantly updates saliency value, finally gives the notable figure of fusion.
In multilayer cellular automata, each pixel of notable figure is exactly a cellular, in N layer automatic machines, arbitrarily Cellular in one notable figure has N-1 neighbours, the same position being respectively in other notable figures.With pixel i saliency value Indicate it as prospect F probability P (i ∈ F)=Si, then probability P (i ∈ the B)=1-S as background BiRepresent.By carrying Take the adaptive threshold of each width figure, the binary-state threshold γ of m width notable figuresmRepresent, and binary conversion treatment is carried out to it. If pixel i saliency value Si≥γm, then prospect is marked as, and use ηi=+1 represents, on the contrary, ηi=-1, represent pixel i For background.
If pixel i is marked as prospect, then its neighbours j of same position in other notable figures is marked as prospect Probability be that the probability of prospect is λ=P (ηj=+1 | i ∈ F).So, with μ=P (ηj=-1 | i ∈ B) represent that pixel i is labeled For background when, its neighbours j turn into background probability.It is assumed that λ and μ are a pair of equal constants, then and posterior probability P (i ∈ F | ηj =+1) formula (5) can be used to represent:
P(i∈F|ηj=+1) ∝ P (i ∈ F) P (ηj=+1 | i ∈ F)=Si·λ (5)
Prior probability ratio is defined as Λ (i ∈ F), calculation formula is then (6):
Then posterior probability ratio Λ (i ∈ F | ηj=+1) as formula (7) represents:
In order to facilitate calculating, make l=ln (Λ) conversion to above formula, as shown in formula (8):
For the ease of representing the update mechanism of cellular automata, by prior probability ratio and posterior probability such as formula (9) institute Show:
WhereinRepresent pixel i in the saliency value of t, synchronized update mechanism f:SM-1→ S is defined such as formula (3.10) institute Show:
Wherein,T is represented, the saliency value of all cellular automatas in m width notable figures. Matrix 1 is the matrix [1 ..., 1] for having N number of elementT.If pixel its neighbour are judged as prospect, accordingly increase The saliency value of itself, i.e., should haveThen there are λ > 0.5. rule of thumb to setThen N2Secondary renewal Afterwards, final notable figure can be obtained by formula (11)
Due to having carried out multi-scale division to image at the beginning of conspicuousness target detection, splitting the selection of yardstick is Selected on the premise of performance and effect is ensured, finally have selected five yardsticks, 100 pictures are included from each super-pixel unit Element, between 500 pixels, each yardstick is incremented by by 100 pixels.Finally optimized to obtain five yardsticks according to conspicuousness Notable figure, carried out consensus using multilayer cellular automata.In the multiple dimensioned notable figure obtained by above step, because To divide the difference of yardstick, to obtain notable figure also not quite identical.Simultaneously because conspicuousness calculating is in units of super-pixel, The saliency value of well-marked target is also piecemeal, discontinuous, so multiple dimensioned notable figure is merged afterwards still using in general method So exist inconsistent, in order that final notable figure is consistent as far as possible, carries out uniformity using multilayer cellular automata and melt Close optimization.

Claims (7)

1. the conspicuousness algorithm of target detection according to claim 1 based on Faster R-CNN, it is characterised in that include Following steps:I. Faster R-CNN target detections, including the pixel of the gray value of each pixel, square are carried out to input picture The grayscale digital image of battle array, obtains candidate target region frame;II. use super-pixel segmentation algorithm by input picture in three yardsticks On split, super-pixel segmentation algorithm is according to low-level image features such as color, texture and brightness, and adjacent similar pixel is gathered Into image-region of different sizes, the complexity that conspicuousness calculates can be effectively reduced;III. by Faster R-CNN god Certain kinds target detection training is carried out through network, the certain kinds target detection in image is come out, recycles its characteristic will likely Objective extraction is come out, and obtained target is extracted like physical property fraction, is generated like thing figure.IV. with point of different scale Figure is cut as one layer of cellular, the super-pixel segmentation figure of each different scale is merged using multilayer cellular automata, merged Notable figure afterwards is final notable figure.
2. the conspicuousness algorithm of target detection of the certain kinds according to claim 1 based on Faster R-CNN, its feature Conspicuousness target in the image after by super-pixel segmentation substantially chooses the image using Faster R-CNN, reduces it The complexity of conspicuousness detection afterwards.The uniform segmentation result of shape size for obtain using multiple dimensioned super-pixel segmentation, Ensure the uniformity of saliency detection.
3. the conspicuousness algorithm of target detection of the certain kinds according to claim 1 based on Faster R-CNN, its feature It is to carry out Faster R-CNN target detections to input picture, includes the ash of the pixel of the gray value of each pixel, matrix Rank digital picture, start to calculate candidate region window first with Fast R-CNN, simultaneously, RPN is also defeated using convolutional layer The characteristic pattern gone out calculates candidate region, and after both calculating terminate, RPN threshold values according to expected from the user of setting pass to Fast The candidate region window that R-CNN should be selected, obtain the substantially target in saliency region.
4. the conspicuousness algorithm of target detection of the certain kinds based on Faster R-CNN, it is characterised in that Faster R-CNN are applied Optimize the result of notable figure, including following characteristics for certain kinds, while the methods of application is multiple dimensioned, cellular automata:
1) this algorithm is the algorithm that Faster R-CNN first Applications detect for the conspicuousness of certain kinds (aircraft, ship, cat etc.);
2) this algorithm is in Faster R-CNN by multi-scale division first Application;
3) this algorithm is first Application Faster R-CNN acquisitions like thing figure;
4) multilayer cellular automata is applied to Faster R-CNN by this algorithm first.
5. the conspicuousness algorithm of target detection of the certain kinds according to claim 4 based on Faster R-CNN, its feature It is the algorithm for first detecting Faster R-CNN first Applications for the conspicuousness of certain kinds (aircraft, ship, cat etc.).In recent years Coming, conspicuousness algorithm has a lot, including based on the preferential detection model of background, based on the preferential detection model of contrast, but base It is limited in the feature that the conspicuousness detection model of low-level feature extracts, it is impossible to make full use of the feature of image.This algorithm passes through Faster R-CNN extract the further feature in image.
6. the conspicuousness algorithm of target detection of the certain kinds according to claim 4 based on Faster R-CNN, its feature It is to apply multi-scale segmentation method and Faster R-CNN.Conventional super-pixel segmentation algorithm has two kinds of SLIC and watershed Partitioning algorithm.Using multiple multi-scale segmentation images, candidate region, feature extraction, target are divided using Faster R-CNN afterwards Class and position refine merge, and region suggests that network shares convolution feature calculation with Fast R-CNN, and input picture is shown Show and be input in sorter network common on ImageNet, then Fast R-CNN start to calculate candidate region window, same in this When, the characteristic pattern that RPN is also exported using convolutional layer calculates candidate region, both calculate terminate after, RPN is according to the user of setting Expected threshold value tells which candidate region window Fast R-CNN should select.Using the Faster R-CNN trained come to defeated The image entered carries out target detection, then obtains candidate target region frame, extracts this regional frame and store, and is next step Conspicuousness detection provides preparation.
7. the conspicuousness algorithm of target detection of the certain kinds according to claim 4 based on Faster R-CNN, its feature It is to utilize Faster R-CNN, by carrying out it certain kinds target detection training, by the certain kinds target detection in image Out, recycle its characteristic will likely Objective extraction come out, finally provide the target window of optimization, it would be possible to the position of target Extract, and obtained target is extracted like physical property fraction.Then generation is started like thing figure.
CN201710974083.8A 2017-10-13 2017-10-13 A kind of conspicuousness object detection method based on Faster R CNN Pending CN107680106A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710974083.8A CN107680106A (en) 2017-10-13 2017-10-13 A kind of conspicuousness object detection method based on Faster R CNN

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710974083.8A CN107680106A (en) 2017-10-13 2017-10-13 A kind of conspicuousness object detection method based on Faster R CNN

Publications (1)

Publication Number Publication Date
CN107680106A true CN107680106A (en) 2018-02-09

Family

ID=61140621

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710974083.8A Pending CN107680106A (en) 2017-10-13 2017-10-13 A kind of conspicuousness object detection method based on Faster R CNN

Country Status (1)

Country Link
CN (1) CN107680106A (en)

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108648192A (en) * 2018-05-17 2018-10-12 杭州依图医疗技术有限公司 A kind of method and device of detection tubercle
CN108776815A (en) * 2018-06-04 2018-11-09 国网辽宁省电力有限公司信息通信分公司 A kind of conspicuousness object detection method based on depth characteristic
CN109101897A (en) * 2018-07-20 2018-12-28 中国科学院自动化研究所 Object detection method, system and the relevant device of underwater robot
CN109242032A (en) * 2018-09-21 2019-01-18 桂林电子科技大学 A kind of object detection method based on deep learning
CN109598241A (en) * 2018-12-05 2019-04-09 武汉大学 Satellite image marine vessel recognition methods based on Faster R-CNN
CN109902693A (en) * 2019-02-16 2019-06-18 太原理工大学 One kind being based on more attention spatial pyramid characteristic image recognition methods
CN109902806A (en) * 2019-02-26 2019-06-18 清华大学 Method is determined based on the noise image object boundary frame of convolutional neural networks
CN110084247A (en) * 2019-04-17 2019-08-02 上海师范大学 A kind of multiple dimensioned conspicuousness detection method and device based on fuzzy characteristics
CN110211135A (en) * 2019-06-05 2019-09-06 广东工业大学 A kind of diatom image partition method, device and equipment towards complex background interference
CN110473212A (en) * 2019-08-15 2019-11-19 广东工业大学 A kind of Electronic Speculum diatom image partition method and device merging conspicuousness and super-pixel
CN110472639A (en) * 2019-08-05 2019-11-19 山东工商学院 A kind of target extraction method based on conspicuousness prior information
CN110751155A (en) * 2019-10-14 2020-02-04 西北工业大学 Novel target detection method based on Faster R-CNN
CN111292334A (en) * 2018-12-10 2020-06-16 北京地平线机器人技术研发有限公司 Panoramic image segmentation method and device and electronic equipment
CN111353508A (en) * 2019-12-19 2020-06-30 华南理工大学 Saliency detection method and device based on RGB image pseudo-depth information
WO2020181685A1 (en) * 2019-03-12 2020-09-17 南京邮电大学 Vehicle-mounted video target detection method based on deep learning
CN111680702A (en) * 2020-05-28 2020-09-18 杭州电子科技大学 Method for realizing weak supervision image significance detection by using detection frame
CN111860672A (en) * 2020-07-28 2020-10-30 北京邮电大学 Fine-grained image classification method based on block convolutional neural network
CN111914850A (en) * 2019-05-07 2020-11-10 百度在线网络技术(北京)有限公司 Picture feature extraction method, device, server and medium
CN112085020A (en) * 2020-09-08 2020-12-15 北京印刷学院 Visual saliency target detection method and device
CN112434618A (en) * 2020-11-26 2021-03-02 西安电子科技大学 Video target detection method based on sparse foreground prior, storage medium and equipment
CN112541912A (en) * 2020-12-23 2021-03-23 中国矿业大学 Method and device for rapidly detecting saliency target in mine sudden disaster scene
CN112766032A (en) * 2020-11-26 2021-05-07 电子科技大学 SAR image saliency map generation method based on multi-scale and super-pixel segmentation
CN114332572A (en) * 2021-12-15 2022-04-12 南方医科大学 Method for extracting breast lesion ultrasonic image multi-scale fusion characteristic parameters based on saliency map guided hierarchical dense characteristic fusion network
CN116645368A (en) * 2023-07-27 2023-08-25 青岛伟东包装有限公司 Online visual detection method for edge curl of casting film

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106611427A (en) * 2015-10-21 2017-05-03 中国人民解放军理工大学 A video saliency detection method based on candidate area merging
CN106780582A (en) * 2016-12-16 2017-05-31 西安电子科技大学 Based on the image significance detection method that textural characteristics and color characteristic are merged

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106611427A (en) * 2015-10-21 2017-05-03 中国人民解放军理工大学 A video saliency detection method based on candidate area merging
CN106780582A (en) * 2016-12-16 2017-05-31 西安电子科技大学 Based on the image significance detection method that textural characteristics and color characteristic are merged

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
SAI SRIVATSA R ETC.: ""SALIENT OBJECT DETECTION VIA OBJECTNESS MEASURE"", 《ARXIV:1506.07363V1[CS.CV]》 *
YAO QIN ETC.: ""Saliency Detection via Cellular Automata"", 《IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION》 *
王刚等: ""一种多尺度超像素显著性检测算法"", 《计算机工程》 *

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108648192A (en) * 2018-05-17 2018-10-12 杭州依图医疗技术有限公司 A kind of method and device of detection tubercle
CN108648192B (en) * 2018-05-17 2021-08-17 杭州依图医疗技术有限公司 Method and device for detecting nodule
CN108776815A (en) * 2018-06-04 2018-11-09 国网辽宁省电力有限公司信息通信分公司 A kind of conspicuousness object detection method based on depth characteristic
CN109101897A (en) * 2018-07-20 2018-12-28 中国科学院自动化研究所 Object detection method, system and the relevant device of underwater robot
CN109242032A (en) * 2018-09-21 2019-01-18 桂林电子科技大学 A kind of object detection method based on deep learning
CN109242032B (en) * 2018-09-21 2021-11-16 桂林电子科技大学 Target detection method based on deep learning
CN109598241A (en) * 2018-12-05 2019-04-09 武汉大学 Satellite image marine vessel recognition methods based on Faster R-CNN
CN109598241B (en) * 2018-12-05 2022-08-12 武汉大学 Satellite image marine ship identification method based on Faster R-CNN
CN111292334A (en) * 2018-12-10 2020-06-16 北京地平线机器人技术研发有限公司 Panoramic image segmentation method and device and electronic equipment
CN111292334B (en) * 2018-12-10 2023-06-09 北京地平线机器人技术研发有限公司 Panoramic image segmentation method and device and electronic equipment
CN109902693A (en) * 2019-02-16 2019-06-18 太原理工大学 One kind being based on more attention spatial pyramid characteristic image recognition methods
CN109902806A (en) * 2019-02-26 2019-06-18 清华大学 Method is determined based on the noise image object boundary frame of convolutional neural networks
WO2020181685A1 (en) * 2019-03-12 2020-09-17 南京邮电大学 Vehicle-mounted video target detection method based on deep learning
JP2021530062A (en) * 2019-03-12 2021-11-04 南京郵電大学Nanjing University Of Posts And Telecommunications In-vehicle video target detection method based on deep learning
JP7120689B2 (en) 2019-03-12 2022-08-17 南京郵電大学 In-Vehicle Video Target Detection Method Based on Deep Learning
CN110084247A (en) * 2019-04-17 2019-08-02 上海师范大学 A kind of multiple dimensioned conspicuousness detection method and device based on fuzzy characteristics
CN111914850B (en) * 2019-05-07 2023-09-19 百度在线网络技术(北京)有限公司 Picture feature extraction method, device, server and medium
CN111914850A (en) * 2019-05-07 2020-11-10 百度在线网络技术(北京)有限公司 Picture feature extraction method, device, server and medium
CN110211135A (en) * 2019-06-05 2019-09-06 广东工业大学 A kind of diatom image partition method, device and equipment towards complex background interference
CN110472639A (en) * 2019-08-05 2019-11-19 山东工商学院 A kind of target extraction method based on conspicuousness prior information
CN110472639B (en) * 2019-08-05 2023-04-18 山东工商学院 Target extraction method based on significance prior information
CN110473212A (en) * 2019-08-15 2019-11-19 广东工业大学 A kind of Electronic Speculum diatom image partition method and device merging conspicuousness and super-pixel
CN110751155A (en) * 2019-10-14 2020-02-04 西北工业大学 Novel target detection method based on Faster R-CNN
CN111353508A (en) * 2019-12-19 2020-06-30 华南理工大学 Saliency detection method and device based on RGB image pseudo-depth information
CN111680702A (en) * 2020-05-28 2020-09-18 杭州电子科技大学 Method for realizing weak supervision image significance detection by using detection frame
CN111860672B (en) * 2020-07-28 2021-03-16 北京邮电大学 Fine-grained image classification method based on block convolutional neural network
CN111860672A (en) * 2020-07-28 2020-10-30 北京邮电大学 Fine-grained image classification method based on block convolutional neural network
CN112085020B (en) * 2020-09-08 2023-08-01 北京印刷学院 Visual saliency target detection method and device
CN112085020A (en) * 2020-09-08 2020-12-15 北京印刷学院 Visual saliency target detection method and device
CN112434618B (en) * 2020-11-26 2023-06-23 西安电子科技大学 Video target detection method, storage medium and device based on sparse foreground priori
CN112766032A (en) * 2020-11-26 2021-05-07 电子科技大学 SAR image saliency map generation method based on multi-scale and super-pixel segmentation
CN112434618A (en) * 2020-11-26 2021-03-02 西安电子科技大学 Video target detection method based on sparse foreground prior, storage medium and equipment
CN112541912A (en) * 2020-12-23 2021-03-23 中国矿业大学 Method and device for rapidly detecting saliency target in mine sudden disaster scene
CN112541912B (en) * 2020-12-23 2024-03-12 中国矿业大学 Rapid detection method and device for salient targets in mine sudden disaster scene
CN114332572A (en) * 2021-12-15 2022-04-12 南方医科大学 Method for extracting breast lesion ultrasonic image multi-scale fusion characteristic parameters based on saliency map guided hierarchical dense characteristic fusion network
CN114332572B (en) * 2021-12-15 2024-03-26 南方医科大学 Method for extracting breast lesion ultrasonic image multi-scale fusion characteristic parameters based on saliency map-guided hierarchical dense characteristic fusion network
CN116645368A (en) * 2023-07-27 2023-08-25 青岛伟东包装有限公司 Online visual detection method for edge curl of casting film
CN116645368B (en) * 2023-07-27 2023-10-03 青岛伟东包装有限公司 Online visual detection method for edge curl of casting film

Similar Documents

Publication Publication Date Title
CN107680106A (en) A kind of conspicuousness object detection method based on Faster R CNN
Shen et al. Detection of stored-grain insects using deep learning
Elhoseny et al. Optimal bilateral filter and convolutional neural network based denoising method of medical image measurements
CN109584248B (en) Infrared target instance segmentation method based on feature fusion and dense connection network
CN109919108B (en) Remote sensing image rapid target detection method based on deep hash auxiliary network
Mousavian et al. Joint semantic segmentation and depth estimation with deep convolutional networks
CN106547880B (en) Multi-dimensional geographic scene identification method fusing geographic area knowledge
Byeon et al. Scene labeling with lstm recurrent neural networks
CN112184752A (en) Video target tracking method based on pyramid convolution
CN109002755B (en) Age estimation model construction method and estimation method based on face image
CN109241982A (en) Object detection method based on depth layer convolutional neural networks
CN104392228A (en) Unmanned aerial vehicle image target class detection method based on conditional random field model
CN108171249B (en) RGBD data-based local descriptor learning method
CN107506792B (en) Semi-supervised salient object detection method
CN113436227A (en) Twin network target tracking method based on inverted residual error
CN113743417B (en) Semantic segmentation method and semantic segmentation device
CN111401380A (en) RGB-D image semantic segmentation method based on depth feature enhancement and edge optimization
Mohaghegh et al. Aggregation of rich depth-aware features in a modified stacked generalization model for single image depth estimation
CN115115863A (en) Water surface multi-scale target detection method, device and system and storage medium
Yuan et al. A lightweight network for smoke semantic segmentation
Baha et al. Accurate real-time neural disparity MAP estimation with FPGA
Bagwari et al. A comprehensive review on segmentation techniques for satellite images
CN114743045B (en) Small sample target detection method based on double-branch area suggestion network
CN110853052A (en) Tujia brocade pattern primitive segmentation method based on deep learning
Li et al. Ship segmentation via encoder–decoder network with global attention in high-resolution SAR images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20180209

WD01 Invention patent application deemed withdrawn after publication