CN106157319B - The conspicuousness detection method in region and Pixel-level fusion based on convolutional neural networks - Google Patents

The conspicuousness detection method in region and Pixel-level fusion based on convolutional neural networks Download PDF

Info

Publication number
CN106157319B
CN106157319B CN201610604732.0A CN201610604732A CN106157319B CN 106157319 B CN106157319 B CN 106157319B CN 201610604732 A CN201610604732 A CN 201610604732A CN 106157319 B CN106157319 B CN 106157319B
Authority
CN
China
Prior art keywords
pixel
conspicuousness
level
cnn
region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201610604732.0A
Other languages
Chinese (zh)
Other versions
CN106157319A (en
Inventor
邬向前
卜巍
唐有宝
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin Institute of Technology
Original Assignee
Harbin Institute of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin Institute of Technology filed Critical Harbin Institute of Technology
Priority to CN201610604732.0A priority Critical patent/CN106157319B/en
Publication of CN106157319A publication Critical patent/CN106157319A/en
Application granted granted Critical
Publication of CN106157319B publication Critical patent/CN106157319B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses the conspicuousness detection methods that a kind of region based on convolutional neural networks and Pixel-level are merged, the object of this method research is still image, wherein the content of image can be arbitrary, the target of research is the target for attracting human eye attention to be found out from image, and different significance value is assigned for it.The present invention mainly proposes a kind of adaptive Area growth, and devises two CNN network structures, is respectively used to the prediction of Pixel-level conspicuousness and conspicuousness fusion.The two CNN network models are used for the training of network model, and final output and input picture Saliency maps of the same size using the legitimate reading of image as supervisory signals using image as input.The estimation of region class conspicuousness and the prediction of Pixel-level conspicuousness can be effectively performed in the present invention, obtain two Saliency maps, finally merged to obtain final Saliency maps by two Saliency maps and original image using the CNN for carrying out conspicuousness fusion.

Description

The conspicuousness detection method in region and Pixel-level fusion based on convolutional neural networks
Technical field
The present invention relates to a kind of image processing methods based on deep learning, and in particular to one kind being based on convolutional neural networks Region and Pixel-level fusion conspicuousness detection method.
Background technology
With the development and rise of deep learning, the conspicuousness detection technique based on deep learning also continues to develop.Significantly Property detection can be divided into bottom-up data-driven model and top-down task-driven model two major classes.It is bottom-up aobvious The detection of work property refers to finding out attractive target in figure, which can be any classification for given arbitrary piece image Things.And top-down conspicuousness detection method usually finds out the target of given classification from given picture, and assign not Same significance value.Currently, to the most study of bottom-up conspicuousness detection method.
Existing bottom-up conspicuousness detection method can be divided into two classes, the respectively method based on hand-designed feature With the method based on convolutional neural networks.The surface information in image is usually utilized due to the method based on hand-designed feature (Such as color, texture etc.)These manual features to carry out feature extraction, and extract cannot capture the deep layer of conspicuousness target Characteristic and multi-scale information, to which good performance cannot be obtained.Recently with the rise of deep learning, there is part researcher to open Begin to carry out conspicuousness target detection using convolutional neural networks.It is existing at present to be mostly based on convolutional neural networks(CNN) Conspicuousness object detection method divide the image into multiple regions first, then to the trained CNN models in each region To predict its significance value.But these methods cannot obtain the conspicuousness prediction result of accurate Pixel-level.
Invention content
In order to overcome the above problem, the present invention to propose a kind of new conspicuousness detection method based on CNN, that is, be based on volume The conspicuousness detection method in the region and Pixel-level fusion of product neural network.This method includes three phases, respectively region class Conspicuousness estimation, the prediction of Pixel-level conspicuousness and conspicuousness fusion, each stage are directed to a CNN model.This method can The Saliency maps of accurate Pixel-level are obtained, so as to more effectively push the development of the related application based on conspicuousness.
The purpose of the present invention is what is be achieved through the following technical solutions:
The present invention provides the conspicuousness detection methods that a kind of region based on convolutional neural networks and Pixel-level are merged, should The object of technique study is still image, and wherein the content of image can be arbitrary, and the target of research is found out from image Attract the target of human eye attention, and assigns different significance value for it.The present invention mainly proposes a kind of adaptive area Domain generation technique, and two CNN network structures are devised, it is respectively used to the prediction of Pixel-level conspicuousness and conspicuousness fusion.This two A CNN network models are used for the training of network model using the legitimate reading of image as supervisory signals using image as input, and Final output and input picture Saliency maps of the same size.The estimation of region class conspicuousness and pixel can be effectively performed in the present invention Grade conspicuousness prediction, obtains two Saliency maps, finally using carrying out the CNN of conspicuousness fusion by two Saliency maps and original Image is merged to obtain final Saliency maps, and whole system block diagram is as shown in Figure 1.
The specific implementation step of the present invention is as follows:
One, region class conspicuousness is estimated
The first step is split input picture I using adaptive Area growth
(1)Super-pixel segmentation is carried out to input picture I using SLIC algorithms, obtains n super-pixel;
(2)A simple feature vector is extracted from each super-pixel, for characterizing the characteristic of the super-pixel;
(3)Super-pixel is clustered to obtain different regions using an agglomerative clustering algorithm based on figure;
Second step carries out region significance estimation using Clarifai network models
(1)M super-pixel is randomly selected around the center line in each region;
(2)The m window of center construction of center as window of m super-pixel is set, and these windows include entire figure Picture;
(3)M video in window of construction is passed through into CNN models, obtains m significance value;
(4)Calculate the mean value of m significance value and the significance value as the region;
Two, Pixel-level conspicuousness is predicted
(1)Using model VGGNet as pre-training model, by the last one module removal in VGGNet, to the 4th and The output of five modules carries out operation of deconvoluting, and splices them in feature channel direction, learns for Analysis On Multi-scale Features; Then it uses size to carry out convolution to spliced characteristic pattern for the convolution kernel of 1* 1 and obtains a probability graph;
(2)In Pixel-level CNN model training stages, calculated between probability graph and legitimate reading figure using fork entropy loss function Error, and error is returned to update Pixel-level CNN model parameters;
(3)After Pixel-level CNN model trainings are complete, input picture I is directly inputted in Pixel-level CNN models and predicts it Corresponding Pixel-level Saliency maps;
Three, conspicuousness merges
(1)Structure fusion CNN network structures:CNN network structures include a splicing layer, three convolutional layers and a loss Layer;
(2)By input picture I and Step 1: two two Saliency maps are spliced into the image in 5 channels, then by it It is sent into three convolutional layers;
(3)In the fusion CNN network training stages, the last one convolutional layer is calculated using the fork entropy loss function in loss layer Output and true Saliency maps between error, and error is returned to update fusion CNN model parameters;
(4)When test, input picture I is directly inputted in trained fusion CNN models, the last one volume of the model The Saliency maps that the output of lamination is as finally predicted.
The invention has the advantages that:
1, the present invention proposes a kind of new conspicuousness detection method based on CNN, and this method has given full play to region class Conspicuousness estimates and the advantage of Pixel-level conspicuousness prediction, and achieves good conspicuousness detection performance.
2, the present invention proposes a kind of adaptive Area growth, which can be that different images generates different numbers The region of amount, and the edge of object can be kept well.
3, the present invention devises a CNN network structure, which can effectively excavate the multiple dimensioned letter in image Breath can also carry out and the relevant task of pixel classifications, such as image segmentation other than it can be used for the prediction of Pixel-level conspicuousness.
4, the present invention proposes a kind of new conspicuousness convergence strategy based on CNN, not only takes full advantage of Saliency maps Between complementary information, also use information abundant in original image, to largely improve conspicuousness detection Performance.
Description of the drawings
Fig. 1 is whole system block diagram of the present invention;
Fig. 2 is that adaptive region generates result example, and a- original images, b- legitimate readings, c- super-pixel segmentations are as a result, d- Area generation result;
Fig. 3 is region class conspicuousness estimated result example, a- original images, b- legitimate readings, c- region class results;
Fig. 4 is Pixel-level CNN network structures;
Fig. 5 is Pixel-level conspicuousness prediction result example, a- original images, b- legitimate readings, c- Pixel-level results;
Fig. 6 is fusion CNN network structures;
Fig. 7 is present invention progress conspicuousness detection as a result, a- original images, b- legitimate readings, c- fusion results, d- pictures Plain grade is as a result, e- region class results.
Specific implementation mode
Technical scheme of the present invention is further described below in conjunction with the accompanying drawings, however, it is not limited to this, every to this Inventive technique scheme is modified or replaced equivalently, and without departing from the spirit of the technical scheme of the invention and range, should all be covered In protection scope of the present invention.
The present invention provides the conspicuousness detection method that a kind of region based on convolutional neural networks and Pixel-level are merged, tools Body implementation steps are as follows:
One, region class conspicuousness is estimated
In region class conspicuousness estimation procedure, wherein the first step is exactly that a large amount of region is generated from input picture.Most Simple method is to carry out conspicuousness estimation using super-pixel as region, so that how to determine the super-pixel of segmentation Number becomes highly difficult.If super-pixel number is very little, may be owed so that belonging to same conspicuousness mesh target area Segmentation.If super-pixel number is too many, so that the region for belonging to conspicuousness target or background may be by over-segmentation.No matter It is less divided or over-segmentation, all the significance value of conspicuousness target or background may be made inconsistent.Therefore, for different Image, due to their different characteristics, it should be divided into the super-pixel of different number.To solve the above-mentioned problems, of the invention A kind of adaptive Area growth is proposed to carry out image segmentation.An input picture I is given, which generates The process of technology is as follows:
(1)Super-pixel segmentation is carried out to I using SLIC algorithms and obtains n super-pixel.Consider method effect and Efficiency, in the present invention, n=300.
(2)A simple feature vector is extracted from each super-pixel(Contain the average color on Lab color spaces With average spatial position coordinate), for characterizing the characteristic of the super-pixel.
(3)Super-pixel is clustered to obtain different regions using an agglomerative clustering algorithm based on figure.
After the above process, the similar and adjacent super-pixel of color is usually clustered the same area in image I In domain.For different images, the areal obtained after final cluster is also different, and will be far smaller than super-pixel Number n.Fig. 2 provides the result example of adaptive region generation technique three obtained generation.
It is region significance estimation in next step after the region generated.The present invention uses Clarifai network models (It is the CNN models that image classification task obtains top performance in ImageNet2013)To carry out region significance estimation.Tool For body, m super-pixel is randomly selected around the center line in each region first, then the center of this m super-pixel is set M window is built as the center of window, and these windows contain whole image.Super picture around selection region center line Element be in order to(1)So that boundary of the center of constructed window as far away from region,(2)Window from different zones In content it is as different as possible.In the present invention, when the number for the super-pixel for including in region is more than 5, m=5 are set, otherwise, The value of m is set to the number of super-pixel.From the foregoing, it will be observed that for each region, m video in window will be constructed, by CNN models Afterwards, m significance value will be obtained, their mean value and the significance value as the region are calculated, so that the region Significance value it is more robust to noise.Fig. 3 gives the result example of three region class conspicuousnesses estimation.
Two, Pixel-level conspicuousness is predicted
Although the estimation of region class conspicuousness can obtain consistent and edge and keep good Saliency maps, picture cannot be obtained The Saliency maps of plain class precision.In this regard, the present invention proposes a kind of CNN network structures(It is denoted as Pixel-level CNN)For carrying out pixel Grade conspicuousness prediction.Pixel-level CNN is input with original image, to be output with the Saliency maps of the sizes such as original image. Accurate conspicuousness prediction, the CNN structures should be deep layer and possess different strides in order to obtain(strides)It is multiple dimensioned Stage, so that for image pixel study to the strong Analysis On Multi-scale Features of discriminating power.When training sample scale is smaller, accent Start effectively to train such a network structure to be a very difficult task.Such issues that in order to overcome, one very well Way be exactly use some trained models on large-scale dataset(For example those are extremely successful on ImageNet Network model VGGNet and GoogleNet)As pre-training model, then to model in the small data set in the task of needs It is finely adjusted, a strong model of learning ability is obtained so as to training.
The present invention modifies to it on the basis of VGGNet models, to build Pixel-level CNN models.VGGNet By six modules(block)Composition, first five module is by convolutional layer(It is denoted as conv)With pond layer(It is denoted as pooling)Composition, As shown in Figure 4.The last one module is made of a pond layer and two full articulamentums.The present invention is by last in VGGNet A module removal.In order to which using the multi-scale information of image, the output that the present invention merges the 4th and the 5th module is more to realize Scale feature learns.Since the output size of most latter two module is different and is much smaller than the size of original image, in order to make Pixel-level CNN models can learn Analysis On Multi-scale Features and predict that the present invention first will be to last two for Pixel-level conspicuousness automatically The output of a module carries out operation of deconvoluting(It is denoted as deconv)So that their size is consistent with original image, and Feature channel direction splices them(It is denoted as concat).Then it is the convolution kernel of 1* 1 to spliced spy to use size Sign figure carries out convolution and obtains a probability graph, and in the probability graph, value means more greatly more notable.In test, the probability graph The actually Saliency maps of input picture.In training, fork entropy loss function is used(It is denoted as loss)To calculate the probability Error between figure and legitimate reading figure, and error is returned to update model parameter.This is arrived, entire Pixel-level CNN Network structure all complete by structure, as shown in Figure 4.In model training stage, the stochastic gradient descent algorithm of standard by with To minimize loss function.After model training is complete, image is directly inputted in model to predict that its corresponding Pixel-level is shown Work property figure.Fig. 5 gives the example of three Pixel-level conspicuousness prediction results.
Three, conspicuousness merges
For a given image, two Saliency maps can effectively be obtained by the above process, respectively:Region class is aobvious Work property figure and Pixel-level Saliency maps.Since the CNN models that they are the different information being utilized in image are calculated, They have complementarity.If can effectively be merged to it, the performance of conspicuousness detection will be further increased.
The present invention designs a simple CNN network structure(It is denoted as fusion CNN)And learn a kind of nonlinear transformation to fill Divide the complementary information between excavation regions grade Saliency maps and Pixel-level Saliency maps, high performance purpose is put forward to reach.It should CNN network structures contain a splicing layer(concat), three convolutional layers(conv)With a loss layer(loss), such as Fig. 6 It is shown.Original image and its two Saliency maps are spliced into the image in 5 channels first, are then sent to subsequent Three convolutional layers(Concrete configuration is shown in Fig. 6).In test, conspicuousness that the output of the last one convolutional layer is as finally predicted Figure.In training, using the fork entropy loss function in loss layer come the output for calculating the last one convolutional layer and true conspicuousness Error between figure.From the foregoing, it will be observed that in conspicuousness fusion method proposed by the present invention, in addition to using two to obtain Saliency maps Outside, original image is also used.This is because introduce the abundant information of original image can correct it is certain can not be by using only notable The mistake that property figure is corrected when being merged.
Fusion CNN can be trained individually, also can carry out combination learning with the CNN networks of front to obtain lastness The promotion of energy.In the estimation of region class conspicuousness, most start to need to generate multiple regions from input picture, then using area Grade CNN carries out significance value estimation to each region.And Pixel-level CNN and fusion CNN then directly using image as input, and directly It connects output and obtains Saliency maps, therefore be a process end to end.Therefore, it is difficult to which above three CNN network structures are incorporated In the network unified to one, and carry out combination learning end to end.In order to simplify this process, finally, three CNN are advanced Row individually training, then, Pixel-level CNN and fusion CNN further carry out combination learning on the basis of pre-training, such as Fig. 1 institutes Show.When test, image is inputted into the frame in Fig. 1, the output for merging CNN is the Saliency maps finally predicted.Fig. 7 gives The testing result example of four conspicuousness detection methods proposed by the present invention, as shown in Figure 7, result that the present invention detects with it is true As a result very close to thus illustrating effectiveness of the invention.

Claims (2)

1. a kind of conspicuousness detection method of region and Pixel-level fusion based on convolutional neural networks, it is characterised in that the side Steps are as follows for method:
Step 1: region class conspicuousness is estimated
The first step is split input picture I using adaptive Area growth
(1)Super-pixel segmentation is carried out to input picture I using SLIC algorithms, obtains n super-pixel;
(2)A simple feature vector is extracted from each super-pixel, for characterizing the characteristic of the super-pixel;
(3)Super-pixel is clustered to obtain different regions using an agglomerative clustering algorithm based on figure;
Second step carries out region significance estimation using Clarifai network models
(1)M super-pixel is randomly selected around the center line in each region;
(2)The m window of center construction of center as window of m super-pixel is set, and these windows include whole image;
(3)M video in window of construction is passed through into CNN models, obtains m significance value;
(4)Calculate the mean value of m significance value and the significance value as the region;
Step 2: Pixel-level conspicuousness is predicted
(1)Using model VGGNet as pre-training model, by the last one module removal in VGGNet, to the 4th and the 5th mould The output of block carries out operation of deconvoluting, and splices them in feature channel direction, learns for Analysis On Multi-scale Features;Then It uses size to carry out convolution to spliced characteristic pattern for the convolution kernel of 1* 1 and obtains a probability graph;
(2)In Pixel-level CNN model training stages, the mistake between probability graph and legitimate reading figure is calculated using fork entropy loss function Difference, and error is returned to update Pixel-level CNN model parameters;
(3)After Pixel-level CNN model trainings are complete, input picture I is directly inputted in Pixel-level CNN models and predicts its correspondence Pixel-level Saliency maps;
Step 3: conspicuousness merges
(1)Structure fusion CNN network structures:CNN network structures include a splicing layer, three convolutional layers and a loss layer;
(2)By input picture I and Step 1: two two Saliency maps are spliced into the image in 5 channels, then it is sent to Three convolutional layers;
(3)In the fusion CNN network training stages, the defeated of the last one convolutional layer is calculated using the fork entropy loss function in loss layer Go out the error between true Saliency maps, and error is returned to update fusion CNN model parameters;
(4)When test, input picture I is directly inputted in trained fusion CNN models, the last one convolutional layer of the model The Saliency maps as finally predicted of output.
2. the conspicuousness detection method of the region and Pixel-level fusion according to claim 1 based on convolutional neural networks, It is characterized in that n=300.
CN201610604732.0A 2016-07-28 2016-07-28 The conspicuousness detection method in region and Pixel-level fusion based on convolutional neural networks Active CN106157319B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610604732.0A CN106157319B (en) 2016-07-28 2016-07-28 The conspicuousness detection method in region and Pixel-level fusion based on convolutional neural networks

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610604732.0A CN106157319B (en) 2016-07-28 2016-07-28 The conspicuousness detection method in region and Pixel-level fusion based on convolutional neural networks

Publications (2)

Publication Number Publication Date
CN106157319A CN106157319A (en) 2016-11-23
CN106157319B true CN106157319B (en) 2018-11-02

Family

ID=58060262

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610604732.0A Active CN106157319B (en) 2016-07-28 2016-07-28 The conspicuousness detection method in region and Pixel-level fusion based on convolutional neural networks

Country Status (1)

Country Link
CN (1) CN106157319B (en)

Families Citing this family (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106709532B (en) * 2017-01-25 2020-03-10 京东方科技集团股份有限公司 Image processing method and device
CN106911930A (en) * 2017-03-03 2017-06-30 深圳市唯特视科技有限公司 It is a kind of that the method for perceiving video reconstruction is compressed based on recursive convolution neutral net
CN106934397B (en) * 2017-03-13 2020-09-01 北京市商汤科技开发有限公司 Image processing method and device and electronic equipment
CN107016409A (en) * 2017-03-20 2017-08-04 华中科技大学 A kind of image classification method and system based on salient region of image
CN107169954B (en) * 2017-04-18 2020-06-19 华南理工大学 Image significance detection method based on parallel convolutional neural network
CN107194933A (en) * 2017-04-24 2017-09-22 天津大学 With reference to convolutional neural networks and the brain tumor dividing method and device of fuzzy reasoning
CN109118459B (en) * 2017-06-23 2022-07-19 南开大学 Image salient object detection method and device
CN107369160B (en) * 2017-06-28 2020-04-03 苏州比格威医疗科技有限公司 Choroid neogenesis blood vessel segmentation algorithm in OCT image
CN107437246B (en) * 2017-07-05 2020-08-18 浙江大学 Common significance detection method based on end-to-end full-convolution neural network
JP6907774B2 (en) * 2017-07-14 2021-07-21 オムロン株式会社 Object detectors, object detection methods, and programs
CN107368831B (en) * 2017-07-19 2019-08-02 中国人民解放军国防科学技术大学 English words and digit recognition method in a kind of natural scene image
CN107506792B (en) * 2017-08-16 2020-09-29 广西荷福智能科技有限公司 Semi-supervised salient object detection method
US11222255B2 (en) * 2017-08-17 2022-01-11 Samsung Electronics Co., Ltd. Neuromorphic processing apparatus
CN107730546B (en) * 2017-08-25 2020-11-03 华北电力大学(保定) Image depth feature determination method and system
CN107784308B (en) * 2017-10-09 2020-04-03 哈尔滨工业大学 Saliency target detection method based on chain type multi-scale full-convolution network
CN107766810B (en) * 2017-10-10 2021-05-14 湖南省测绘科技研究所 Cloud and shadow detection method
CN107886533B (en) * 2017-10-26 2021-05-04 深圳大学 Method, device and equipment for detecting visual saliency of three-dimensional image and storage medium
CN107945204B (en) * 2017-10-27 2021-06-25 西安电子科技大学 Pixel-level image matting method based on generation countermeasure network
CN107767383B (en) * 2017-11-01 2021-05-11 太原理工大学 Road image segmentation method based on superpixels
CN107945109B (en) * 2017-11-06 2020-07-28 清华大学 Image splicing method and device based on convolutional network
CN107967474A (en) * 2017-11-24 2018-04-27 上海海事大学 A kind of sea-surface target conspicuousness detection method based on convolutional neural networks
CN108345850B (en) * 2018-01-23 2021-06-01 哈尔滨工业大学 Scene text detection method based on region classification of stroke feature transformation and deep learning of superpixel
CN108389182B (en) * 2018-01-24 2020-07-17 北京卓视智通科技有限责任公司 Image quality detection method and device based on deep neural network
CN110111295B (en) * 2018-02-01 2021-06-11 北京中科奥森数据科技有限公司 Image collaborative saliency detection method and device
CN108961220B (en) * 2018-06-14 2022-07-12 上海大学 Image collaborative saliency detection method based on multilayer convolution feature fusion
CN109086777B (en) * 2018-07-09 2021-09-28 南京师范大学 Saliency map refining method based on global pixel characteristics
CN108960261B (en) * 2018-07-25 2021-09-24 扬州万方电子技术有限责任公司 Salient object detection method based on attention mechanism
CN109409222B (en) * 2018-09-20 2020-10-30 中国地质大学(武汉) Multi-view facial expression recognition method based on mobile terminal
CN109389056B (en) * 2018-09-21 2020-05-26 北京航空航天大学 Space-based multi-view-angle collaborative track surrounding environment detection method
RU2703270C1 (en) 2018-10-31 2019-10-16 Общество с ограниченной ответственностью "Аби Продакшн" Optical character recognition using specialized confidence functions, implemented on the basis of neural networks
CN109409435B (en) * 2018-11-01 2022-07-15 上海大学 Depth perception significance detection method based on convolutional neural network
CN109934241B (en) * 2019-03-28 2022-12-09 南开大学 Image multi-scale information extraction method capable of being integrated into neural network architecture
CN110084221B (en) * 2019-05-08 2023-02-03 南京云智控产业技术研究院有限公司 Serialized human face key point detection method with relay supervision based on deep learning
CN110166850B (en) * 2019-05-30 2020-11-06 上海交通大学 Method and system for predicting panoramic video watching position by multiple CNN networks
CN110222704B (en) * 2019-06-12 2022-04-01 北京邮电大学 Weak supervision target detection method and device
CN110390327B (en) * 2019-06-25 2022-06-28 北京百度网讯科技有限公司 Foreground extraction method and device, computer equipment and storage medium
CN112149459B (en) * 2019-06-27 2023-07-25 哈尔滨工业大学(深圳) Video saliency object detection model and system based on cross attention mechanism
CN110472639B (en) * 2019-08-05 2023-04-18 山东工商学院 Target extraction method based on significance prior information
CN111583173B (en) * 2020-03-20 2023-12-01 北京交通大学 RGB-D image saliency target detection method
CN111311532B (en) * 2020-03-26 2022-11-11 深圳市商汤科技有限公司 Image processing method and device, electronic device and storage medium
CN111598841B (en) * 2020-04-23 2022-04-15 南开大学 Example significance detection method based on regularized dense connection feature pyramid
CN111260653B (en) * 2020-04-27 2020-08-25 腾讯科技(深圳)有限公司 Image segmentation method and device, storage medium and electronic equipment
CN111696021B (en) * 2020-06-10 2023-03-28 中国人民武装警察部队工程大学 Image self-adaptive steganalysis system and method based on significance detection
CN111915613B (en) * 2020-08-11 2023-06-13 华侨大学 Image instance segmentation method, device, equipment and storage medium
CN112465700B (en) * 2020-11-26 2022-04-26 北京航空航天大学 Image splicing positioning device and method based on depth clustering
CN112541912B (en) * 2020-12-23 2024-03-12 中国矿业大学 Rapid detection method and device for salient targets in mine sudden disaster scene
CN113034365B (en) * 2021-03-19 2023-09-22 西安电子科技大学 Multi-picture splicing method and system based on super pixels
CN114004775B (en) * 2021-11-30 2023-07-04 四川大学 Infrared and visible light image fusion method combining potential low-rank representation and convolutional neural network
CN114255351B (en) * 2022-02-28 2022-05-27 魔门塔(苏州)科技有限公司 Image processing method, device, medium, equipment and driving system
CN115965844B (en) * 2023-01-04 2023-08-18 哈尔滨工业大学 Multi-focus image fusion method based on visual saliency priori knowledge

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102222231A (en) * 2011-05-26 2011-10-19 厦门大学 Visual attention computational model based on guidance of dorsal pathway and processing method thereof
CN102567731A (en) * 2011-12-06 2012-07-11 北京航空航天大学 Extraction method for region of interest

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102222231A (en) * 2011-05-26 2011-10-19 厦门大学 Visual attention computational model based on guidance of dorsal pathway and processing method thereof
CN102567731A (en) * 2011-12-06 2012-07-11 北京航空航天大学 Extraction method for region of interest

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
Offline Signature Verification based on ASIFT;Youbao Tang et.al;《International Confer-ence on Biometrics (ICB)》;20131231(第1期);第1-6页 *
Offline Text-Independent Writer Identification Based on Scale Invariant Feature Transform;Xiangqian Wu et.al;《IEEE Transactions on Information Forensics and Security》;20140331;第9卷(第3期);第526-536页 *
Saliency Detection based on Graph-Structural Agglomerative Clustering;Youbao Tang et.al;《ACM International Conference on Mul-timedia (ACMMM)》;20151231(第1期);第1083-1086页 *
深度卷积神经网络的显著性检测;李岳云 等;《中国图像图形学报》;20160131;第21卷(第1期);第53-59页 *

Also Published As

Publication number Publication date
CN106157319A (en) 2016-11-23

Similar Documents

Publication Publication Date Title
CN106157319B (en) The conspicuousness detection method in region and Pixel-level fusion based on convolutional neural networks
CN110111335B (en) Urban traffic scene semantic segmentation method and system for adaptive countermeasure learning
CN107203781B (en) End-to-end weak supervision target detection method based on significance guidance
CN110852316B (en) Image tampering detection and positioning method adopting convolution network with dense structure
CN103984959B (en) A kind of image classification method based on data and task-driven
CN107480726A (en) A kind of Scene Semantics dividing method based on full convolution and shot and long term mnemon
CN109711413A (en) Image, semantic dividing method based on deep learning
CN105139395B (en) SAR image segmentation method based on small echo pond convolutional neural networks
CN109902806A (en) Method is determined based on the noise image object boundary frame of convolutional neural networks
CN106920243A (en) The ceramic material part method for sequence image segmentation of improved full convolutional neural networks
CN107403430A (en) A kind of RGBD image, semantics dividing method
CN107680106A (en) A kind of conspicuousness object detection method based on Faster R CNN
CN109002755B (en) Age estimation model construction method and estimation method based on face image
CN109583340A (en) A kind of video object detection method based on deep learning
CN107066916B (en) Scene semantic segmentation method based on deconvolution neural network
CN106981080A (en) Night unmanned vehicle scene depth method of estimation based on infrared image and radar data
CN105912999A (en) Human behavior identification method based on depth information
CN104966286A (en) 3D video saliency detection method
CN107247952B (en) Deep supervision-based visual saliency detection method for cyclic convolution neural network
CN110334589A (en) A kind of action identification method of the high timing 3D neural network based on empty convolution
CN105184772A (en) Adaptive color image segmentation method based on super pixels
CN109858487A (en) Weakly supervised semantic segmentation method based on watershed algorithm and image category label
CN110472634A (en) Change detecting method based on multiple dimensioned depth characteristic difference converged network
CN112488229A (en) Domain self-adaptive unsupervised target detection method based on feature separation and alignment
CN111882620A (en) Road drivable area segmentation method based on multi-scale information

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant