CN111652317A - Hyper-parameter image segmentation method based on Bayesian deep learning - Google Patents
Hyper-parameter image segmentation method based on Bayesian deep learning Download PDFInfo
- Publication number
- CN111652317A CN111652317A CN202010501892.9A CN202010501892A CN111652317A CN 111652317 A CN111652317 A CN 111652317A CN 202010501892 A CN202010501892 A CN 202010501892A CN 111652317 A CN111652317 A CN 111652317A
- Authority
- CN
- China
- Prior art keywords
- target
- image
- bayesian
- hyper
- edge
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2415—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
- G06F18/24155—Bayesian classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Multimedia (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Biophysics (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Probability & Statistics with Applications (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a Bayesian depth learning-based hyper-parameter image segmentation method, and aims to solve the technical problems of large extraction calculation amount and low precision of hyper-parameters in the conventional image segmentation. The method comprises the steps of selecting an image training set, carrying out a Gaussian process on image information, preprocessing a data set by adopting an L2 regular operator, obtaining image contour edge characteristics, constructing a target characteristic edge characteristic recognition training set, classifying the data set according to Bayesian theorem, setting image target edge segmentation labels based on semantic recognition, further extracting a target edge characteristic data set by adopting the Gaussian process, and calculating a target set edge characteristic Gaussian hyperparameter set. The invention has the beneficial effects that: the efficiency and the precision of target identification are improved.
Description
Technical Field
The invention relates to the technical field of machine learning, in particular to a super-parameter image segmentation method based on Bayes deep learning.
Background
In the field of computer vision, image segmentation refers to the task of assigning a label to each pixel in an image, which can also be considered as pixel classification. Unlike target detection using rectangular candidate frames, image segmentation needs to be accurate to pixel level positions, so it plays a very important role in tasks such as medical analysis, satellite image object detection, iris recognition, and automatic driving of cars.
Human beings distinguish targets more by experience, and deep learning is to identify targets by constructing a convolutional neural network and extracting target features by training. The recognition result of the traditional target recognition method is often some objects of a certain category defined in advance, such as human faces, vehicles and the like, while one image contains far more than some mutually independent objects, and also contains information of a plurality of objects and the attributes, spatial relationships, logical relationships and the like of the objects, and the information cannot be described only by some class labels, but needs to be described by using natural language. Any mathematical model is difficult to satisfy all target recognition, so that a plurality of condition classification recognition is formed, and the efficiency of cross-domain fusion recognition of deep learning is low.
Pixel-level image segmentation is a research hotspot in the field of artificial intelligence, and is a mathematical modeling problem relating to multiple subjects of image processing, pattern recognition, visual perception, psychological cognition and the like. It is very easy for human beings to identify targets by experience in the long-term evolution and learning process, but the targets are automatically identified from a complex background by a machine, complex mathematical modeling is needed for realization, and therefore, selection of a recognition model and hyper-parameter optimization are important. The hyper-parameters of deep learning have the characteristics of difficult selection and no regularity, unpredictable influence exists among different hyper-parameters, the debugging is very time-consuming, and a large amount of iterative computation is needed for the evaluation of each hyper-parameter combination. For such problems, classical optimization algorithms: such as particle swarm algorithm, simulated annealing algorithm, local search algorithm, etc. are no longer applicable. Some researchers propose a method of using a proxy model to reduce the evaluation cost of the problem by simulating the estimation value of the objective function. However, no matter the monte carlo algorithm or the adaptive simulation algorithm proposed in the reinforcement learning field is adopted, the learning process is always time-consuming, and is limited under a certain condition or in a certain field, the precision is difficult to guarantee, and the cross-boundary fusion is difficult to realize.
Disclosure of Invention
The invention provides a Bayesian deep learning-based hyper-parameter image segmentation method, which aims to solve the technical problems of large extraction calculation amount and low precision of hyper-parameters in the conventional image segmentation.
In order to solve the technical problems, the invention adopts the following technical scheme:
a super-parameter image segmentation method based on Bayesian deep learning is designed, and comprises the following steps:
step 1: data preprocessing, namely regularizing data elements in the image to generate an image segmentation class data set;
step 2: extracting target edge characteristics of the image by using a Gaussian mask;
and step 3: extracting a boundary frame and a target mask of the image through target edge characteristics by using Bayesian estimation;
and 4, step 4: putting the bounding box and the target mask into a feature dictionary for comparison, so as to obtain the category of each target in the image;
the construction method of the feature dictionary comprises the following steps:
(1) establishing a training set and a testing set of images;
(2) carrying out the operation in the step 1-3 on each image in the training set to obtain a boundary frame and a target mask of each image;
(3) collecting the boundary frame and the target mask in the step (2) to obtain a feature dictionary formed by the boundary frame and the target mask;
(4) and (4) inputting the test set into the feature dictionary in the step (3), checking the accuracy of the obtained feature dictionary, and if the accuracy does not meet the requirement, adjusting the parameters of the model for retraining until the accuracy of the feature dictionary meets the requirement.
Further, in step 1, the image segmentation class data set comprises N object segmentation class attributes and M data attributes of each object class, and when the probability of the N class attributes and the probability of the M data attributes are maximum, a Bayesian classification matcher is adopted to select an object and segment the image; the software used was python and the framework used tensorflow.
Further, in step 2, the specific steps of extracting the target edge feature are as follows:
the first step is as follows: assuming that the edge probability of the image pixel f (x, y) satisfies the gaussian distribution, the two-dimensional gaussian function is:
the second step is that: the gradient function is calculated for the x, y directions of the image:
the third step: convolving the image data set:
the fourth step: calculating the probability density distribution of the image target edge, namely the target edge characteristic:
further, in step 2, before extracting the object class label, calculating a prior probability of the object appearing:
wherein, CiIs a C type object set (C)1、C2、C3...Cn) Any one element of (1), NiRepresenting the number of occurrences of the target and N representing the total amount of the target set.
Further, in step 2, in the process of extracting the object class label, the conditional probability of the object occurrence is calculated:
wherein x isaRepresenting the abscissa, y, of the target pointaRepresenting the ordinate coordinates of the target point. P (x)a)、P(ya) Representing the probability of the edge feature of the object.
Further, in step 3, the specific steps of extracting the bounding box of the image and the target mask are as follows:
(1) obtaining a target area of the image and the classification weight of each pixel in the area by learning the extracted target boundary characteristics;
(2) after obtaining the target area of the image, combining the internal and external feature maps of each target area into two complete feature maps, and then synchronously performing image segmentation and image classification on two branch data sets D1 and D2;
(3) in image segmentation, classifying the internal and external feature maps of a target region by using a Bayesian classifier so as to distinguish the foreground and the background in an image and generate a mask;
(4) in image classification, the maximum value is taken according to pixel probability distribution in the two types of feature maps to obtain a new feature map, and then the class of objects in the target area is obtained by using a maximum likelihood estimation classifier.
Further, in step 4, the method for comparing the bounding box, the target mask and the feature dictionary comprises: firstly, similarity weights of a boundary box, a target mask and a feature dictionary are calculated by using an L2 regular operator, then a similarity Gaussian process is carried out, a target edge feature data set is extracted, and a semantic segmentation result can be obtained through Bayesian classification matching.
Further, before the output of the semantic segmentation result, a side Gaussian hyperparametric function is calculated, then the target matching degree is calculated according to the value of the score, and the better the hyperparametric set is, the higher the precision score of the obtained semantic segmentation is.
Further, the training set of Images included an Open Images V4 test set containing 190 ten thousand pictures and 1540 ten thousand bounding boxes for 600 categories on the pictures.
Compared with the prior art, the invention has the beneficial technical effects that:
the method mainly comprises the steps of utilizing a Bayesian formula, a python language and a tensoflow frame, carrying out pixel preprocessing on an image according to the principle that after an image edge feature Gaussian process, the edge feature projection is steep, carrying out the Gaussian process on the whole image, obtaining an edge feature data set, preprocessing the data set by utilizing an L2 criterion, utilizing a Bayesian estimation model for image target edge feature recognition aiming at the image target segmentation requirement based on semantic recognition, constructing an image target edge feature data dictionary based on semantic recognition by combining deep learning, applying the trained model to a complex target recognition system, and improving the efficiency and the precision of target recognition.
Drawings
FIG. 1 is a flow chart of a hyper-parametric image segmentation method based on Bayesian deep learning according to the present invention.
Detailed Description
The following examples are intended to illustrate the present invention in detail and should not be construed as limiting the scope of the present invention in any way.
The programs referred to or relied on in the following embodiments are all conventional programs or simple programs in the art, and those skilled in the art can make routine selection or adaptation according to specific application scenarios.
Example 1: a hyperparametric image segmentation method based on Bayesian deep learning is disclosed, and referring to FIG. 1, the overall steps are as follows: selecting an image training set, carrying out a Gaussian process on image information, preprocessing a data set by adopting an L2 regular operator to obtain image contour edge characteristics, constructing a target characteristic edge characteristic recognition training set, classifying the data set according to Bayesian theorem, setting an image target edge segmentation label based on semantic recognition, further adopting the Gaussian process to extract a target edge characteristic data set, calculating a target set edge characteristic Gaussian hyperparameter set, calculating a target edge posterior probability according to the data set, taking the maximum posterior probability as the target image segmentation and recognition probability based on the semantic, matching Bayesian scores, judging that the recognition is correct if the score is higher than 90, and carrying out Gaussian process training again by adopting deep learning and 0.618 coefficient adjustment Gaussian and hyperparameter until super-optimal parameters are obtained. The final result of the invention can be realized as follows: and inputting the image and the target label into the model for image target segmentation, wherein the target can be separated from a corresponding background. That is, for a trained model, given an image and target information to be queried, a corresponding target can be detected from the image.
The method for manufacturing the feature dictionary comprises the following steps:
(1) data pre-processing
And carrying out regularization processing on data elements in the image to generate an image segmentation class data set. The data set comprises N object segmentation class attributes and M data attributes representing each object class, and when the probability of the N class attributes and the probability of the M data attributes are the maximum, the model is adopted to segment the image according to the object edge contour. The specific implementation is realized by using Python language under an artificial intelligence tenserflow framework.
(2) Extracting target edge features
The method mainly extracts the characteristics of the target super-edge parameters through Gaussian transformation, and the main extracted characteristics are as follows:
1. the gray level of the edge pixel of the target image jumps;
2. the pixel boundary between different materials, textures, colors and brightness of the target generates jump;
3. the target contour line and the background have different reflection characteristics and can also form pixel value jump;
4. the object is illuminated to form shadows, which also form inter-pixel gray value transitions.
Calculating the attributes, including: adopting a Gaussian mask edge feature extraction algorithm for the image f (x, y), and being characterized in that: according to the characteristics of image target edge data jump and large probability density change, selecting a proper Gaussian mask, and obtaining a position with high probability density of a data set, namely an edge pixel point, on the basis of calculating the maximum value in the pixel point neighborhood.
The specific process is as follows:
the first step is as follows: assuming that the edge probability of an image pixel f (x, y) satisfies the gaussian distribution, the two-dimensional gaussian function is:
the second step is that: gradient function for x, y direction:
the third step: convolving the image data set:
the fourth step: calculating the probability density distribution of the image target edge, namely the target edge characteristic:
the method for calculating the Gaussian edge hyper-parameter feature extraction module is essentially to calculate a Gaussian kernel function, and then the Gaussian kernel function is used as a mask to be convolved with a target pixel, so that the target edge feature can be extracted, and the process is also a hyper-parameter estimation process.
After the edge characteristics of image pixels are obtained, characteristic distribution parameters are calculated by using a Gaussian process, then the minimum (optimal) loss function of a Gaussian characteristic function is calculated by using an L2 regularization operator, and meanwhile a penalty function is added in an algorithm to prevent the model from being over-fitted. The L2 regularization operator is:
where Loss is a Loss function, EinIs the training sample error without the regularization term and λ is the regularization parameter (penalty function). To make the model more optimal, the rule function is defined as follows:
i.e. the sum of the squares of all w (errors) does not exceed the parameter C (threshold), it can be ensured that the training sample error E is minimizedinThe loss function value is minimal.
(3) Extracting bounding boxes and target masks of an image
The Bayesian estimation algorithm is mainly adopted. The Bayes estimation model is used for recognizing the objective world by using a prior probability recognition algorithm, the basic idea is that on the basis of accumulating a large number of samples, the maximum probability estimation is carried out on a recognized object through the prior probability and the conditional probability, the maximum probability is a recognition result, and when the analysis sample is close to the total number, the probability of the event occurrence in the sample is close to the probability of the event occurrence in the total, so that the minimum error prediction can be realized. The Bayesian estimation core is hyper-parameter selection, in order to improve the local dominance involved in the image segmentation process, a hyper-parameter image data classification model based on Gaussian distribution is constructed and combined with deep learning, a 0.618 estimation method (golden segmentation method) is adopted to select hyper-parameters, the balance between hyper-parameter errors and weights is effectively adjusted, and the efficient pixel segmentation method based on semantic understanding is realized.
The Bayes classifier is a classification method based on statistical theory, and for a sample set C containing M class samples, C is { C ═ C1C2C3......CnThe classifier first calculates N-dimensional feature vector X ═ X1x2......xn]The maximum likelihood estimate of the labels belonging to each category is ranked and the maximum value is taken to calculate the category label C to which x belongsiThe Bayesian equation is as follows:
wherein, Pr (c)i| x) is the posterior probability, Pr (x | c)i) Conditional probability, Pr (c)i) As a priori probability, P (x)a)、P(ya) Representing the edge of an objectThe feature probability. The classification problem is resolved into the x attribute class CiMaximum value problem:
Ci=argmaxPr(x|ci)Pr(ci) (11)
experiments prove that the accuracy of the naive Bayes classifier and other classification classifiers is much higher.
According to different image gray levels, obvious protruding edges generally exist at the image boundary, and the image can be segmented by utilizing the characteristic. The kernel operator is a filter with Gaussian hyperparameter, has the characteristics of denoising, smoothing and strengthening edge characteristic attributes of an image, and the calculation process comprises four steps: firstly, smoothing the image by a Gaussian filter; secondly, calculating the amplitude and the direction of the gradient by using the finite difference of the first-order partial derivatives; thirdly, carrying out non-maximum suppression on the gradient amplitude; and fourthly, detecting and connecting edges by using a double-threshold algorithm.
The image segmentation process is as follows:
(1) obtaining a target area of the image and the classification weight of each pixel in the area by learning the extracted target boundary characteristics;
(2) after obtaining the target area of the image, combining the internal and external feature maps of each target area into two complete feature maps, and then synchronously performing image segmentation and image classification on two branch data sets D1 and D2;
(3) in image segmentation, classifying the internal and external feature maps of a target region by using a Bayesian classifier so as to distinguish the foreground and the background in an image and generate a mask;
(4) in image classification, the maximum value is taken according to pixel probability distribution in the two types of feature maps to obtain a new feature map, and then the class of objects in the target area is obtained by using a maximum likelihood estimation classifier.
After the image segmentation is completed, an image target contour edge data set needs to be established, which specifically comprises the following steps: the image target edge segmentation task comprises 5 basic subtasks of target candidate generation, candidate target edge feature extraction, candidate target Bayes classification, candidate target hyper-parameter correction, candidate target edge feature dictionary construction and the like. The alternative target data sets include the OpenImages V4 test set, which contains 190 ten thousand pictures and 1540 ten thousand bounding boxes for 600 categories on the pictures. Preprocessing image data by adopting a pixel L2 regular operator to form an image classification set and a classification image pixel subset, calculating a target edge feature nuclear mask by adopting a multi-dimensional Gaussian distribution probability model, acquiring a target edge feature data set by convolution, and acquiring the prior probability of the target edge feature classification set by adopting Bayesian evidence learning.
(4) Implementation procedure
The device mainly comprises three parts: (1) preprocessing image target pixels: and (3) setting a threshold function to generate an initial point set by adopting an L2 regular algorithm: x, Y ═ (X1, Y1), (X2, Y2),. · (xt, yt); (2) constructing a data set D { (x1, y1) · (xt, yt) } by adopting a Gaussian kernel model; (3) and entering Bayes maximum posterior probability estimation. The specific operation is as follows:
A. setting labels according to target classification;
B. counting image target classification;
C. calculating the prior probability of each type of target: pr (C)i);
D. According to the target classification, selecting all data in a corresponding data set D, constructing a Gaussian process model, and extracting target edge points Xi and Yi sets;
E. calculating a Gaussian distribution hyperparametric function (mu)i,σi);
F. Further, an acquisition function (μ) is usedi,σi) Calculating the value of the next evaluation point xi, i is 1-t, xi is argmaxu (x | D), and calculating response yi;
G. adding new data points to a set D, D ← D { xi, yi }, i ← i + 1;
H. calculating posterior probability using Bayesian posterior estimation formula
I. Calculating posterior probability distribution: max (Pr (C) ═ R1|X),Pr(C2|X)......Pr(CtX));
J. And (4) checking and calculating results: matching the calculated posterior probability with the edge characteristics of the target, wherein if the matching score is higher, the identified month is close to the actual target, and the score is set to be more than 90 by the model;
K. correcting, if the score is less than 90, adopting deep learning to measure Gaussian super parameter mui,σiCorrecting with a correction coefficient of 0.618;
and L, placing the target marginal probability parameters with the scores larger than 90 into a data dictionary to form a hyper-parameter data dictionary set.
(5) Training and testing
190 ten thousand pictures in the Open Images V4 detection set are divided into a training set and a test set, and a supervised learning training model is used for inputting the test set to verify whether the trained model meets the requirements. At this point, the feature dictionary establishment is complete.
(6) Image detection
Inputting the image to be detected into the detection model, extracting the boundary frame and the target mask of the image to be detected, and then putting the boundary frame and the target mask into the feature dictionary for comparison to obtain the category of each target in the image. The method comprises the following steps: firstly, similarity weights of a boundary box, a target mask and the feature dictionary are calculated by using an L2 regular operator, then a target edge feature data set is extracted by the similarity Gaussian process, and a semantic segmentation result can be obtained through Bayesian classification matching. Before the output of the semantic segmentation result, a side Gaussian hyperparametric function is calculated, then the target matching degree is calculated according to the size of the score, and the better the hyperparametric set is, the higher the precision score of the obtained semantic segmentation is.
Through the steps, the image segmentation method based on the Bayes deep learning hyper-parameter optimization can be completed, and the target edge segmentation hyper-parameter dictionary is obtained on the basis of the training of the image set. With the hyper-parameter dictionary, the recognition and the segmentation of the target can be realized, namely: after training, the system can realize the separation and recognition of the target and the background through model calculation in the environment of inputting images and voice. The method can effectively overcome the defects of long time consumption, large performance fluctuation, large occupied resources and the like of the traditional deep learning optimization algorithm, and the model can be applied to plug-ins in image semantic recognition software of the smart phone after being trained.
While the present invention has been described in detail with reference to the drawings and the embodiments, those skilled in the art will understand that various specific parameters in the above embodiments can be changed without departing from the spirit of the present invention, and a plurality of specific embodiments are formed, which are common variation ranges of the present invention, and will not be described in detail herein.
Claims (9)
1. A hyper-parametric image segmentation method based on Bayesian deep learning is characterized by comprising the following steps:
step 1: data preprocessing, namely regularizing data elements in the image to generate an image segmentation class data set;
step 2: extracting target edge characteristics of the image by using a Gaussian mask;
and step 3: extracting a boundary frame and a target mask of the image through the target edge characteristics by using Bayesian estimation;
and 4, step 4: putting the bounding box and the target mask into a feature dictionary for comparison, so as to obtain the category of each target in the image;
the construction method of the feature dictionary comprises the following steps:
(1) establishing a training set and a testing set of images;
(2) performing the operation in the step 1-3 on each image in the training set to obtain a boundary frame and a target mask of the image;
(3) collecting the boundary frame and the target mask in the step (2) to obtain a feature dictionary formed by the boundary frame and the target mask;
(4) and (4) inputting the test set into the feature dictionary in the step (3), checking the accuracy of the obtained feature dictionary, and if the accuracy does not meet the requirement, adjusting the parameters of the model for retraining until the accuracy of the feature dictionary meets the requirement.
2. The Bayesian deep learning-based hyper-parametric image segmentation method as recited in claim 1, wherein in step 1, the image segmentation class data set comprises N object segmentation class attributes and M data attributes of each object class, and when the probability of the N class attributes and the probability of the M data attributes are the maximum, a Bayesian classification matcher is adopted to select an object and segment the image.
3. The hyperparametric image segmentation method based on Bayesian deep learning as recited in claim 1, wherein in the step 2, the specific steps of extracting the target edge feature are as follows:
the first step is as follows: assuming that the edge probability of the image pixel f (x, y) satisfies the gaussian distribution, the two-dimensional gaussian function is:
the second step is that: the gradient function is calculated for the x, y directions of the image:
the third step: convolving the image data set:
the fourth step: calculating the probability density distribution of the image target edge, namely the target edge characteristic:
4. the Bayesian deep learning-based hyper-parametric image segmentation method as recited in claim 3, wherein before obtaining the class label of the target in step 3, the prior probability of the target occurrence needs to be calculated:
wherein, CiIs a C type object set (C)1、C2、C3...Cn) Any one element of (1), NiRepresenting the number of occurrences of the target and N representing the total amount of the target set.
5. The Bayesian deep learning-based hyper-parametric image segmentation method as recited in claim 3, wherein in step 3, in the process of obtaining the class label of the target, a conditional probability of the target occurrence is calculated:
wherein x isaRepresenting the abscissa, y, of the target pointaRepresenting the ordinate coordinates of the target point. P (x)a)、P(ya) Representing the probability of the edge feature of the object.
6. The Bayesian deep learning-based hyper-parametric image segmentation method as recited in claim 1, wherein in step 3, the specific steps of extracting the bounding box and the target mask of the image are as follows:
(1) obtaining a target area of the image and the classification weight of each pixel in the area by learning the extracted target boundary characteristics;
(2) after obtaining the target area of the image, combining the internal and external feature maps of each target area into two complete feature maps, and then synchronously performing image segmentation and image classification on two branch data sets D1 and D2;
(3) in image segmentation, classifying the internal and external feature maps of the target region by using a Bayesian classifier so as to distinguish the foreground and the background in the image and generate a mask;
(4) in image classification, the maximum value is taken according to pixel probability distribution in the two types of feature maps to obtain a new feature map, and then the class of objects in the target area is obtained by using a maximum likelihood estimation classifier.
7. The Bayesian deep learning-based hyper-parametric image segmentation method as recited in claim 1, wherein in step 4, the method for comparing the bounding box and the target mask with the feature dictionary comprises: firstly, similarity weights of a boundary box, a target mask and the feature dictionary are calculated by using an L2 regular operator, then a target edge feature data set is extracted by the similarity Gaussian process, and a semantic segmentation result can be obtained through Bayesian classification matching.
8. The Bayesian deep learning-based hyper-parameter image segmentation method as recited in claim 7, wherein before outputting the semantic segmentation result, a side Gaussian hyper-parameter function is calculated, and then a target matching degree is calculated according to the value, and the better the hyper-parameter set is, the higher the precision value of the semantic segmentation is.
9. The Bayesian deep learning-based hyper-parametric image segmentation method as recited in claim 1, wherein the training set of Images comprises an Open Images V4 detection set containing 190 ten thousand pictures and 1540 ten thousand bounding boxes for 600 classes on the pictures.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010501892.9A CN111652317B (en) | 2020-06-04 | 2020-06-04 | Super-parameter image segmentation method based on Bayes deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010501892.9A CN111652317B (en) | 2020-06-04 | 2020-06-04 | Super-parameter image segmentation method based on Bayes deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111652317A true CN111652317A (en) | 2020-09-11 |
CN111652317B CN111652317B (en) | 2023-08-25 |
Family
ID=72347345
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010501892.9A Active CN111652317B (en) | 2020-06-04 | 2020-06-04 | Super-parameter image segmentation method based on Bayes deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111652317B (en) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112464948A (en) * | 2020-11-11 | 2021-03-09 | 常州码库数据科技有限公司 | Natural scene target contour extraction method and system based on bionics |
CN112906704A (en) * | 2021-03-09 | 2021-06-04 | 深圳海翼智新科技有限公司 | Method and apparatus for cross-domain target detection |
CN113111928A (en) * | 2021-04-01 | 2021-07-13 | 中国地质大学(北京) | Semi-supervised learning mineral resource quantitative prediction method based on geoscience database |
CN113158960A (en) * | 2021-05-06 | 2021-07-23 | 吴国军 | Medical image recognition model construction and recognition method and device |
CN113269782A (en) * | 2021-04-21 | 2021-08-17 | 青岛小鸟看看科技有限公司 | Data generation method and device and electronic equipment |
CN113392933A (en) * | 2021-07-06 | 2021-09-14 | 湖南大学 | Self-adaptive cross-domain target detection method based on uncertainty guidance |
CN113450268A (en) * | 2021-05-24 | 2021-09-28 | 南京中医药大学 | Image noise reduction method based on posterior probability |
CN113468939A (en) * | 2020-11-30 | 2021-10-01 | 电子科技大学 | SAR target recognition method based on supervised minimization deep learning model |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030147558A1 (en) * | 2002-02-07 | 2003-08-07 | Loui Alexander C. | Method for image region classification using unsupervised and supervised learning |
WO2010129711A1 (en) * | 2009-05-05 | 2010-11-11 | The Trustees Of Columbia University In The City Of New York | Devices, systems, and methods for evaluating vision and diagnosing and compensating impairment of vision |
US7979363B1 (en) * | 2008-03-06 | 2011-07-12 | Thomas Cecil Minter | Priori probability and probability of error estimation for adaptive bayes pattern recognition |
CN110245587A (en) * | 2019-05-29 | 2019-09-17 | 西安交通大学 | A kind of remote sensing image object detection method based on Bayes's transfer learning |
-
2020
- 2020-06-04 CN CN202010501892.9A patent/CN111652317B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20030147558A1 (en) * | 2002-02-07 | 2003-08-07 | Loui Alexander C. | Method for image region classification using unsupervised and supervised learning |
US7979363B1 (en) * | 2008-03-06 | 2011-07-12 | Thomas Cecil Minter | Priori probability and probability of error estimation for adaptive bayes pattern recognition |
WO2010129711A1 (en) * | 2009-05-05 | 2010-11-11 | The Trustees Of Columbia University In The City Of New York | Devices, systems, and methods for evaluating vision and diagnosing and compensating impairment of vision |
CN110245587A (en) * | 2019-05-29 | 2019-09-17 | 西安交通大学 | A kind of remote sensing image object detection method based on Bayes's transfer learning |
Non-Patent Citations (2)
Title |
---|
包晓敏;汪亚明;: "基于最小风险贝叶斯决策的织物图像分割" * |
王生生;王顺;张航;温长吉;: "基于轻量和积网络及无人机遥感图像的大豆田杂草识别" * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112464948A (en) * | 2020-11-11 | 2021-03-09 | 常州码库数据科技有限公司 | Natural scene target contour extraction method and system based on bionics |
CN113468939A (en) * | 2020-11-30 | 2021-10-01 | 电子科技大学 | SAR target recognition method based on supervised minimization deep learning model |
CN112906704A (en) * | 2021-03-09 | 2021-06-04 | 深圳海翼智新科技有限公司 | Method and apparatus for cross-domain target detection |
CN113111928A (en) * | 2021-04-01 | 2021-07-13 | 中国地质大学(北京) | Semi-supervised learning mineral resource quantitative prediction method based on geoscience database |
CN113111928B (en) * | 2021-04-01 | 2023-12-29 | 中国地质大学(北京) | Semi-supervised learning mineral resource quantitative prediction method based on geometrics database |
CN113269782A (en) * | 2021-04-21 | 2021-08-17 | 青岛小鸟看看科技有限公司 | Data generation method and device and electronic equipment |
CN113158960A (en) * | 2021-05-06 | 2021-07-23 | 吴国军 | Medical image recognition model construction and recognition method and device |
CN113450268A (en) * | 2021-05-24 | 2021-09-28 | 南京中医药大学 | Image noise reduction method based on posterior probability |
CN113392933A (en) * | 2021-07-06 | 2021-09-14 | 湖南大学 | Self-adaptive cross-domain target detection method based on uncertainty guidance |
Also Published As
Publication number | Publication date |
---|---|
CN111652317B (en) | 2023-08-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111652317B (en) | Super-parameter image segmentation method based on Bayes deep learning | |
US10380759B2 (en) | Posture estimating apparatus, posture estimating method and storing medium | |
TWI742382B (en) | Neural network system for vehicle parts recognition executed by computer, method for vehicle part recognition through neural network system, device and computing equipment for vehicle part recognition | |
CN110135503B (en) | Deep learning identification method for parts of assembly robot | |
CN107610087B (en) | Tongue coating automatic segmentation method based on deep learning | |
US10216979B2 (en) | Image processing apparatus, image processing method, and storage medium to detect parts of an object | |
KR100647322B1 (en) | Apparatus and method of generating shape model of object and apparatus and method of automatically searching feature points of object employing the same | |
CN106991388B (en) | Key point positioning method | |
Jia et al. | Visual tracking via coarse and fine structural local sparse appearance models | |
CN106570480B (en) | A kind of human action classification method based on gesture recognition | |
CN110909618B (en) | Method and device for identifying identity of pet | |
CN109711283A (en) | A kind of joint doubledictionary and error matrix block Expression Recognition algorithm | |
CN111967464B (en) | Weak supervision target positioning method based on deep learning | |
CN114758288A (en) | Power distribution network engineering safety control detection method and device | |
CN108734200B (en) | Human target visual detection method and device based on BING (building information network) features | |
CN110853070A (en) | Underwater sea cucumber image segmentation method based on significance and Grabcut | |
WO2018100668A1 (en) | Image processing device, image processing method, and image processing program | |
CN113221956B (en) | Target identification method and device based on improved multi-scale depth model | |
Yuan et al. | Neighborloss: a loss function considering spatial correlation for semantic segmentation of remote sensing image | |
Ravi et al. | Sign language recognition with multi feature fusion and ANN classifier | |
US11367206B2 (en) | Edge-guided ranking loss for monocular depth prediction | |
Pramunendar et al. | A Robust Image Enhancement Techniques for Underwater Fish Classification in Marine Environment. | |
Yang et al. | An improved algorithm for the detection of fastening targets based on machine vision | |
Goyal et al. | Fuzzy similarity measure based spectral clustering framework for noisy image segmentation | |
Jeevanantham et al. | Deep Learning Based Plant Diseases Monitoring and Detection System |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |