CN102750702A - Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model - Google Patents
Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model Download PDFInfo
- Publication number
- CN102750702A CN102750702A CN2012102067011A CN201210206701A CN102750702A CN 102750702 A CN102750702 A CN 102750702A CN 2012102067011 A CN2012102067011 A CN 2012102067011A CN 201210206701 A CN201210206701 A CN 201210206701A CN 102750702 A CN102750702 A CN 102750702A
- Authority
- CN
- China
- Prior art keywords
- depth
- pixel
- infrared image
- yardstick
- image block
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Landscapes
- Image Analysis (AREA)
Abstract
The invention relates to a monocular infrared image depth estimation method based on an optimized BP (Back Propagation) neural network model. The method comprises the following steps of: acquiring a monocular infrared image and a depth map to which the monocular infrared image corresponds; setting at least three feature regions with different scales for pixel points in the monocular infrared image; calculating feature vectors of the feature regions to which the pixel points in the monocular infrared image correspond; screening all the feature vectors by successively using stepwise linear regression and independent component analysis methods to obtain feature vectors conforming to depth information of the infrared image; constructing a depth training sample set by using the obtained feature vectors and the depth map to which the infrared image corresponds, and performing nonlinear fitting on the feature vectors in the set and depth values of the depth map by using a BP neuron network, and optimizing the BP neuron network through a genetic algorithm, and then constructing a depth model; and analyzing the monocular infrared image through the depth model to obtain a depth estimated value. By using the monocular infrared image depth estimation method based on the optimized BP neural network model, the depth information of the infrared image can be relatively accurately estimated.
Description
Technical field
The present invention relates to infrared image estimation of Depth technical field, particularly relate to a kind of monocular infrared image depth estimation method based on the genetic algorithm optimization BP neural network model.
Background technology
The estimation of Depth of image promptly is from image, to obtain depth distance information, is the problem of a depth perception in essence.By depth perception make up spatial positional information characterized be from observer to the scene in the relative distance on detected surface.Recover the now existing more satisfactory algorithm of depth distance information in the coloured image, but for infrared image, because of its reflection be the Temperature Distribution of scene, defectives such as low signal-to-noise ratio, low contrast are arranged, the degree of depth algorithm that recovers this image still belongs to blank.If can recover the depth information of infrared image, will greatly improve the understanding effect of human eye so to this image.
The picture depth method of estimation mainly is to launch to binocular degree of depth clue with based on the estimation of Depth of image sequence at present, and these two kinds of methods all depend on the feature difference between image.And estimate for monocular depth; Comparatively classical in the middle of traditional in early days algorithm is " by shadow recovery shape (shape from shading) "; This algorithm is a theoretical foundation how much with space multistory; Changing to light and shade that body surface produced according to light source irradiation is that the shade of image recovers the object degree of depth, but because this algorithm needs priori (like reflection model and light source direction etc.) to make the application limit increase.Afterwards, some researchers find the importance of experience gradually, begin to utilize the method for machine learning to go to address this problem.The team of the Andrew Ng of Stanford University carries out estimation of Depth through the model that utilizes the Markov field training to single image, has reached good effect; Simple classification in the scene is then manually demarcated by the Aloysha Efros team of CMU before training; Such as sky, trees, ground and perpendicular line or the like; Utilize these classifications of lot of data study then, thereby and finally new images is classified and recover depth information through the structure Bayesian model.Though it is this method is comparatively suitable to the simple picture of a series of scenes, and produces a desired effect, inaccurate often for the classification that does not have study in the scene.
Summary of the invention
Technical matters to be solved by this invention provides a kind of based on the monocular infrared image depth estimation method of optimizing the BP neural network model, and this method can be estimated the depth information of infrared image comparatively exactly.
The technical solution adopted for the present invention to solve the technical problems is: provide a kind of based on the monocular infrared image depth estimation method of optimizing the BP neural network model, may further comprise the steps:
(1) (x y) reaches and the pairing depth map of said monocular infrared image I (x, y) to obtain any monocular infrared image I;
(2) be said monocular infrared image I (x; Y) each pixel in is set the characteristic area at least three different scales; Three ascending being respectively of different scale: first yardstick, second yardstick and the 3rd yardstick; Wherein, the characteristic area of each yardstick comprises that at least the image block that is positioned at the center reaches and the adjacent image block in said image block upper and lower, left and right, and the image block that be positioned at center of i pixel on first yardstick is i pixel itself; I pixel comprises all images piece on first yardstick at the image block that is positioned at the center on second yardstick; I pixel comprises all images piece on second yardstick at the image block that is positioned at the center on the 3rd yardstick; By that analogy;
(3) calculate monocular infrared image I (x; Y) proper vector of the pairing characteristic area of each pixel in; The characteristic component of the proper vector of i pixel comprises at least: the gray-scale value of all images piece of i pixel on first yardstick; The texture energy of each image block of i pixel on other yardsticks except that first yardstick; The gradient energy of the different directions of each image block of i pixel on other yardsticks except that first yardstick and the average and the variance of all gradient energy, and the sharpness of each image block of i pixel on other yardsticks except that first yardstick;
(4) utilize the method for progressively linear regression analysis and independent component analysis to screen successively to all proper vectors of obtaining, obtain meeting the proper vector of infrared image depth information;
(5) utilize the middle screening of step (4) to obtain proper vector and said monocular infrared image I (x; Y) pairing depth map makes up the set of degree of depth training sample; The proper vector in the set of degree of depth training sample and the depth value of depth map are carried out nonlinear fitting with the BP neural network; And be optimized, and then make up depth model through the initial weight and the threshold value of genetic algorithm to the BP neural network;
The new monocular infrared image that (6) will collect obtains the estimation of Depth value through the depth model analysis that structure obtains.
Texture energy in the said step (3) calculates through the Louth mask, and concrete steps are: adopt N basic two-dimentional Louth mask, be designated as M
1..., M
N, with said monocular infrared image I (x y) does convolution with each two-dimentional Louth mask, then monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after value be: T
k(x, y)=I (x, y) * M
k, k=1 ..., N, then i m the image block N of pixel on the j yardstick
i(m) monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after the texture energy that obtains do
The gradient energy of the different directions of each image block of i pixel on second yardstick and the 3rd yardstick and the average and the variance of all gradient energy calculate through following steps in the said step (3); Concrete steps are: (x y) tries to achieve x axial gradient figure I respectively on x direction of principal axis and y direction of principal axis to monocular infrared image I
Gradx(x is y) with y axial gradient figure I
Grady(x, y), then (x is y) at angle θ for monocular infrared image I
lGrad G on the direction
l(x, y)=I
Gradx(x, y) * cos (θ
l)+I
Grady(x, y) * sin (θ
l), wherein,
L=0 ..., 7, calculate the gradient energy of the different directions of each image block of each pixel on other yardsticks except that first yardstick subsequently, wherein, i m the image block N of pixel on the j yardstick
i(m) at angle θ
lGradient energy on the direction does
I m the image block N of pixel on the j yardstick then
i(m) at angle θ
lThe average of the gradient energy on the direction
I m the image block N of pixel on the j yardstick
i(m) at angle θ
lGradient energy variance on the direction
Wherein, size is image block N
iThe number of the pixel that (m) is comprised,
The sharpness of each image block of i pixel on second yardstick and the 3rd yardstick calculates through following steps in the said step (3), and concrete steps are: i m the image block N of pixel on the j yardstick
i(m) sharpness
Wherein, (x y) is image block N to I
i(m) contain the gray scale of pixel,
Be image block N
i(m) average of the gray scale of contained pixel,
Size is image block N
iThe number of the pixel that (m) is comprised.
In the said step (4) progressively linear regression analysis comprise following substep:
(411) related coefficient of each characteristic component and depth value in the calculated characteristics vector is according to the descending ordering that obtains each characteristic component to the effect of depth degree of the absolute value of related coefficient;
(412) the absolute value biggest characteristic component from related coefficient begins progressively to introduce regression equation; And do regression equation significance test; If do not think significantly that selected whole characteristic component all is not the principal element that influences depth value, if significantly again from depth value is influenced the descending regression equation of introducing one by one successively;
(413) new characteristic component of every introducing all need carry out significance test to each contained in regression equation characteristic component; With that in new regression equation significantly and to the minimum characteristic component rejection of depth value influence, repeat this step each characteristic component in regression equation all significantly till;
(414) introduce a characteristic component that in the characteristic component of not introducing depth value is had the greatest impact again, repeating step (413) and step (414) up to rejecting selected characteristic component, also can't be introduced till the new characteristic component.
What independent component analysis was used in the said step (4) is quick ICA algorithm, to analyzing through the proper vector of progressively linear regression analysis, makes between each component independently as much as possible through quick ICA algorithm, comprises following substep:
(421) specifying the proper vector through M pixel after the linear regression analysis progressively is observation data X, and observation data X is carried out centralization, and making it average is 0;
(422), be about to become the albefaction vector Z, Z=W after observation data X projects to new subspace with the observation data X albefaction of centralization
0X, wherein, W
0Be albefaction matrix, W
0=Λ
-1/2U
T, Λ is the eigenvalue matrix of the covariance matrix of observation data X, U is the eigenvectors matrix of the covariance matrix of observation data X;
(423) upgrade W
*Make W
*=E{Zg (W
TZ) }-E{g ' (W
TZ}W, wherein g () is a nonlinear function, and then to W
*Standardization: W=W
*/ || W
*||, if do not restrain and then do not repeat this step, wherein, selecting a mould is 1 the initial random weight vector initial value as W.
Initial weight and the threshold value of utilizing genetic algorithm to be optimized the BP neural network in the said step (5) also comprise following substep:
(511) make up initial BP neural network according to training sample, and confirm the number of input layer, hidden layer and the output layer of this network;
(512) carry out the initialization of genetic algorithm population; Wherein each individuality of population all is to be made up of the threshold value of the weights between the weights between input layer and the hidden layer, hidden layer and the output layer, hidden layer and threshold value four parts of output layer, and all weights and threshold value are listed as into a real number vector;
(513) maximum iteration time is set as stopping criterion for iteration, iterative process each time comprises selection, intersection and mutation operation, when reaching stopping criterion for iteration, output the last reign of a dynasty population optimum individual be the approximate optimal solution of initial weight and threshold value.
Beneficial effect
Owing to adopted above-mentioned technical scheme; The present invention compared with prior art; Have following advantage and good effect: the present invention utilizes infrared image " spatial context " and " multiple dimensioned " information extraction depth characteristic vector; Through the method for progressively linear regression analysis and independent component analysis the characteristic of extracting is screened successively, help finding the depth characteristic that is fit to infrared image more, and make up degree of depth training set with this; Adopt the back propagation learning theory that nonlinear fitting is carried out in the training set, and be optimized, improved the training speed of experiment and the accuracy rate of match through genetic algorithm.
Description of drawings
Fig. 1 is a process flow diagram of the present invention.
Embodiment
Below in conjunction with specific embodiment, further set forth the present invention.Should be understood that these embodiment only to be used to the present invention is described and be not used in the restriction scope of the present invention.Should be understood that in addition those skilled in the art can do various changes or modification to the present invention after the content of having read the present invention's instruction, these equivalent form of values fall within the application's appended claims institute restricted portion equally.
As shown in Figure 1, present embodiment discloses a kind of monocular infrared image depth estimation method of the BP neural network model based on genetic algorithm optimization, the steps include:
Step 1, obtain any monocular infrared image I (x, y) and with this monocular infrared image I (x, y) pairing depth map;
Step 2, be monocular infrared image I (x; Y) each pixel in is set the characteristic area on three different scales; Three ascending being respectively of different scale: first yardstick, second yardstick and the 3rd yardstick; Wherein, The characteristic area of each yardstick comprises image block and the image block adjacent with these image block upper and lower, left and right that is positioned at the center; The image block that be positioned at center of i pixel on first yardstick is i pixel itself, and this i pixel comprises all images piece on first yardstick at the image block that is positioned at the center on second yardstick, and this i pixel comprises all images piece on second yardstick at the image block that is positioned at the center on the 3rd yardstick;
Step 3, calculating monocular infrared image I (x; Y) proper vector of the pairing characteristic area of each pixel in; The characteristic component of the proper vector of i pixel comprises at least: the gray-scale value of all images piece of i pixel on first yardstick; The texture energy of each image block of i pixel on second yardstick and the 3rd yardstick; I pixel gradient energy and the average and variance of all gradient energy on 8 directions of each image block on second yardstick and the 3rd yardstick, and the sharpness of each image block of i pixel on second yardstick and the 3rd yardstick;
The texture energy of each image block of i pixel on second yardstick and the 3rd yardstick calculates through following steps:
Adopt 9 basic two-dimentional Louth masks, be designated as M
1..., M
9, with monocular infrared image I (x y) does convolution with each two-dimentional Louth mask, then monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after value be: T
k(x, y)=I (x, y) * M
k, k=1 ..., 9, i m the image block N of pixel on the j yardstick then
i(m) monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after the texture energy that obtains do
The gradient energy of the different directions of each image block of i pixel on second yardstick and the 3rd yardstick and the average and the variance of all gradient energy calculate through following steps:
(x y) tries to achieve x axial gradient figure I respectively on x direction of principal axis and y direction of principal axis to monocular infrared image I
Gradx(x is y) with y axial gradient figure I
Grady(x, y), then (x is y) at angle θ for monocular infrared image I
lGrad G on the direction
l(x, y)=I
Gradx(x, y) * cos (θ
l)+I
Grady(x, y) * sin (θ
l), wherein,
L=0 ..., 7, calculate the gradient energy of the different directions of each image block of each pixel on other yardsticks except that first yardstick subsequently, wherein, i m the image block N of pixel on the j yardstick
i(m) at angle θ
lGradient energy on the direction does
I m the image block N of pixel on the j yardstick then
i(m) at angle θ
lThe average of the gradient energy on the direction
I m the image block N of pixel on the j yardstick
i(m) at angle θ
lGradient energy variance on the direction
Wherein, size is image block N
iThe number of the pixel that (m) is comprised,
The sharpness of each image block of i pixel on second yardstick and the 3rd yardstick calculates through following steps:
I m the image block N of pixel on the j yardstick
i(m) sharpness
Wherein, (x y) is image block N to I
i(m) contain the gray scale of pixel,
Be image block N
i(m) average of the gray scale of contained pixel,
Step 4, utilize the method for progressively linear regression analysis and independent component analysis to screen successively, comparatively met the proper vector of infrared image depth information all proper vectors of having obtained in the step 3.
In the step 4 progressively the concrete steps of linear regression analysis be:
The related coefficient of each characteristic component and depth value in step 4.1.1, the calculated characteristics vector is according to the descending ordering that obtains each characteristic component to the effect of depth degree of the absolute value of related coefficient;
Step 4.1.2, begin progressively to introduce regression equation from the absolute value biggest characteristic component of related coefficient; And do regression equation significance test; If can not think significantly that selected whole characteristic component all is not the principal element that influences depth value, if significantly again from depth value is influenced the descending regression equation of introducing one by one successively;
Step 4.1.3, new characteristic component of every introducing all need carry out significance test to each contained in regression equation characteristic component; With that in new regression equation significantly and to the minimum characteristic component rejection of depth value influence, repeat this step each characteristic component in regression equation all significantly till;
Step 4.1.4, introduce a characteristic component that in the characteristic component of not introducing depth value is had the greatest impact again, repeating step 4.1.3 and step 4.1.4 up to rejecting selected characteristic component, also can't introduce till the new characteristic component.
What independent component analysis was used in the step 4 is quick ICA algorithm, to analyzing through the proper vector of progressively linear regression analysis, makes between each component independent as much as possible through quick ICA algorithm.In the present embodiment, the ICA algorithm is maximum as searching direction with negentropy fast, and its concrete steps are following:
Step 4.2.1, to specify the proper vector through M pixel after the linear regression analysis progressively be observation data X, and observation data X is carried out centralization, and making it average is 0;
Step 4.2.2, with the observation data X albefaction of centralization, be about to become the albefaction vector Z, Z=W after observation data X will project to new subspace
0X, wherein, W
0Be albefaction matrix, W
0=Λ
-1/2U
T, Λ is the eigenvalue matrix of the covariance matrix of observation data X, U is the eigenvectors matrix of the covariance matrix of observation data X;
Step 4.2.3, renewal W
*Be to find the solution the separation matrix that draws, the separation matrix W that finally draws
*And observation data X multiplies each other and can obtain between each component as far as possible independently source data and make W
*=E{Zg (W
TZ) }-E{g ' (W
TZ) } W, wherein E () is the expectation computing, wherein g () is a nonlinear function, and then to W
*Standardization: W=W
*/ || W
*||, if do not restrain and then do not repeat this step, wherein, selecting a mould is 1 the initial random weight vector initial value as W.
The depth map of step 5, the proper vector of utilizing the infrared image that is drawn by step 4 and infrared image makes up degree of depth training sample set { f
i, depth
i, wherein, f
iBe the proper vector of i pixel, i=1 ..., l, f
i∈ χ, depth
iBe i the pairing depth value of pixel.Utilize back propagation learning The Theory Construction depth model then, and be optimized, and then new infrared image is carried out estimation of Depth through genetic algorithm.
The concrete steps of BP neural metwork training are following in the step 5:
Step 5.1.1, according to proper vector f
iWith depth value depth
iConfirm the input layer of BP neural network and the node number of output layer.Then through counting scope with reference to the roughly definite hidden layer node of formula
; Wherein n is the input layer number; M is an output layer node number; A is the constant between 0 ~ 10, and l is the hidden layer node number.Confirm the hidden layer node number that square error is minimum through trial method again;
The weights and the threshold value of step 5.1.2, initialization BP neural network, and given maximum iteration time, learning rate parameter η and neuron excitation function
wherein a be constant;
The output H of step 5.1.3, calculating hidden layer:
Wherein l is the hidden layer node number, ω
IjFor connecting the weights between input layer and the hidden layer, x
iBe the proper vector f that mentions among the step 5.1.1
i, a
jBe the threshold value of each hidden layer node, the neuron excitation function of f (.) for mentioning among the step 5.1.2;
Step 5.1.4, calculating prediction output O:
Wherein m is an output layer node number, ω
JkFor connecting the weights between hidden layer and the output layer, b
kThreshold value for each output layer node;
Step 5.1.5, according to prediction output and desired output error of calculation e, and upgrade the weights and the threshold value of network according to error e:
e
k=Y
k-O
k,k=1,...,m
ω
jk=ω
jk+ηH
je
k,j=1,...,l;k=1,...,m
b
k=b
k+e
k,k=1,...,m
Wherein n is the input layer number, Y
kBe desired output, i.e. depth value in the training set;
Step 5.1.6, circulation step 5.1.3 finish after reaching maximum iteration time to step 5.1.5;
Step 5.1.7, to obtaining the proper vector of each pixel after feature extraction and the screening of new infrared image through step 1 and step 2; Each component of proper vector is corresponded to the input layer of the BP neural network of having trained respectively; Draw corresponding depth value through prediction, just accomplished estimation new infrared image depth value.
BP neural network initial weight and threshold value are following through the concrete steps that genetic algorithm is optimized in the step 5:
Step 5.2.1, carry out the initialization of genetic algorithm population, wherein each individuality of population all is to be made up of the threshold value of the weights between the weights between input layer and the hidden layer, hidden layer and the output layer, hidden layer and threshold value four parts of output layer.All weights and threshold value are listed as into a real number vector;
Step 5.2.2, carry out the selection operation of genetic algorithm, at first individuality acted on the BP neural network, with the Error Absolute Value between prediction output that draws and the desired output as this individual fitness value F.Select according to each individual adaptive value, then the selection Probability p of each individual i
iFor:
f
i=k/F
i
F wherein
iBe the fitness value of individuality, the coefficient of k for being provided with is a constant; N is the individual number of population, because of F
iWhat reflect is the error between desired output and the prediction output, and the bright individuality of novel is excellent more more for error, and selecteed probability is just big more, so with F
iGet inverse and become f
iSo that with select the big or small proportional of probability; f
jMeaning and f
iIdentical, wherein subscript j represents the population individual amount;
Step 5.2.3, carry out the interlace operation of genetic algorithm, need chromosome that intersects and the position that needs to intersect to select at random.Suppose k chromosome a
kWith l chromosome a
lIntersect j position, then method of operating is following:
Wherein b is the random number between [0,1];
Step 5.2.4, carry out the mutation operation of genetic algorithm, need the chromosome of variation and the position of variation to select at random.Suppose the gene a of j position of i individuals
IjNeed make a variation, then method of operating is following:
A wherein
MaxBe gene a
IjThe upper bound, a
MinBe gene a
IjLower bound;
R and r
2All be the random number between [0,1], g is the current iteration number of times, G
MsxIt is maximum iteration time.
Step 5.2.5, circulation step 5.2.2 are provided with the circulation maximum times to step 5.2.4, when circulation reaches maximum times, stop, and take out initial weight and the threshold value of the minimum individuality of fitness value F as neural metwork training.
Claims (7)
1. the monocular infrared image depth estimation method based on optimization BP neural network model is characterized in that, may further comprise the steps:
(1) (x y) reaches and said monocular infrared image I (x, y) pairing depth map to obtain any monocular infrared image I;
(2) be said monocular infrared image I (x; Y) each pixel in is set the characteristic area at least three different scales; Three ascending being respectively of different scale: first yardstick, second yardstick and the 3rd yardstick; Wherein, the characteristic area of each yardstick comprises that at least the image block that is positioned at the center reaches and the adjacent image block in said image block upper and lower, left and right, and the image block that be positioned at center of i pixel on first yardstick is i pixel itself; I pixel comprises all images piece on first yardstick at the image block that is positioned at the center on second yardstick; I pixel comprises all images piece on second yardstick at the image block that is positioned at the center on the 3rd yardstick; By that analogy;
(3) calculate monocular infrared image I (x; Y) proper vector of the pairing characteristic area of each pixel in; The characteristic component of the proper vector of i pixel comprises at least: the gray-scale value of all images piece of i pixel on first yardstick; The texture energy of each image block of i pixel on other yardsticks except that first yardstick; The gradient energy of the different directions of each image block of i pixel on other yardsticks except that first yardstick and the average and the variance of all gradient energy, and the sharpness of each image block of i pixel on other yardsticks except that first yardstick;
(4) utilize the method for progressively linear regression analysis and independent component analysis to screen successively to all proper vectors of obtaining, obtain meeting the proper vector of infrared image depth information;
(5) utilize the middle screening of step (4) to obtain proper vector and said monocular infrared image I (x; Y) pairing depth map makes up the set of degree of depth training sample; The proper vector in the set of degree of depth training sample and the depth value of depth map are carried out nonlinear fitting with the BP neural network; And be optimized, and then make up depth model through the initial weight and the threshold value of genetic algorithm to the BP neural network;
The new monocular infrared image that (6) will collect obtains the estimation of Depth value through the depth model analysis that structure obtains.
2. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that; Texture energy in the said step (3) calculates through the Louth mask, and concrete steps are: adopt N basic two-dimentional Louth mask, be designated as M
1..., M
N, with said monocular infrared image I (x y) does convolution with each two-dimentional Louth mask, then monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after value be:
3. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that; The gradient energy of the different directions of each image block of i pixel on second yardstick and the 3rd yardstick and the average and the variance of all gradient energy calculate through following steps in the said step (3), and concrete steps are:
(x y) tries to achieve x axial gradient figure I respectively on x direction of principal axis and y direction of principal axis to monocular infrared image I
Gradx(x is y) with y axial gradient figure I
Grady(x, y), then (x is y) at angle θ for monocular infrared image I
lGrad G on the direction
l(x, y)=I
Gradx(x, y) * cos (θ
l)+I
Grady(x, y) * sin (θ
l), wherein,
L=0 ..., 7, calculate the gradient energy of the different directions of each image block of each pixel on other yardsticks except that first yardstick subsequently, wherein, i m the image block N of pixel on the j yardstick
i(m) at angle θ
lGradient energy on the direction does
I m the image block N of pixel on the j yardstick then
i(m) at angle θ
lThe average of the gradient energy on the direction
I m the image block N of pixel on the j yardstick
i(m) at angle θ
lGradient energy variance on the direction
Wherein, size is image block N
iThe number of the pixel that (m) is comprised,
4. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that; The sharpness of each image block of i pixel on second yardstick and the 3rd yardstick calculates through following steps in the said step (3), and concrete steps are: i m the image block N of pixel on the j yardstick
i(m) sharpness
Wherein, (x y) is image block N to I
i(m) contain the gray scale of pixel,
Be image block N
i(m) average of the gray scale of contained pixel,
Size is image block N
iThe number of the pixel that (m) is comprised.
5. according to claim 1ly it is characterized in that based on the monocular infrared image depth estimation method of optimizing the BP neural network model, in the said step (4) progressively linear regression analysis comprise following substep:
(411) related coefficient of each characteristic component and depth value in the calculated characteristics vector is according to the descending ordering that obtains each characteristic component to the effect of depth degree of the absolute value of related coefficient;
(412) the absolute value biggest characteristic component from related coefficient begins progressively to introduce regression equation; And do regression equation significance test; If do not think significantly that selected whole characteristic component all is not the principal element that influences depth value, if significantly again from depth value is influenced the descending regression equation of introducing one by one successively;
(413) new characteristic component of every introducing all need carry out significance test to each contained in regression equation characteristic component, and that significantly and to depth value is not influenced minimum characteristic component rejection in new regression equation, repeats
Till this step each characteristic component in regression equation is all remarkable;
(414) introduce a characteristic component that in the characteristic component of not introducing depth value is had the greatest impact, repeating step again
(413) and step (414), up to rejecting selected characteristic component, also can't introduce till the new characteristic component.
6. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that; What independent component analysis was used in the said step (4) is quick ICA algorithm; To analyzing through the proper vector of progressively linear regression analysis, make between each component independently as much as possible through quick ICA algorithm, comprise following substep:
(421) specifying the proper vector through M pixel after the linear regression analysis progressively is observation data X, and observation data X is carried out centralization, and making it average is 0;
(422), be about to become the albefaction vector Z, Z=W after observation data X projects to new subspace with the observation data X albefaction of centralization
0X, wherein, W
0Be albefaction matrix, W
0=Λ
-1/2U
T, Λ is the eigenvalue matrix of the covariance matrix of observation data X, U is the eigenvectors matrix of the covariance matrix of observation data X;
(423) upgrade W
*Make W
*=E{Zg (W
TZ) }-E{g ' (W
TZ) } W, wherein g () is a nonlinear function, and then to W
*Standardization: W=W
*/ || W
*||, if do not restrain and then do not repeat this step, wherein, selecting a mould is 1 the initial random weight vector initial value as W.
7. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that initial weight and the threshold value of utilizing genetic algorithm to be optimized the BP neural network in the said step (5) also comprise following substep:
(511) make up initial BP neural network according to training sample, and confirm the number of input layer, hidden layer and the output layer of this network;
(512) carry out the initialization of genetic algorithm population; Wherein each individuality of population all is to be made up of the threshold value of the weights between the weights between input layer and the hidden layer, hidden layer and the output layer, hidden layer and threshold value four parts of output layer, and all weights and threshold value are listed as into a real number vector;
(513) maximum iteration time is set as stopping criterion for iteration, iterative process each time comprises selection, intersection and mutation operation, when reaching stopping criterion for iteration, output the last reign of a dynasty population optimum individual be the approximate optimal solution of initial weight and threshold value.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210206701.1A CN102750702B (en) | 2012-06-21 | 2012-06-21 | Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201210206701.1A CN102750702B (en) | 2012-06-21 | 2012-06-21 | Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model |
Publications (2)
Publication Number | Publication Date |
---|---|
CN102750702A true CN102750702A (en) | 2012-10-24 |
CN102750702B CN102750702B (en) | 2014-10-15 |
Family
ID=47030851
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201210206701.1A Expired - Fee Related CN102750702B (en) | 2012-06-21 | 2012-06-21 | Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN102750702B (en) |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106408015A (en) * | 2016-09-13 | 2017-02-15 | 电子科技大学成都研究院 | Road fork identification and depth estimation method based on convolutional neural network |
CN107204010A (en) * | 2017-04-28 | 2017-09-26 | 中国科学院计算技术研究所 | A kind of monocular image depth estimation method and system |
CN107578436A (en) * | 2017-08-02 | 2018-01-12 | 南京邮电大学 | A kind of monocular image depth estimation method based on full convolutional neural networks FCN |
CN107607207A (en) * | 2017-08-22 | 2018-01-19 | 中国科学院深圳先进技术研究院 | A kind of power equipment thermal fault detection method, system and electronic equipment |
CN107690663A (en) * | 2015-06-05 | 2018-02-13 | 谷歌有限责任公司 | Albefaction neural net layer |
CN107993239A (en) * | 2017-12-25 | 2018-05-04 | 北京邮电大学 | A kind of method and apparatus for the depth order for calculating monocular image |
CN108335322A (en) * | 2018-02-01 | 2018-07-27 | 深圳市商汤科技有限公司 | Depth estimation method and device, electronic equipment, program and medium |
CN109035319A (en) * | 2018-07-27 | 2018-12-18 | 深圳市商汤科技有限公司 | Monocular image depth estimation method and device, equipment, program and storage medium |
CN109146944A (en) * | 2018-10-30 | 2019-01-04 | 浙江科技学院 | A kind of space or depth perception estimation method based on the revoluble long-pending neural network of depth |
CN109377524A (en) * | 2018-10-29 | 2019-02-22 | 山东师范大学 | A kind of single image depth recovery method and system |
CN109461177A (en) * | 2018-09-29 | 2019-03-12 | 浙江科技学院 | A kind of monocular image depth prediction approach neural network based |
CN109598704A (en) * | 2018-11-19 | 2019-04-09 | 电子科技大学 | A kind of night soil microscopic image clarity evaluation method based on BP neural network |
CN109903393A (en) * | 2019-02-22 | 2019-06-18 | 清华大学 | New Century Planned Textbook Scene Composition methods and device based on deep learning |
CN109964237A (en) * | 2016-09-15 | 2019-07-02 | 谷歌有限责任公司 | Picture depth prediction neural network |
CN112036470A (en) * | 2020-08-28 | 2020-12-04 | 扬州大学 | Cloud transmission-based multi-sensor fusion cucumber bemisia tabaci identification method |
CN112215410A (en) * | 2020-09-24 | 2021-01-12 | 国网福建省电力有限公司 | Power load prediction method based on improved deep learning |
CN112329189A (en) * | 2019-07-30 | 2021-02-05 | 天津大学青岛海洋技术研究院 | Transient electromagnetic useful signal extraction method based on independent quantity analysis |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100135581A1 (en) * | 2008-12-02 | 2010-06-03 | Samsung Electronics Co., Ltd. | Depth estimation apparatus and method |
CN101789123A (en) * | 2010-01-27 | 2010-07-28 | 中国科学院半导体研究所 | Method for creating distance map based on monocular camera machine vision |
-
2012
- 2012-06-21 CN CN201210206701.1A patent/CN102750702B/en not_active Expired - Fee Related
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20100135581A1 (en) * | 2008-12-02 | 2010-06-03 | Samsung Electronics Co., Ltd. | Depth estimation apparatus and method |
CN101789123A (en) * | 2010-01-27 | 2010-07-28 | 中国科学院半导体研究所 | Method for creating distance map based on monocular camera machine vision |
Non-Patent Citations (4)
Title |
---|
CLAUDIA HENDRIX ET AL.: "Relationship between monocular and binocular depth cues for judgements of spatial information and spatial instrument design", 《DISPLAYS》 * |
TIEN-YING KUO ET AL.: "Depth Estimation from a Monocular Outdoor Image", 《2011 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS》 * |
蓝建梁等: "基于多尺度纹理能量测度的单幅图像深度估计", 《计算机工程与设计》 * |
谷小婧: "基于图像分析的自然彩色夜视成像方法研究", 《中国博士学位论文全文数据库》 * |
Cited By (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107690663A (en) * | 2015-06-05 | 2018-02-13 | 谷歌有限责任公司 | Albefaction neural net layer |
CN107690663B (en) * | 2015-06-05 | 2022-04-12 | 渊慧科技有限公司 | Whitening neural network layer |
CN106408015A (en) * | 2016-09-13 | 2017-02-15 | 电子科技大学成都研究院 | Road fork identification and depth estimation method based on convolutional neural network |
CN109964237B (en) * | 2016-09-15 | 2020-07-17 | 谷歌有限责任公司 | Image depth prediction neural network |
CN109964237A (en) * | 2016-09-15 | 2019-07-02 | 谷歌有限责任公司 | Picture depth prediction neural network |
US10929996B2 (en) | 2016-09-15 | 2021-02-23 | Google Llc | Image depth prediction neural networks |
US11734847B2 (en) | 2016-09-15 | 2023-08-22 | Google Llc | Image depth prediction neural networks |
CN107204010A (en) * | 2017-04-28 | 2017-09-26 | 中国科学院计算技术研究所 | A kind of monocular image depth estimation method and system |
CN107204010B (en) * | 2017-04-28 | 2019-11-19 | 中国科学院计算技术研究所 | A kind of monocular image depth estimation method and system |
CN107578436A (en) * | 2017-08-02 | 2018-01-12 | 南京邮电大学 | A kind of monocular image depth estimation method based on full convolutional neural networks FCN |
CN107607207A (en) * | 2017-08-22 | 2018-01-19 | 中国科学院深圳先进技术研究院 | A kind of power equipment thermal fault detection method, system and electronic equipment |
CN107607207B (en) * | 2017-08-22 | 2019-06-28 | 中国科学院深圳先进技术研究院 | A kind of power equipment thermal fault detection method, system and electronic equipment |
CN107993239B (en) * | 2017-12-25 | 2022-04-12 | 北京邮电大学 | Method and device for calculating depth order of monocular image |
CN107993239A (en) * | 2017-12-25 | 2018-05-04 | 北京邮电大学 | A kind of method and apparatus for the depth order for calculating monocular image |
CN108335322A (en) * | 2018-02-01 | 2018-07-27 | 深圳市商汤科技有限公司 | Depth estimation method and device, electronic equipment, program and medium |
US11308638B2 (en) | 2018-02-01 | 2022-04-19 | Shenzhen Sensetime Technology Co., Ltd. | Depth estimation method and apparatus, electronic device, program, and medium |
CN108335322B (en) * | 2018-02-01 | 2021-02-12 | 深圳市商汤科技有限公司 | Depth estimation method and apparatus, electronic device, program, and medium |
US11443445B2 (en) | 2018-07-27 | 2022-09-13 | Shenzhen Sensetime Technology Co., Ltd. | Method and apparatus for depth estimation of monocular image, and storage medium |
JP2021500689A (en) * | 2018-07-27 | 2021-01-07 | 深▲せん▼市商▲湯▼科技有限公司Shenzhen Sensetime Technology Co., Ltd. | Monocular image depth estimation method and equipment, equipment, programs and storage media |
CN109035319A (en) * | 2018-07-27 | 2018-12-18 | 深圳市商汤科技有限公司 | Monocular image depth estimation method and device, equipment, program and storage medium |
CN109035319B (en) * | 2018-07-27 | 2021-04-30 | 深圳市商汤科技有限公司 | Monocular image depth estimation method, monocular image depth estimation device, monocular image depth estimation apparatus, monocular image depth estimation program, and storage medium |
CN109461177A (en) * | 2018-09-29 | 2019-03-12 | 浙江科技学院 | A kind of monocular image depth prediction approach neural network based |
CN109461177B (en) * | 2018-09-29 | 2021-12-10 | 浙江科技学院 | Monocular image depth prediction method based on neural network |
CN109377524A (en) * | 2018-10-29 | 2019-02-22 | 山东师范大学 | A kind of single image depth recovery method and system |
CN109377524B (en) * | 2018-10-29 | 2021-02-23 | 山东师范大学 | Method and system for recovering depth of single image |
CN109146944A (en) * | 2018-10-30 | 2019-01-04 | 浙江科技学院 | A kind of space or depth perception estimation method based on the revoluble long-pending neural network of depth |
CN109146944B (en) * | 2018-10-30 | 2020-06-26 | 浙江科技学院 | Visual depth estimation method based on depth separable convolutional neural network |
CN109598704A (en) * | 2018-11-19 | 2019-04-09 | 电子科技大学 | A kind of night soil microscopic image clarity evaluation method based on BP neural network |
CN109903393A (en) * | 2019-02-22 | 2019-06-18 | 清华大学 | New Century Planned Textbook Scene Composition methods and device based on deep learning |
CN112329189A (en) * | 2019-07-30 | 2021-02-05 | 天津大学青岛海洋技术研究院 | Transient electromagnetic useful signal extraction method based on independent quantity analysis |
CN112036470A (en) * | 2020-08-28 | 2020-12-04 | 扬州大学 | Cloud transmission-based multi-sensor fusion cucumber bemisia tabaci identification method |
CN112215410A (en) * | 2020-09-24 | 2021-01-12 | 国网福建省电力有限公司 | Power load prediction method based on improved deep learning |
CN112215410B (en) * | 2020-09-24 | 2023-01-10 | 国网福建省电力有限公司 | Power load prediction method based on improved deep learning |
Also Published As
Publication number | Publication date |
---|---|
CN102750702B (en) | 2014-10-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN102750702B (en) | Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model | |
CN102708569B (en) | Based on the monocular infrared image depth estimation method of SVM model | |
CN108038445B (en) | SAR automatic target identification method based on multi-view deep learning framework | |
CN109614985B (en) | Target detection method based on densely connected feature pyramid network | |
CN112766199B (en) | Hyperspectral image classification method based on self-adaptive multi-scale feature extraction model | |
CN107132516B (en) | A kind of Radar range profile's target identification method based on depth confidence network | |
CN105869178A (en) | Method for unsupervised segmentation of complex targets from dynamic scene based on multi-scale combination feature convex optimization | |
CN107247938A (en) | A kind of method of high-resolution remote sensing image City Building function classification | |
CN110110621A (en) | The oblique photograph point cloud classifications method of deep learning model is integrated based on multiple features | |
CN105787557A (en) | Design method of deep nerve network structure for computer intelligent identification | |
CN112862090B (en) | Air temperature forecasting method based on deep space-time neural network | |
CN112347970B (en) | Remote sensing image ground object identification method based on graph convolution neural network | |
CN114092697B (en) | Building facade semantic segmentation method with attention fused with global and local depth features | |
CN110852393A (en) | Remote sensing image segmentation method and system | |
CN109784288B (en) | Pedestrian re-identification method based on discrimination perception fusion | |
CN105095913B (en) | The Classifying Method in Remote Sensing Image and system represented based on neighbour's canonical joint sparse | |
CN106991411B (en) | Remote Sensing Target based on depth shape priori refines extracting method | |
CN114758252A (en) | Image-based distributed photovoltaic roof resource segmentation and extraction method and system | |
CN112766280A (en) | Remote sensing image road extraction method based on graph convolution | |
CN104318215B (en) | A kind of cross-view face identification method based on domain robust convolution feature learning | |
CN116206185A (en) | Lightweight small target detection method based on improved YOLOv7 | |
CN105701493A (en) | Methods and systems for image matting and foreground estimation based on hierarchical graphs | |
CN104680167B (en) | Auroral oval location determining method based on deep learning | |
Simon et al. | Parameter-free/pareto-driven procedural 3d reconstruction of buildings from ground-level sequences | |
CN104850677A (en) | Deep learning cellular automaton simulation analysis method based on geographic information system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20141015 Termination date: 20170621 |
|
CF01 | Termination of patent right due to non-payment of annual fee |