CN102750702A - Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model - Google Patents

Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model Download PDF

Info

Publication number
CN102750702A
CN102750702A CN2012102067011A CN201210206701A CN102750702A CN 102750702 A CN102750702 A CN 102750702A CN 2012102067011 A CN2012102067011 A CN 2012102067011A CN 201210206701 A CN201210206701 A CN 201210206701A CN 102750702 A CN102750702 A CN 102750702A
Authority
CN
China
Prior art keywords
depth
pixel
infrared image
yardstick
image block
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2012102067011A
Other languages
Chinese (zh)
Other versions
CN102750702B (en
Inventor
孙韶媛
席林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Donghua University
Original Assignee
Donghua University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Donghua University filed Critical Donghua University
Priority to CN201210206701.1A priority Critical patent/CN102750702B/en
Publication of CN102750702A publication Critical patent/CN102750702A/en
Application granted granted Critical
Publication of CN102750702B publication Critical patent/CN102750702B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The invention relates to a monocular infrared image depth estimation method based on an optimized BP (Back Propagation) neural network model. The method comprises the following steps of: acquiring a monocular infrared image and a depth map to which the monocular infrared image corresponds; setting at least three feature regions with different scales for pixel points in the monocular infrared image; calculating feature vectors of the feature regions to which the pixel points in the monocular infrared image correspond; screening all the feature vectors by successively using stepwise linear regression and independent component analysis methods to obtain feature vectors conforming to depth information of the infrared image; constructing a depth training sample set by using the obtained feature vectors and the depth map to which the infrared image corresponds, and performing nonlinear fitting on the feature vectors in the set and depth values of the depth map by using a BP neuron network, and optimizing the BP neuron network through a genetic algorithm, and then constructing a depth model; and analyzing the monocular infrared image through the depth model to obtain a depth estimated value. By using the monocular infrared image depth estimation method based on the optimized BP neural network model, the depth information of the infrared image can be relatively accurately estimated.

Description

Based on the monocular infrared image depth estimation method of optimizing the BP neural network model
Technical field
The present invention relates to infrared image estimation of Depth technical field, particularly relate to a kind of monocular infrared image depth estimation method based on the genetic algorithm optimization BP neural network model.
Background technology
The estimation of Depth of image promptly is from image, to obtain depth distance information, is the problem of a depth perception in essence.By depth perception make up spatial positional information characterized be from observer to the scene in the relative distance on detected surface.Recover the now existing more satisfactory algorithm of depth distance information in the coloured image, but for infrared image, because of its reflection be the Temperature Distribution of scene, defectives such as low signal-to-noise ratio, low contrast are arranged, the degree of depth algorithm that recovers this image still belongs to blank.If can recover the depth information of infrared image, will greatly improve the understanding effect of human eye so to this image.
The picture depth method of estimation mainly is to launch to binocular degree of depth clue with based on the estimation of Depth of image sequence at present, and these two kinds of methods all depend on the feature difference between image.And estimate for monocular depth; Comparatively classical in the middle of traditional in early days algorithm is " by shadow recovery shape (shape from shading) "; This algorithm is a theoretical foundation how much with space multistory; Changing to light and shade that body surface produced according to light source irradiation is that the shade of image recovers the object degree of depth, but because this algorithm needs priori (like reflection model and light source direction etc.) to make the application limit increase.Afterwards, some researchers find the importance of experience gradually, begin to utilize the method for machine learning to go to address this problem.The team of the Andrew Ng of Stanford University carries out estimation of Depth through the model that utilizes the Markov field training to single image, has reached good effect; Simple classification in the scene is then manually demarcated by the Aloysha Efros team of CMU before training; Such as sky, trees, ground and perpendicular line or the like; Utilize these classifications of lot of data study then, thereby and finally new images is classified and recover depth information through the structure Bayesian model.Though it is this method is comparatively suitable to the simple picture of a series of scenes, and produces a desired effect, inaccurate often for the classification that does not have study in the scene.
Summary of the invention
Technical matters to be solved by this invention provides a kind of based on the monocular infrared image depth estimation method of optimizing the BP neural network model, and this method can be estimated the depth information of infrared image comparatively exactly.
The technical solution adopted for the present invention to solve the technical problems is: provide a kind of based on the monocular infrared image depth estimation method of optimizing the BP neural network model, may further comprise the steps:
(1) (x y) reaches and the pairing depth map of said monocular infrared image I (x, y) to obtain any monocular infrared image I;
(2) be said monocular infrared image I (x; Y) each pixel in is set the characteristic area at least three different scales; Three ascending being respectively of different scale: first yardstick, second yardstick and the 3rd yardstick; Wherein, the characteristic area of each yardstick comprises that at least the image block that is positioned at the center reaches and the adjacent image block in said image block upper and lower, left and right, and the image block that be positioned at center of i pixel on first yardstick is i pixel itself; I pixel comprises all images piece on first yardstick at the image block that is positioned at the center on second yardstick; I pixel comprises all images piece on second yardstick at the image block that is positioned at the center on the 3rd yardstick; By that analogy;
(3) calculate monocular infrared image I (x; Y) proper vector of the pairing characteristic area of each pixel in; The characteristic component of the proper vector of i pixel comprises at least: the gray-scale value of all images piece of i pixel on first yardstick; The texture energy of each image block of i pixel on other yardsticks except that first yardstick; The gradient energy of the different directions of each image block of i pixel on other yardsticks except that first yardstick and the average and the variance of all gradient energy, and the sharpness of each image block of i pixel on other yardsticks except that first yardstick;
(4) utilize the method for progressively linear regression analysis and independent component analysis to screen successively to all proper vectors of obtaining, obtain meeting the proper vector of infrared image depth information;
(5) utilize the middle screening of step (4) to obtain proper vector and said monocular infrared image I (x; Y) pairing depth map makes up the set of degree of depth training sample; The proper vector in the set of degree of depth training sample and the depth value of depth map are carried out nonlinear fitting with the BP neural network; And be optimized, and then make up depth model through the initial weight and the threshold value of genetic algorithm to the BP neural network;
The new monocular infrared image that (6) will collect obtains the estimation of Depth value through the depth model analysis that structure obtains.
Texture energy in the said step (3) calculates through the Louth mask, and concrete steps are: adopt N basic two-dimentional Louth mask, be designated as M 1..., M N, with said monocular infrared image I (x y) does convolution with each two-dimentional Louth mask, then monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after value be: T k(x, y)=I (x, y) * M k, k=1 ..., N, then i m the image block N of pixel on the j yardstick i(m) monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after the texture energy that obtains do E n N i ( m ) j , E n N i ( m ) j = Σ x , y ∈ N i ( m ) | T k ( x , y ) | .
The gradient energy of the different directions of each image block of i pixel on second yardstick and the 3rd yardstick and the average and the variance of all gradient energy calculate through following steps in the said step (3); Concrete steps are: (x y) tries to achieve x axial gradient figure I respectively on x direction of principal axis and y direction of principal axis to monocular infrared image I Gradx(x is y) with y axial gradient figure I Grady(x, y), then (x is y) at angle θ for monocular infrared image I lGrad G on the direction l(x, y)=I Gradx(x, y) * cos (θ l)+I Grady(x, y) * sin (θ l), wherein,
Figure BDA00001795989100031
L=0 ..., 7, calculate the gradient energy of the different directions of each image block of each pixel on other yardsticks except that first yardstick subsequently, wherein, i m the image block N of pixel on the j yardstick i(m) at angle θ lGradient energy on the direction does I m the image block N of pixel on the j yardstick then i(m) at angle θ lThe average of the gradient energy on the direction
Figure BDA00001795989100033
I m the image block N of pixel on the j yardstick i(m) at angle θ lGradient energy variance on the direction
Figure BDA00001795989100034
Wherein, size is image block N iThe number of the pixel that (m) is comprised,
Figure BDA00001795989100035
The sharpness of each image block of i pixel on second yardstick and the 3rd yardstick calculates through following steps in the said step (3), and concrete steps are: i m the image block N of pixel on the j yardstick i(m) sharpness
Figure BDA00001795989100036
Wherein, (x y) is image block N to I i(m) contain the gray scale of pixel,
Figure BDA00001795989100037
Be image block N i(m) average of the gray scale of contained pixel,
Figure BDA00001795989100038
Size is image block N iThe number of the pixel that (m) is comprised.
In the said step (4) progressively linear regression analysis comprise following substep:
(411) related coefficient of each characteristic component and depth value in the calculated characteristics vector is according to the descending ordering that obtains each characteristic component to the effect of depth degree of the absolute value of related coefficient;
(412) the absolute value biggest characteristic component from related coefficient begins progressively to introduce regression equation; And do regression equation significance test; If do not think significantly that selected whole characteristic component all is not the principal element that influences depth value, if significantly again from depth value is influenced the descending regression equation of introducing one by one successively;
(413) new characteristic component of every introducing all need carry out significance test to each contained in regression equation characteristic component; With that in new regression equation significantly and to the minimum characteristic component rejection of depth value influence, repeat this step each characteristic component in regression equation all significantly till;
(414) introduce a characteristic component that in the characteristic component of not introducing depth value is had the greatest impact again, repeating step (413) and step (414) up to rejecting selected characteristic component, also can't be introduced till the new characteristic component.
What independent component analysis was used in the said step (4) is quick ICA algorithm, to analyzing through the proper vector of progressively linear regression analysis, makes between each component independently as much as possible through quick ICA algorithm, comprises following substep:
(421) specifying the proper vector through M pixel after the linear regression analysis progressively is observation data X, and observation data X is carried out centralization, and making it average is 0;
(422), be about to become the albefaction vector Z, Z=W after observation data X projects to new subspace with the observation data X albefaction of centralization 0X, wherein, W 0Be albefaction matrix, W 0-1/2U T, Λ is the eigenvalue matrix of the covariance matrix of observation data X, U is the eigenvectors matrix of the covariance matrix of observation data X;
(423) upgrade W *Make W *=E{Zg (W TZ) }-E{g ' (W TZ}W, wherein g () is a nonlinear function, and then to W *Standardization: W=W */ || W *||, if do not restrain and then do not repeat this step, wherein, selecting a mould is 1 the initial random weight vector initial value as W.
Initial weight and the threshold value of utilizing genetic algorithm to be optimized the BP neural network in the said step (5) also comprise following substep:
(511) make up initial BP neural network according to training sample, and confirm the number of input layer, hidden layer and the output layer of this network;
(512) carry out the initialization of genetic algorithm population; Wherein each individuality of population all is to be made up of the threshold value of the weights between the weights between input layer and the hidden layer, hidden layer and the output layer, hidden layer and threshold value four parts of output layer, and all weights and threshold value are listed as into a real number vector;
(513) maximum iteration time is set as stopping criterion for iteration, iterative process each time comprises selection, intersection and mutation operation, when reaching stopping criterion for iteration, output the last reign of a dynasty population optimum individual be the approximate optimal solution of initial weight and threshold value.
Beneficial effect
Owing to adopted above-mentioned technical scheme; The present invention compared with prior art; Have following advantage and good effect: the present invention utilizes infrared image " spatial context " and " multiple dimensioned " information extraction depth characteristic vector; Through the method for progressively linear regression analysis and independent component analysis the characteristic of extracting is screened successively, help finding the depth characteristic that is fit to infrared image more, and make up degree of depth training set with this; Adopt the back propagation learning theory that nonlinear fitting is carried out in the training set, and be optimized, improved the training speed of experiment and the accuracy rate of match through genetic algorithm.
Description of drawings
Fig. 1 is a process flow diagram of the present invention.
Embodiment
Below in conjunction with specific embodiment, further set forth the present invention.Should be understood that these embodiment only to be used to the present invention is described and be not used in the restriction scope of the present invention.Should be understood that in addition those skilled in the art can do various changes or modification to the present invention after the content of having read the present invention's instruction, these equivalent form of values fall within the application's appended claims institute restricted portion equally.
As shown in Figure 1, present embodiment discloses a kind of monocular infrared image depth estimation method of the BP neural network model based on genetic algorithm optimization, the steps include:
Step 1, obtain any monocular infrared image I (x, y) and with this monocular infrared image I (x, y) pairing depth map;
Step 2, be monocular infrared image I (x; Y) each pixel in is set the characteristic area on three different scales; Three ascending being respectively of different scale: first yardstick, second yardstick and the 3rd yardstick; Wherein, The characteristic area of each yardstick comprises image block and the image block adjacent with these image block upper and lower, left and right that is positioned at the center; The image block that be positioned at center of i pixel on first yardstick is i pixel itself, and this i pixel comprises all images piece on first yardstick at the image block that is positioned at the center on second yardstick, and this i pixel comprises all images piece on second yardstick at the image block that is positioned at the center on the 3rd yardstick;
Step 3, calculating monocular infrared image I (x; Y) proper vector of the pairing characteristic area of each pixel in; The characteristic component of the proper vector of i pixel comprises at least: the gray-scale value of all images piece of i pixel on first yardstick; The texture energy of each image block of i pixel on second yardstick and the 3rd yardstick; I pixel gradient energy and the average and variance of all gradient energy on 8 directions of each image block on second yardstick and the 3rd yardstick, and the sharpness of each image block of i pixel on second yardstick and the 3rd yardstick;
The texture energy of each image block of i pixel on second yardstick and the 3rd yardstick calculates through following steps:
Adopt 9 basic two-dimentional Louth masks, be designated as M 1..., M 9, with monocular infrared image I (x y) does convolution with each two-dimentional Louth mask, then monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after value be: T k(x, y)=I (x, y) * M k, k=1 ..., 9, i m the image block N of pixel on the j yardstick then i(m) monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after the texture energy that obtains do E n N i ( m ) j , E n N i ( m ) j = Σ x , y ∈ N i ( m ) | T k ( x , y ) | , j = 2,3 .
The gradient energy of the different directions of each image block of i pixel on second yardstick and the 3rd yardstick and the average and the variance of all gradient energy calculate through following steps:
(x y) tries to achieve x axial gradient figure I respectively on x direction of principal axis and y direction of principal axis to monocular infrared image I Gradx(x is y) with y axial gradient figure I Grady(x, y), then (x is y) at angle θ for monocular infrared image I lGrad G on the direction l(x, y)=I Gradx(x, y) * cos (θ l)+I Grady(x, y) * sin (θ l), wherein, L=0 ..., 7, calculate the gradient energy of the different directions of each image block of each pixel on other yardsticks except that first yardstick subsequently, wherein, i m the image block N of pixel on the j yardstick i(m) at angle θ lGradient energy on the direction does
Figure BDA00001795989100064
I m the image block N of pixel on the j yardstick then i(m) at angle θ lThe average of the gradient energy on the direction
Figure BDA00001795989100065
I m the image block N of pixel on the j yardstick i(m) at angle θ lGradient energy variance on the direction Wherein, size is image block N iThe number of the pixel that (m) is comprised,
Figure BDA00001795989100067
The sharpness of each image block of i pixel on second yardstick and the 3rd yardstick calculates through following steps:
I m the image block N of pixel on the j yardstick i(m) sharpness
Figure BDA00001795989100068
Wherein, (x y) is image block N to I i(m) contain the gray scale of pixel,
Figure BDA00001795989100069
Be image block N i(m) average of the gray scale of contained pixel,
Figure BDA000017959891000610
Step 4, utilize the method for progressively linear regression analysis and independent component analysis to screen successively, comparatively met the proper vector of infrared image depth information all proper vectors of having obtained in the step 3.
In the step 4 progressively the concrete steps of linear regression analysis be:
The related coefficient of each characteristic component and depth value in step 4.1.1, the calculated characteristics vector is according to the descending ordering that obtains each characteristic component to the effect of depth degree of the absolute value of related coefficient;
Step 4.1.2, begin progressively to introduce regression equation from the absolute value biggest characteristic component of related coefficient; And do regression equation significance test; If can not think significantly that selected whole characteristic component all is not the principal element that influences depth value, if significantly again from depth value is influenced the descending regression equation of introducing one by one successively;
Step 4.1.3, new characteristic component of every introducing all need carry out significance test to each contained in regression equation characteristic component; With that in new regression equation significantly and to the minimum characteristic component rejection of depth value influence, repeat this step each characteristic component in regression equation all significantly till;
Step 4.1.4, introduce a characteristic component that in the characteristic component of not introducing depth value is had the greatest impact again, repeating step 4.1.3 and step 4.1.4 up to rejecting selected characteristic component, also can't introduce till the new characteristic component.
What independent component analysis was used in the step 4 is quick ICA algorithm, to analyzing through the proper vector of progressively linear regression analysis, makes between each component independent as much as possible through quick ICA algorithm.In the present embodiment, the ICA algorithm is maximum as searching direction with negentropy fast, and its concrete steps are following:
Step 4.2.1, to specify the proper vector through M pixel after the linear regression analysis progressively be observation data X, and observation data X is carried out centralization, and making it average is 0;
Step 4.2.2, with the observation data X albefaction of centralization, be about to become the albefaction vector Z, Z=W after observation data X will project to new subspace 0X, wherein, W 0Be albefaction matrix, W 0-1/2U T, Λ is the eigenvalue matrix of the covariance matrix of observation data X, U is the eigenvectors matrix of the covariance matrix of observation data X;
Step 4.2.3, renewal W *Be to find the solution the separation matrix that draws, the separation matrix W that finally draws *And observation data X multiplies each other and can obtain between each component as far as possible independently source data and make W *=E{Zg (W TZ) }-E{g ' (W TZ) } W, wherein E () is the expectation computing, wherein g () is a nonlinear function, and then to W *Standardization: W=W */ || W *||, if do not restrain and then do not repeat this step, wherein, selecting a mould is 1 the initial random weight vector initial value as W.
The depth map of step 5, the proper vector of utilizing the infrared image that is drawn by step 4 and infrared image makes up degree of depth training sample set { f i, depth i, wherein, f iBe the proper vector of i pixel, i=1 ..., l, f i∈ χ, depth iBe i the pairing depth value of pixel.Utilize back propagation learning The Theory Construction depth model then, and be optimized, and then new infrared image is carried out estimation of Depth through genetic algorithm.
The concrete steps of BP neural metwork training are following in the step 5:
Step 5.1.1, according to proper vector f iWith depth value depth iConfirm the input layer of BP neural network and the node number of output layer.Then through counting scope with reference to the roughly definite hidden layer node of formula
Figure BDA00001795989100071
; Wherein n is the input layer number; M is an output layer node number; A is the constant between 0 ~ 10, and l is the hidden layer node number.Confirm the hidden layer node number that square error is minimum through trial method again;
The weights and the threshold value of step 5.1.2, initialization BP neural network, and given maximum iteration time, learning rate parameter η and neuron excitation function wherein a be constant;
The output H of step 5.1.3, calculating hidden layer:
H j = f ( Σ i = 1 n ω ij x i - a j ) , j = 1,2 , . . . , l
Wherein l is the hidden layer node number, ω IjFor connecting the weights between input layer and the hidden layer, x iBe the proper vector f that mentions among the step 5.1.1 i, a jBe the threshold value of each hidden layer node, the neuron excitation function of f (.) for mentioning among the step 5.1.2;
Step 5.1.4, calculating prediction output O:
O k = Σ j = 1 l H j ω jk - b k , k = 1,2 , . . . , m
Wherein m is an output layer node number, ω JkFor connecting the weights between hidden layer and the output layer, b kThreshold value for each output layer node;
Step 5.1.5, according to prediction output and desired output error of calculation e, and upgrade the weights and the threshold value of network according to error e:
e k=Y k-O k,k=1,...,m
ω ij = ω ij + η H j ( 1 - H j ) x ( i ) Σ k = 1 m ω jk e k , i = 1 , . . . , n ; j = 1 , . . . , l
ω jkjk+ηH je k,j=1,...,l;k=1,...,m
a j = a j + η ( 1 - H j ) Σ k = 1 m ω jk e k , j = 1 , . . . , l
b k=b k+e k,k=1,...,m
Wherein n is the input layer number, Y kBe desired output, i.e. depth value in the training set;
Step 5.1.6, circulation step 5.1.3 finish after reaching maximum iteration time to step 5.1.5;
Step 5.1.7, to obtaining the proper vector of each pixel after feature extraction and the screening of new infrared image through step 1 and step 2; Each component of proper vector is corresponded to the input layer of the BP neural network of having trained respectively; Draw corresponding depth value through prediction, just accomplished estimation new infrared image depth value.
BP neural network initial weight and threshold value are following through the concrete steps that genetic algorithm is optimized in the step 5:
Step 5.2.1, carry out the initialization of genetic algorithm population, wherein each individuality of population all is to be made up of the threshold value of the weights between the weights between input layer and the hidden layer, hidden layer and the output layer, hidden layer and threshold value four parts of output layer.All weights and threshold value are listed as into a real number vector;
Step 5.2.2, carry out the selection operation of genetic algorithm, at first individuality acted on the BP neural network, with the Error Absolute Value between prediction output that draws and the desired output as this individual fitness value F.Select according to each individual adaptive value, then the selection Probability p of each individual i iFor:
f i=k/F i
P i = f i / Σ j = 1 N f i
F wherein iBe the fitness value of individuality, the coefficient of k for being provided with is a constant; N is the individual number of population, because of F iWhat reflect is the error between desired output and the prediction output, and the bright individuality of novel is excellent more more for error, and selecteed probability is just big more, so with F iGet inverse and become f iSo that with select the big or small proportional of probability; f jMeaning and f iIdentical, wherein subscript j represents the population individual amount;
Step 5.2.3, carry out the interlace operation of genetic algorithm, need chromosome that intersects and the position that needs to intersect to select at random.Suppose k chromosome a kWith l chromosome a lIntersect j position, then method of operating is following:
a kj = a kj ( 1 - b ) + a lj b a lj = a lj ( 1 - b ) + a kj b
Wherein b is the random number between [0,1];
Step 5.2.4, carry out the mutation operation of genetic algorithm, need the chromosome of variation and the position of variation to select at random.Suppose the gene a of j position of i individuals IjNeed make a variation, then method of operating is following:
a ij = a ij + ( a ij - a max ) * f ( g ) , r &GreaterEqual; 0.5 a ij + ( a min - a ij ) * f ( g ) , r < 0.5
A wherein MaxBe gene a IjThe upper bound, a MinBe gene a IjLower bound;
Figure BDA00001795989100094
R and r 2All be the random number between [0,1], g is the current iteration number of times, G MsxIt is maximum iteration time.
Step 5.2.5, circulation step 5.2.2 are provided with the circulation maximum times to step 5.2.4, when circulation reaches maximum times, stop, and take out initial weight and the threshold value of the minimum individuality of fitness value F as neural metwork training.

Claims (7)

1. the monocular infrared image depth estimation method based on optimization BP neural network model is characterized in that, may further comprise the steps:
(1) (x y) reaches and said monocular infrared image I (x, y) pairing depth map to obtain any monocular infrared image I;
(2) be said monocular infrared image I (x; Y) each pixel in is set the characteristic area at least three different scales; Three ascending being respectively of different scale: first yardstick, second yardstick and the 3rd yardstick; Wherein, the characteristic area of each yardstick comprises that at least the image block that is positioned at the center reaches and the adjacent image block in said image block upper and lower, left and right, and the image block that be positioned at center of i pixel on first yardstick is i pixel itself; I pixel comprises all images piece on first yardstick at the image block that is positioned at the center on second yardstick; I pixel comprises all images piece on second yardstick at the image block that is positioned at the center on the 3rd yardstick; By that analogy;
(3) calculate monocular infrared image I (x; Y) proper vector of the pairing characteristic area of each pixel in; The characteristic component of the proper vector of i pixel comprises at least: the gray-scale value of all images piece of i pixel on first yardstick; The texture energy of each image block of i pixel on other yardsticks except that first yardstick; The gradient energy of the different directions of each image block of i pixel on other yardsticks except that first yardstick and the average and the variance of all gradient energy, and the sharpness of each image block of i pixel on other yardsticks except that first yardstick;
(4) utilize the method for progressively linear regression analysis and independent component analysis to screen successively to all proper vectors of obtaining, obtain meeting the proper vector of infrared image depth information;
(5) utilize the middle screening of step (4) to obtain proper vector and said monocular infrared image I (x; Y) pairing depth map makes up the set of degree of depth training sample; The proper vector in the set of degree of depth training sample and the depth value of depth map are carried out nonlinear fitting with the BP neural network; And be optimized, and then make up depth model through the initial weight and the threshold value of genetic algorithm to the BP neural network;
The new monocular infrared image that (6) will collect obtains the estimation of Depth value through the depth model analysis that structure obtains.
2. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that; Texture energy in the said step (3) calculates through the Louth mask, and concrete steps are: adopt N basic two-dimentional Louth mask, be designated as M 1..., M N, with said monocular infrared image I (x y) does convolution with each two-dimentional Louth mask, then monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after value be:
T k(x, y)=I (x, y) * M k, k=1 ..., N, then i m the image block N of pixel on the j yardstick i(m) monocular infrared image I (x, y) with k two-dimentional Louth mask convolution after the texture energy that obtains do
Figure FDA00001795989000011
En N i ( m ) j = &Sigma; x , y &Element; N i ( m ) | T k ( x , y ) | .
3. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that; The gradient energy of the different directions of each image block of i pixel on second yardstick and the 3rd yardstick and the average and the variance of all gradient energy calculate through following steps in the said step (3), and concrete steps are:
(x y) tries to achieve x axial gradient figure I respectively on x direction of principal axis and y direction of principal axis to monocular infrared image I Gradx(x is y) with y axial gradient figure I Grady(x, y), then (x is y) at angle θ for monocular infrared image I lGrad G on the direction l(x, y)=I Gradx(x, y) * cos (θ l)+I Grady(x, y) * sin (θ l), wherein, L=0 ..., 7, calculate the gradient energy of the different directions of each image block of each pixel on other yardsticks except that first yardstick subsequently, wherein, i m the image block N of pixel on the j yardstick i(m) at angle θ lGradient energy on the direction does I m the image block N of pixel on the j yardstick then i(m) at angle θ lThe average of the gradient energy on the direction
Figure FDA00001795989000024
I m the image block N of pixel on the j yardstick i(m) at angle θ lGradient energy variance on the direction Wherein, size is image block N iThe number of the pixel that (m) is comprised, G l &OverBar; ( x , y ) = Mn N i ( m ) Lj .
4. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that; The sharpness of each image block of i pixel on second yardstick and the 3rd yardstick calculates through following steps in the said step (3), and concrete steps are: i m the image block N of pixel on the j yardstick i(m) sharpness Sharp N i ( m ) j = 1 Size &Sigma; x , y &Element; N i ( m ) ( I ( x , y ) - I &OverBar; ( x , y ) ) 2 , Wherein, (x y) is image block N to I i(m) contain the gray scale of pixel,
Figure FDA00001795989000028
Be image block N i(m) average of the gray scale of contained pixel,
Figure FDA00001795989000029
Size is image block N iThe number of the pixel that (m) is comprised.
5. according to claim 1ly it is characterized in that based on the monocular infrared image depth estimation method of optimizing the BP neural network model, in the said step (4) progressively linear regression analysis comprise following substep:
(411) related coefficient of each characteristic component and depth value in the calculated characteristics vector is according to the descending ordering that obtains each characteristic component to the effect of depth degree of the absolute value of related coefficient;
(412) the absolute value biggest characteristic component from related coefficient begins progressively to introduce regression equation; And do regression equation significance test; If do not think significantly that selected whole characteristic component all is not the principal element that influences depth value, if significantly again from depth value is influenced the descending regression equation of introducing one by one successively;
(413) new characteristic component of every introducing all need carry out significance test to each contained in regression equation characteristic component, and that significantly and to depth value is not influenced minimum characteristic component rejection in new regression equation, repeats
Till this step each characteristic component in regression equation is all remarkable;
(414) introduce a characteristic component that in the characteristic component of not introducing depth value is had the greatest impact, repeating step again
(413) and step (414), up to rejecting selected characteristic component, also can't introduce till the new characteristic component.
6. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that; What independent component analysis was used in the said step (4) is quick ICA algorithm; To analyzing through the proper vector of progressively linear regression analysis, make between each component independently as much as possible through quick ICA algorithm, comprise following substep:
(421) specifying the proper vector through M pixel after the linear regression analysis progressively is observation data X, and observation data X is carried out centralization, and making it average is 0;
(422), be about to become the albefaction vector Z, Z=W after observation data X projects to new subspace with the observation data X albefaction of centralization 0X, wherein, W 0Be albefaction matrix, W 0-1/2U T, Λ is the eigenvalue matrix of the covariance matrix of observation data X, U is the eigenvectors matrix of the covariance matrix of observation data X;
(423) upgrade W *Make W *=E{Zg (W TZ) }-E{g ' (W TZ) } W, wherein g () is a nonlinear function, and then to W *Standardization: W=W */ || W *||, if do not restrain and then do not repeat this step, wherein, selecting a mould is 1 the initial random weight vector initial value as W.
7. according to claim 1 based on the monocular infrared image depth estimation method of optimizing the BP neural network model; It is characterized in that initial weight and the threshold value of utilizing genetic algorithm to be optimized the BP neural network in the said step (5) also comprise following substep:
(511) make up initial BP neural network according to training sample, and confirm the number of input layer, hidden layer and the output layer of this network;
(512) carry out the initialization of genetic algorithm population; Wherein each individuality of population all is to be made up of the threshold value of the weights between the weights between input layer and the hidden layer, hidden layer and the output layer, hidden layer and threshold value four parts of output layer, and all weights and threshold value are listed as into a real number vector;
(513) maximum iteration time is set as stopping criterion for iteration, iterative process each time comprises selection, intersection and mutation operation, when reaching stopping criterion for iteration, output the last reign of a dynasty population optimum individual be the approximate optimal solution of initial weight and threshold value.
CN201210206701.1A 2012-06-21 2012-06-21 Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model Expired - Fee Related CN102750702B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201210206701.1A CN102750702B (en) 2012-06-21 2012-06-21 Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201210206701.1A CN102750702B (en) 2012-06-21 2012-06-21 Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model

Publications (2)

Publication Number Publication Date
CN102750702A true CN102750702A (en) 2012-10-24
CN102750702B CN102750702B (en) 2014-10-15

Family

ID=47030851

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201210206701.1A Expired - Fee Related CN102750702B (en) 2012-06-21 2012-06-21 Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model

Country Status (1)

Country Link
CN (1) CN102750702B (en)

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106408015A (en) * 2016-09-13 2017-02-15 电子科技大学成都研究院 Road fork identification and depth estimation method based on convolutional neural network
CN107204010A (en) * 2017-04-28 2017-09-26 中国科学院计算技术研究所 A kind of monocular image depth estimation method and system
CN107578436A (en) * 2017-08-02 2018-01-12 南京邮电大学 A kind of monocular image depth estimation method based on full convolutional neural networks FCN
CN107607207A (en) * 2017-08-22 2018-01-19 中国科学院深圳先进技术研究院 A kind of power equipment thermal fault detection method, system and electronic equipment
CN107690663A (en) * 2015-06-05 2018-02-13 谷歌有限责任公司 Albefaction neural net layer
CN107993239A (en) * 2017-12-25 2018-05-04 北京邮电大学 A kind of method and apparatus for the depth order for calculating monocular image
CN108335322A (en) * 2018-02-01 2018-07-27 深圳市商汤科技有限公司 Depth estimation method and device, electronic equipment, program and medium
CN109035319A (en) * 2018-07-27 2018-12-18 深圳市商汤科技有限公司 Monocular image depth estimation method and device, equipment, program and storage medium
CN109146944A (en) * 2018-10-30 2019-01-04 浙江科技学院 A kind of space or depth perception estimation method based on the revoluble long-pending neural network of depth
CN109377524A (en) * 2018-10-29 2019-02-22 山东师范大学 A kind of single image depth recovery method and system
CN109461177A (en) * 2018-09-29 2019-03-12 浙江科技学院 A kind of monocular image depth prediction approach neural network based
CN109598704A (en) * 2018-11-19 2019-04-09 电子科技大学 A kind of night soil microscopic image clarity evaluation method based on BP neural network
CN109903393A (en) * 2019-02-22 2019-06-18 清华大学 New Century Planned Textbook Scene Composition methods and device based on deep learning
CN109964237A (en) * 2016-09-15 2019-07-02 谷歌有限责任公司 Picture depth prediction neural network
CN112036470A (en) * 2020-08-28 2020-12-04 扬州大学 Cloud transmission-based multi-sensor fusion cucumber bemisia tabaci identification method
CN112215410A (en) * 2020-09-24 2021-01-12 国网福建省电力有限公司 Power load prediction method based on improved deep learning
CN112329189A (en) * 2019-07-30 2021-02-05 天津大学青岛海洋技术研究院 Transient electromagnetic useful signal extraction method based on independent quantity analysis

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100135581A1 (en) * 2008-12-02 2010-06-03 Samsung Electronics Co., Ltd. Depth estimation apparatus and method
CN101789123A (en) * 2010-01-27 2010-07-28 中国科学院半导体研究所 Method for creating distance map based on monocular camera machine vision

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100135581A1 (en) * 2008-12-02 2010-06-03 Samsung Electronics Co., Ltd. Depth estimation apparatus and method
CN101789123A (en) * 2010-01-27 2010-07-28 中国科学院半导体研究所 Method for creating distance map based on monocular camera machine vision

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
CLAUDIA HENDRIX ET AL.: "Relationship between monocular and binocular depth cues for judgements of spatial information and spatial instrument design", 《DISPLAYS》 *
TIEN-YING KUO ET AL.: "Depth Estimation from a Monocular Outdoor Image", 《2011 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS》 *
蓝建梁等: "基于多尺度纹理能量测度的单幅图像深度估计", 《计算机工程与设计》 *
谷小婧: "基于图像分析的自然彩色夜视成像方法研究", 《中国博士学位论文全文数据库》 *

Cited By (33)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107690663A (en) * 2015-06-05 2018-02-13 谷歌有限责任公司 Albefaction neural net layer
CN107690663B (en) * 2015-06-05 2022-04-12 渊慧科技有限公司 Whitening neural network layer
CN106408015A (en) * 2016-09-13 2017-02-15 电子科技大学成都研究院 Road fork identification and depth estimation method based on convolutional neural network
CN109964237B (en) * 2016-09-15 2020-07-17 谷歌有限责任公司 Image depth prediction neural network
CN109964237A (en) * 2016-09-15 2019-07-02 谷歌有限责任公司 Picture depth prediction neural network
US10929996B2 (en) 2016-09-15 2021-02-23 Google Llc Image depth prediction neural networks
US11734847B2 (en) 2016-09-15 2023-08-22 Google Llc Image depth prediction neural networks
CN107204010A (en) * 2017-04-28 2017-09-26 中国科学院计算技术研究所 A kind of monocular image depth estimation method and system
CN107204010B (en) * 2017-04-28 2019-11-19 中国科学院计算技术研究所 A kind of monocular image depth estimation method and system
CN107578436A (en) * 2017-08-02 2018-01-12 南京邮电大学 A kind of monocular image depth estimation method based on full convolutional neural networks FCN
CN107607207A (en) * 2017-08-22 2018-01-19 中国科学院深圳先进技术研究院 A kind of power equipment thermal fault detection method, system and electronic equipment
CN107607207B (en) * 2017-08-22 2019-06-28 中国科学院深圳先进技术研究院 A kind of power equipment thermal fault detection method, system and electronic equipment
CN107993239B (en) * 2017-12-25 2022-04-12 北京邮电大学 Method and device for calculating depth order of monocular image
CN107993239A (en) * 2017-12-25 2018-05-04 北京邮电大学 A kind of method and apparatus for the depth order for calculating monocular image
CN108335322A (en) * 2018-02-01 2018-07-27 深圳市商汤科技有限公司 Depth estimation method and device, electronic equipment, program and medium
US11308638B2 (en) 2018-02-01 2022-04-19 Shenzhen Sensetime Technology Co., Ltd. Depth estimation method and apparatus, electronic device, program, and medium
CN108335322B (en) * 2018-02-01 2021-02-12 深圳市商汤科技有限公司 Depth estimation method and apparatus, electronic device, program, and medium
US11443445B2 (en) 2018-07-27 2022-09-13 Shenzhen Sensetime Technology Co., Ltd. Method and apparatus for depth estimation of monocular image, and storage medium
JP2021500689A (en) * 2018-07-27 2021-01-07 深▲せん▼市商▲湯▼科技有限公司Shenzhen Sensetime Technology Co., Ltd. Monocular image depth estimation method and equipment, equipment, programs and storage media
CN109035319A (en) * 2018-07-27 2018-12-18 深圳市商汤科技有限公司 Monocular image depth estimation method and device, equipment, program and storage medium
CN109035319B (en) * 2018-07-27 2021-04-30 深圳市商汤科技有限公司 Monocular image depth estimation method, monocular image depth estimation device, monocular image depth estimation apparatus, monocular image depth estimation program, and storage medium
CN109461177A (en) * 2018-09-29 2019-03-12 浙江科技学院 A kind of monocular image depth prediction approach neural network based
CN109461177B (en) * 2018-09-29 2021-12-10 浙江科技学院 Monocular image depth prediction method based on neural network
CN109377524A (en) * 2018-10-29 2019-02-22 山东师范大学 A kind of single image depth recovery method and system
CN109377524B (en) * 2018-10-29 2021-02-23 山东师范大学 Method and system for recovering depth of single image
CN109146944A (en) * 2018-10-30 2019-01-04 浙江科技学院 A kind of space or depth perception estimation method based on the revoluble long-pending neural network of depth
CN109146944B (en) * 2018-10-30 2020-06-26 浙江科技学院 Visual depth estimation method based on depth separable convolutional neural network
CN109598704A (en) * 2018-11-19 2019-04-09 电子科技大学 A kind of night soil microscopic image clarity evaluation method based on BP neural network
CN109903393A (en) * 2019-02-22 2019-06-18 清华大学 New Century Planned Textbook Scene Composition methods and device based on deep learning
CN112329189A (en) * 2019-07-30 2021-02-05 天津大学青岛海洋技术研究院 Transient electromagnetic useful signal extraction method based on independent quantity analysis
CN112036470A (en) * 2020-08-28 2020-12-04 扬州大学 Cloud transmission-based multi-sensor fusion cucumber bemisia tabaci identification method
CN112215410A (en) * 2020-09-24 2021-01-12 国网福建省电力有限公司 Power load prediction method based on improved deep learning
CN112215410B (en) * 2020-09-24 2023-01-10 国网福建省电力有限公司 Power load prediction method based on improved deep learning

Also Published As

Publication number Publication date
CN102750702B (en) 2014-10-15

Similar Documents

Publication Publication Date Title
CN102750702B (en) Monocular infrared image depth estimation method based on optimized BP (Back Propagation) neural network model
CN102708569B (en) Based on the monocular infrared image depth estimation method of SVM model
CN108038445B (en) SAR automatic target identification method based on multi-view deep learning framework
CN109614985B (en) Target detection method based on densely connected feature pyramid network
CN112766199B (en) Hyperspectral image classification method based on self-adaptive multi-scale feature extraction model
CN107132516B (en) A kind of Radar range profile&#39;s target identification method based on depth confidence network
CN105869178A (en) Method for unsupervised segmentation of complex targets from dynamic scene based on multi-scale combination feature convex optimization
CN107247938A (en) A kind of method of high-resolution remote sensing image City Building function classification
CN110110621A (en) The oblique photograph point cloud classifications method of deep learning model is integrated based on multiple features
CN105787557A (en) Design method of deep nerve network structure for computer intelligent identification
CN112862090B (en) Air temperature forecasting method based on deep space-time neural network
CN112347970B (en) Remote sensing image ground object identification method based on graph convolution neural network
CN114092697B (en) Building facade semantic segmentation method with attention fused with global and local depth features
CN110852393A (en) Remote sensing image segmentation method and system
CN109784288B (en) Pedestrian re-identification method based on discrimination perception fusion
CN105095913B (en) The Classifying Method in Remote Sensing Image and system represented based on neighbour&#39;s canonical joint sparse
CN106991411B (en) Remote Sensing Target based on depth shape priori refines extracting method
CN114758252A (en) Image-based distributed photovoltaic roof resource segmentation and extraction method and system
CN112766280A (en) Remote sensing image road extraction method based on graph convolution
CN104318215B (en) A kind of cross-view face identification method based on domain robust convolution feature learning
CN116206185A (en) Lightweight small target detection method based on improved YOLOv7
CN105701493A (en) Methods and systems for image matting and foreground estimation based on hierarchical graphs
CN104680167B (en) Auroral oval location determining method based on deep learning
Simon et al. Parameter-free/pareto-driven procedural 3d reconstruction of buildings from ground-level sequences
CN104850677A (en) Deep learning cellular automaton simulation analysis method based on geographic information system

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20141015

Termination date: 20170621

CF01 Termination of patent right due to non-payment of annual fee