CN103279759A - Vehicle front trafficability analyzing method based on convolution nerve network - Google Patents

Vehicle front trafficability analyzing method based on convolution nerve network Download PDF

Info

Publication number
CN103279759A
CN103279759A CN2013102341260A CN201310234126A CN103279759A CN 103279759 A CN103279759 A CN 103279759A CN 2013102341260 A CN2013102341260 A CN 2013102341260A CN 201310234126 A CN201310234126 A CN 201310234126A CN 103279759 A CN103279759 A CN 103279759A
Authority
CN
China
Prior art keywords
image
pixel
gray
layer
scale value
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN2013102341260A
Other languages
Chinese (zh)
Other versions
CN103279759B (en
Inventor
李琳辉
连静
王蒙蒙
丁新立
宗云鹏
化玉伟
王宏旭
常静
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Dalian University of Technology
Original Assignee
Dalian University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Dalian University of Technology filed Critical Dalian University of Technology
Priority to CN201310234126.0A priority Critical patent/CN103279759B/en
Publication of CN103279759A publication Critical patent/CN103279759A/en
Application granted granted Critical
Publication of CN103279759B publication Critical patent/CN103279759B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Image Analysis (AREA)
  • Traffic Control Systems (AREA)

Abstract

The invention discloses a vehicle front trafficability analyzing method based on a convolution nerve network. The method comprises the following steps: first, a vidicon arranged in the front of a vehicle is used for collecting a large number of actual vehicle traveling environment images; a Gamma rectification function is used for pre-processing the images; training of the convolution nerve network is conducted. According to the method, a nonlinear function superimposed Gamma rectification method is used for pre-processing the images, so that influence of light illumination of strong changes on identification of objects is avoided, and the image resolution ratio is improved. According to the method, a geometry normalization method is used, so that the resolution ratio difference caused by identifying the distance of an object distance vidicon is reduced. The convolution nerve network LeNet-5 adopted in the method can extract implicit expression characteristics with class distinguishing capacity and is simple in extracting process. The LeNet-5 is combined with a local receptive field, weight share and secondary sampling to ensure robustness of simple geometry deformation, reduce training parameters of the network, and simplify the structure of the network.

Description

A kind of vehicle front trafficability analytical approach based on convolutional neural networks
Technical field
The invention belongs to safe driver assistance and intelligent transport technology field, relate to vehicle front trafficability analytical approach, specially refer to a kind of by camera acquisition vehicle front video image, based on the vehicle front trafficability analytical approach of convolutional neural networks.
Background technology
The analysis of vehicle front trafficability belongs to the environment sensing category of intelligent transportation field outside car, refer to based on advanced meanses such as sensor technology, computer technology or mechanicss of communication the driving safety of institute's acquisition environment be analyzed, find out the potential safety hazard of existence, send prompting and early warning or lay the foundation for the automatic driving vehicle navigation to the driver.At present, based on camera acquisition vehicle front video image information, the research of adopting the visual pattern understanding method to carry out the trafficability analysis mainly contains detection of obstacles, pedestrian detection, vehicle detection, Road Detection, traffic sign detection, topography and geomorphology classification etc.
Trafficability is analyzed related visual pattern understanding method and can be divided into based on the method for rebuilding with based on the method for identifying.Wherein, be based on three-dimensional or 2.5 dimension reconstruction techniques based on the method for rebuilding, make judgement from the angle in space to whether passing through, problems such as the less and real-time difference of serious polysemy, the reconstruction scope that is difficult to avoid three-dimensional reconstruction intrinsic.In the image understanding method based on identification, mainly contain algorithm based on modeling and template matches, general neural network, support vector machine, self-supervisory study, based on the method for statistical learning etc., these methods need be extracted the explicit features of target, the leaching process complexity, easily cause important information to lose, adaptive capacity to environment is poor.
For the strong structured road environment of illumination variation, if directly original image is identified, interfere information is many, the process complexity that explicit features is extracted, furthermore target can cause the difference of resolution apart from difference apart from video camera.In addition, the variation of illumination can influence picture quality, reduces the resolution of image.
Summary of the invention
For solving the problems referred to above that prior art exists, the present invention will propose a kind of vehicle front trafficability analytical approach based on convolutional neural networks, this method can be extracted the implicit features of target, leaching process is simple, avoid reducing the resolution of image, and can reduce the influence of illumination, be applicable to the structured road environment that illumination variation is strong.
Technical scheme of the present invention is: a kind of vehicle front trafficability analytical approach based on convolutional neural networks may further comprise the steps:
A, image acquisition
At first by a large amount of real vehicle running environment images of the camera acquisition that is installed in vehicle front, described image has m * n pixel; Obtain the zone of image bottom 3/5ths as area-of-interest by cutting then; At last images cut is converted into gray level image.
B, image pre-service
B1, the Gamma of method construct that utilizes nonlinear function to superpose correct function, and the gray level image that steps A obtains is corrected, and concrete functional expression is as follows:
G(x)=1+f 1(x)+f 2(x)+f 3(x) (1)
f 1(x)=acos(πx/255) (2)
f 2(x)=(K(x)+b)cosβ+xsinα (3)
K(x)=ρsin(4πx/255) (4)
α=arctan(-2b/255) (5)
f 3(x)=R(x)cos(3πx/255) (6)
R(x)=c|2x/255-1| (7)
In the formula, x is a certain gray values of pixel points, the Gamma correction value of a certain gray-scale value correspondence of G (x) representative, and a ∈ (0,1) is a weighting coefficient, b represents f 2(x) maximum changing range, ρ are represented the amplitude of K (x), and α represents the deflection angle of K (x), and c represents the amplitude of R (x), and satisfy a+b+c<1.
Gray-scale value computing formula after Gamma corrects is:
g(x)=255(x/255) 1/G(x) (8)
In the formula, a certain gray values of pixel points after g (x) representative is corrected through Gamma.
Correct through Gamma, obtain gray level image P.
B2, at gray level image P, change some gray values of pixel points, concrete change method is as follows:
Choose in the image gray-scale value in the image-region except vehicle and road boundary and be 0 pixel, change its gray-scale value into 1, choose in the image gray-scale value in the image-region except vehicle and road boundary and be 255 pixel, change its gray-scale value into 254; Change vehicle region gray values of pixel points in the image into 0, the gray-scale value of road boundary area pixel point changes 255 into, and the image behind the change pixel is gray level image Q.So far, the pixel of gray level image Q comprises three classes: the first kind is that gray-scale value is 0 pixel, represents vehicle; Second class is that gray-scale value is 255 pixel, represents road boundary; The 3rd class is that to remove gray-scale value be pixel outside 0 and 255, represents the road surface.Give corresponding label respectively with above three class pixels, be about to label " 0 " and compose expression " vehicle " to first kind pixel, label " 1 " is composed to the second class pixel, expression " road boundary " composes label " 2 " to the 3rd class pixel, expression " road surface ".At last the label of each pixel among the gray level image Q is composed to corresponding pixel among the gray level image P.
B3, carry out normalized at the size of gray level image P:
B31, along the picture altitude direction, at interval choose different pixel columns, represent with x, measure by actual samples, obtain the pixel wide of the corresponding target of different pixels row x institute and highly;
B32, being that 0~32 image-region is the reference picture zone with pixels tall in the image, is benchmark with pixel wide W and the height H of the target that will identify in this reference picture zone, and the horizontal and vertical cutting scale-up factor of namely establishing this reference picture zone is 1; With W and H remove respectively target on all the other each pixel columns width and the height, obtain two groups of ratios, represent with Y and Z respectively;
B33, at last pixel column x and two groups of ratio Y and Z are carried out match respectively, obtain two matched curves, as follows:
Y=k 1x+b 1 (9)
Z=k 2x+b 2 (10)
Wherein, the horizontal cutting scale-up factor of Y representative image, the vertical cutting scale-up factor of Z representative image, a certain pixel column of x representative image, k 1, k 2Represent the slope of two matched curves respectively, b 1, b 2Represent the intercept of two matched curves respectively.
B34, the horizontal and vertical cutting ratio in reference picture zone all is made as 1, is about to the image pattern that the reference picture zone is cut to 32 * 32 pixels.Along with the increase of x, according to the also corresponding increase of horizontal and vertical cutting size of formula (9), (10) acquisition.By cutting, obtain a series of image patterns not of uniform size, the image pattern unification that at last cutting is obtained is scaled the image of 32 * 32 pixels.With the image of 32 * 32 pixels that the obtain training sample as convolutional neural networks.
The training of C, convolutional neural networks
Typical case's convolutional neural networks LeNet-5 forms by 8 layers, and input is that layer is 32 * 32 pixel images; Network layer C1, C3 and C5 represent convolutional layer respectively, and network layer S2 and S4 are time sampling layer, and network layer F5 is full articulamentum, and the neuronic number of output layer is identical with the target classification number that will identify, changes according to actual application environment.A face of every layer network represents a characteristic pattern, the set that the neuron that this characteristic pattern is shared by weights in each layer is formed.The neuron of each layer only is connected with the neuron of a local receptive field of last layer.
The general type of convolutional layer is:
x j l = f ( Σ i ∈ M j x i l - 1 * k ij l + b j l ) - - - ( 11 )
In the formula, { 1,2,3,4,5,6,7,8} represents the number of plies to l ∈, and k is convolution kernel, M jRepresent the selection of input feature vector figure, b represents biasing.
The general type of inferior sampling layer is:
x j l = f ( β j l down ( x j l - 1 ) + b j l ) - - - ( 12 )
In the formula, down () expression time sampling function generally is the zone summation to a n * n of last tomographic image, and β represents time weights of sampling layer, and b represents biasing.
According to actual application environment, the number of the output neuron of LeNet-5 to be adjusted, the Pixel Dimensions that adopts step B to obtain then is that 32 * 32 image pattern is trained.By training, when the output valve of convolutional neural networks and the error of expectation value are in the tolerance interval, just obtain can be used for the convolutional neural networks of vehicle front trafficability analysis.
Compared with prior art, effect of the present invention and benefit are:
1, the present invention adopts the Gamma antidote pretreatment image of nonlinear function stack, has avoided the illumination of strong variations to the influence of target identification, has improved image resolution ratio.
2, the present invention has adopted the geometrical normalization method, has reduced the differences in resolution that identification target range video camera distance causes.
3, the convolutional neural networks LeNet-5 of the present invention's employing can extract the implicit features with classification resolution characteristic, and leaching process is simple; LeNet-5 shares and inferior sampling in conjunction with local receptive field, weights, guarantees the robustness to the simple geometry distortion, and has reduced the training parameter of network, has simplified network structure; The neuron number of LeNet-5 output layer can be adjusted according to actual application environment, and adaptive capacity to environment is strong.
Description of drawings
3 in the total accompanying drawing of the present invention, wherein:
Fig. 1 is based on the vehicle front trafficability analytical approach process flow diagram of convolutional neural networks
Fig. 2 convolutional neural networks LeNet-5 structural drawing.
The sample set of Fig. 3 convolutional neural networks training.
Embodiment
The present invention is further described below in conjunction with accompanying drawing.Be illustrated in figure 1 as the process flow diagram based on the vehicle front trafficability analytical approach of convolutional neural networks.The present invention is example with the structured environment of highway, and environment before the car is divided into vehicle, road boundary and road surface.
Analytic process of the present invention comprises: the training of image acquisition, image pre-service, convolutional neural networks.
A, image acquisition
Camera acquisition by being installed in vehicle front a large amount of real highway running environment images (640 * 480 pixel), then with 3/5ths parts under the image as area-of-interest (640 * 288 pixel), to reduce follow-up workload; At last images cut is converted into gray level image.
B, image pre-service
The first step, the Gamma antidote has certain advantage aspect the minimizing illumination effect.Generally speaking, when the Gamma value greater than 1 the time, the high light part of image is compressed and shadow partly is expanded; When the Gamma value less than 1 the time, the high light of image partly is expanded and the shadow part is compressed.Utilize Gamma of method construct of nonlinear function stack to correct function, the gray level image that steps A obtains is corrected, functional expression is as follows:
G(x)=1+f 1(x)+f 2(x)+f 3(x) (1)
f 1(x)=acos(πx/255) (2)
f 2(x)=(K(x)+b)cosβ+xsinα (3)
K(x)=ρsin(4πx/255) (4)
α=arctan(-2b/255) (5)
f 3(x)=R(x)cos(3πx/255) (6)
R(x)=c|2x/255-1| (7)
In the formula, x is a certain gray values of pixel points, the Gamma correction value of a certain gray-scale value correspondence of G (x) representative, and a ∈ (0,1) is a weighting coefficient, b represents f 2(x) maximum changing range, ρ are represented the amplitude of K (x), and α represents the deflection angle of K (x), and c represents the amplitude of R (x), and satisfy a+b+c<1.Generally speaking, a<b≤c, desirable a=0.2, b=0.3, c=0.3.
Gray-scale value computing formula after Gamma corrects is:
g(x)=255(x/255) 1/G(x) (8)
In the formula, a certain gray values of pixel points after g (x) representative is corrected through Gamma.
Correct through Gamma, obtain gray level image P.
Second step, at the gray level image P that corrects through Gamma, choose the gray-scale value except vehicle and road boundary in the image and be 0 pixel, change these gray values of pixel points into 1 by programming, choose the gray-scale value except vehicle and road boundary in the image and be 255 pixel, change these gray values of pixel points into 254 by programming; Change vehicle region gray values of pixel points in the image into 0 by programming then, the gray-scale value of road boundary area pixel point changes 255 into, obtains gray level image Q.So far, the pixel of gray level image Q comprises three classes: the first kind is that gray-scale value is 0 pixel, represents vehicle; Second class is that gray-scale value is 255 pixel, represents road boundary; The 3rd class is that to remove gray-scale value be pixel outside 0 and 255, represents the road surface.Give corresponding label respectively by being programmed for above three class pixels, be about to label " 0 " and compose expression " vehicle " to first kind pixel, label " 1 " is composed to the second class pixel, expression " road boundary " composes label " 2 " to the 3rd class pixel, expression " road surface ".At last the label of each pixel among the image Q is composed the respective pixel point of giving among the gray level image P by programming.
In the 3rd step, carry out normalized at the size of gray level image P:
In image, the shared number of pixels of same target is subjected to himself very big apart from the influence of video camera distance, and namely the shared number of pixels of arbitrary target is inversely proportional to the distance of itself and video camera.The present invention proposes a kind of new geometrical normalization method, carries out normalized at the size of gray level image, avoids reducing the resolution of image, reduces the differences in resolution that identification target range video camera distance causes.At first, on picture altitude, choose different pixel columns at interval, represent with vector x, x={15 32 60 65 68 72 75 82 85 87 92 96 100 108 111 113 124 130 138 143 150 160}, measure by actual samples, obtain the pixel wide of the corresponding target of different pixels row institute and highly; Secondly, be that 0~32 image-region is reference with pixels tall in the image, be benchmark with pixel wide W and the height H of target in this reference picture zone, namely the horizontal and vertical cutting scale-up factor in this reference picture zone is 1; At last, remove width and the height of target on all the other each pixel columns respectively with W and H, obtain two groups of ratios, represent with Y and Z respectively, by programming pixel column x data and Y and Z data are carried out match respectively then, obtain two matched curves, as follows:
Y=0.0312x-0.8339 (9)
Z=0.0360x-1.0590 (10)
Wherein, the horizontal cutting scale-up factor of Y representative image, the vertical cutting scale-up factor of Z representative image, a certain pixel column of x representative image.
The horizontal and vertical cutting ratio in reference picture zone is 1, is about to the image pattern that the reference picture zone is cut to 32 * 32 pixels.Along with the increase of x, according to the also corresponding increase of horizontal and vertical cutting size of formula (9), (10) acquisition.By cutting, obtain a series of image patterns not of uniform size, the image pattern that cutting is obtained is by the unified image that is scaled 32 * 32 pixels of programming at last.With the image of 32 * 32 pixels that the obtain training sample as convolutional neural networks.
The training of C, convolutional neural networks
The structure that the present invention adopts typical convolutional neural networks LeNet-5 as shown in Figure 2.Typical case's convolutional neural networks LeNet-5 forms by 8 layers, and input picture is 32 * 32 pixels; Network layer C1, C3, C5 represent convolutional layer, and network layer S2, S4 are time sampling layer, and network layer F5 is full articulamentum, and the neuronic number of output layer is identical with the target classification number that will identify, can change according to actual application environment.A face of every layer network represents a characteristic pattern, the set that the neuron that this characteristic pattern is shared by weights in each layer is formed.The neuron of each layer only is connected with the neuron of a local receptive field of last layer (counting from input layer).
Convolutional layer C1 is that 28 * 28 characteristic pattern is formed by 6 sizes, and each neuron of characteristic pattern is connected with one 5 * 5 neighborhood of input picture, can train and is connected with 122304 but convolutional layer C1 comprises 156 training parameters.Inferior sampling layer S2 is that 14 * 14 characteristic pattern is formed by 6 sizes, and size is that 2 * 2 neighborhood links to each other among each neuron of characteristic pattern and the convolutional layer C1, but an inferior sampling layer S2 has 12 training parameters can train and be connected with 5880.Convolutional layer C3 is that 10 * 10 characteristic pattern is formed by 16 sizes, and each neuron of characteristic pattern links to each other with time one 5 * 5 the neighborhood of sampling layer S2, can train and is connected with 151600 but convolutional layer C3 includes 1516 training parameters.Inferior sampling layer S4 is that 5 * 5 characteristic pattern is formed by 16 sizes, and each neuron of characteristic pattern and the size of convolutional layer C3 are that 2 * 2 neighborhood is connected, and can train and is connected with 2000 but an inferior sampling layer S4 includes 32 training parameters.Convolutional layer C5 is made up of 120 characteristic patterns, and each neuron of characteristic pattern is connected with time neighborhood of 5 * 5 sizes of all characteristic patterns of sampling layer S4, can train and is connected with 48120 but convolutional layer C5 includes 48120 training parameters.Network layer F6 is connected entirely with convolutional layer C5, but full articulamentum F6 comprises 10164 training parameters.Output layer is made up of the radial basis function unit.
The general type of convolutional layer is:
x j l = f ( Σ i ∈ M j x i l - 1 * k ij l + b j l ) - - - ( 11 )
In the formula, l represents the number of plies, and k is convolution kernel, M jRepresent the selection of input feature vector figure, b represents biasing.
The general type of inferior sampling layer is:
x j l = f ( β j l down ( x j l - 1 ) + b j l ) - - - ( 12 )
In the formula, down () expression time sampling function generally is the zone summation to a n * n of last tomographic image, and β represents time weights of sampling layer, and b represents biasing.
The present invention is example with the highway environment, environment is divided into road surface, vehicle and road boundary three parts before the car, therefore the output neuron number of LeNe-5 should be decided to be 3, and output 0 representative identification target is vehicle, output 1 representative identification target is road boundary, and output 2 representative identification targets are the road surface.The sample set size of network training is 5000, and the part sample as shown in Figure 3.Choosing of network initial weight adopts random approach to produce.With this known classification mode the convolution network is trained, network has just possessed the mapping ability between the inputoutput pair.By training, if satisfy the output valve of convolutional neural networks and the error of expectation value is in the tolerance interval, just obtained can be used for the convolutional neural networks of vehicle front trafficability analysis.

Claims (1)

1. vehicle front trafficability analytical approach based on convolutional neural networks is characterized in that: may further comprise the steps:
A, image acquisition
At first by a large amount of real vehicle running environment images of the camera acquisition that is installed in vehicle front, described image has m * n pixel; Obtain the zone of image bottom 3/5ths as area-of-interest by cutting then; At last images cut is converted into gray level image;
B, image pre-service
B1, the Gamma of method construct that utilizes nonlinear function to superpose correct function, and the gray level image that steps A obtains is corrected, and concrete functional expression is as follows:
G(x)=1+f 1(x)+f 2(x)+f 3(x) (1)
f 1(x)=acos(πx/255) (2)
f 2(x)=(K(x)+b)cosβ+xsinα (3)
K(x)=ρsin(4πx/255) (4)
α=arctan(-2b/255) (5)
f 3(x)=R(x)cos(3πx/255) (6)
R(x)=c|2x/255-1| (7)
In the formula, x is a certain gray values of pixel points, the Gamma correction value of a certain gray-scale value correspondence of G (x) representative, and a ∈ (0,1) is a weighting coefficient, b represents f 2(x) maximum changing range, ρ are represented the amplitude of K (x), and α represents the deflection angle of K (x), and c represents the amplitude of R (x), and satisfy a+b+c<1;
Gray-scale value computing formula after Gamma corrects is:
g(x)=255(x/255) 1/G(x) (8)
In the formula, a certain gray values of pixel points after g (x) representative is corrected through Gamma;
Correct through Gamma, obtain gray level image P;
B2, B2, at gray level image P, change some gray values of pixel points, concrete change method is as follows:
Choose in the image gray-scale value in the image-region except vehicle and road boundary and be 0 pixel, change its gray-scale value into 1, choose in the image gray-scale value in the image-region except vehicle and road boundary and be 255 pixel, change its gray-scale value into 254; Change vehicle region gray values of pixel points in the image into 0, the gray-scale value of road boundary area pixel point changes 255 into, and the image behind the change pixel is gray level image Q; So far, the pixel of gray level image Q comprises three classes: the first kind is that gray-scale value is 0 pixel, represents vehicle; Second class is that gray-scale value is 255 pixel, represents road boundary; The 3rd class is that to remove gray-scale value be pixel outside 0 and 255, represents the road surface; Give corresponding label respectively with above three class pixels, be about to label " 0 " and compose expression " vehicle " to first kind pixel, label " 1 " is composed to the second class pixel, expression " road boundary " composes label " 2 " to the 3rd class pixel, expression " road surface "; At last the label of each pixel among the gray level image Q is composed to corresponding pixel among the gray level image P;
B3, carry out normalized at the size of gray level image P:
B31, along the picture altitude direction, at interval choose different pixel columns, represent with x, measure by actual samples, obtain the pixel wide of the corresponding target of different pixels row x institute and highly;
B32, being that 0~32 image-region is the reference picture zone with pixels tall in the image, is benchmark with pixel wide W and the height H of the target that will identify in this reference picture zone, and the horizontal and vertical cutting scale-up factor of namely establishing this reference picture zone is 1; With W and H remove respectively target on all the other each pixel columns width and the height, obtain two groups of ratios, represent with Y and Z respectively;
B33, at last pixel column x and two groups of ratio Y and Z are carried out match respectively, obtain two matched curves, as follows:
Y=k 1x+b 1 (9)
Z=k 2x+b 2 (10)
Wherein, the horizontal cutting scale-up factor of Y representative image, the vertical cutting scale-up factor of Z representative image, a certain pixel column of x representative image, k 1, k 2Represent the slope of two matched curves respectively, b 1, b 2Represent the intercept of two matched curves respectively;
B34, the horizontal and vertical cutting ratio in reference picture zone all is made as 1, is about to the image pattern that the reference picture zone is cut to 32 * 32 pixels; Along with the increase of x, according to the also corresponding increase of horizontal and vertical cutting size of formula (9), (10) acquisition; By cutting, obtain a series of image patterns not of uniform size, the image pattern unification that at last cutting is obtained is scaled the image of 32 * 32 pixels; With the image of 32 * 32 pixels that the obtain training sample as convolutional neural networks;
The training of C, convolutional neural networks
Typical case's convolutional neural networks LeNet-5 forms by 8 layers, and input is that layer is 32 * 32 pixel images; Network layer C1, C3 and C5 represent convolutional layer respectively, and network layer S2 and S4 are time sampling layer, and network layer F5 is full articulamentum, and the neuronic number of output layer is identical with the target classification number that will identify, changes according to actual application environment; A face of every layer network represents a characteristic pattern, the set that the neuron that this characteristic pattern is shared by weights in each layer is formed; The neuron of each layer only is connected with the neuron of a local receptive field of last layer;
The general type of convolutional layer is:
x j l = f ( Σ i ∈ M j x i l - 1 * k ij l + b j l ) - - - ( 11 )
In the formula, { 1,2,3,4,5,6,7,8} represents the number of plies to l ∈, and k is convolution kernel, M jRepresent the selection of input feature vector figure, b represents biasing;
The general type of inferior sampling layer is:
x j l = f ( β j l down ( x j l - 1 ) + b j l ) - - - ( 12 )
In the formula, down () expression time sampling function generally is the zone summation to a n * n of last tomographic image, and β represents time weights of sampling layer, and b represents biasing;
According to actual application environment, the number of the output neuron of LeNet-5 to be adjusted, the Pixel Dimensions that adopts step B to obtain then is that 32 * 32 image pattern is trained; By training, when the output valve of convolutional neural networks and the error of expectation value are in the tolerance interval, just obtain can be used for the convolutional neural networks of vehicle front trafficability analysis.
CN201310234126.0A 2013-06-09 2013-06-09 A kind of vehicle front trafficability analytical procedure based on convolutional neural networks Expired - Fee Related CN103279759B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310234126.0A CN103279759B (en) 2013-06-09 2013-06-09 A kind of vehicle front trafficability analytical procedure based on convolutional neural networks

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310234126.0A CN103279759B (en) 2013-06-09 2013-06-09 A kind of vehicle front trafficability analytical procedure based on convolutional neural networks

Publications (2)

Publication Number Publication Date
CN103279759A true CN103279759A (en) 2013-09-04
CN103279759B CN103279759B (en) 2016-06-01

Family

ID=49062274

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310234126.0A Expired - Fee Related CN103279759B (en) 2013-06-09 2013-06-09 A kind of vehicle front trafficability analytical procedure based on convolutional neural networks

Country Status (1)

Country Link
CN (1) CN103279759B (en)

Cited By (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103544506A (en) * 2013-10-12 2014-01-29 Tcl集团股份有限公司 Method and device for classifying images on basis of convolutional neural network
CN104063719A (en) * 2014-06-27 2014-09-24 深圳市赛为智能股份有限公司 Method and device for pedestrian detection based on depth convolutional network
CN104077577A (en) * 2014-07-03 2014-10-01 浙江大学 Trademark detection method based on convolutional neural network
CN104182756A (en) * 2014-09-05 2014-12-03 大连理工大学 Method for detecting barriers in front of vehicles on basis of monocular vision
CN104680508A (en) * 2013-11-29 2015-06-03 华为技术有限公司 Convolutional neural network and target object detection method based on convolutional neural network
CN104809426A (en) * 2014-01-27 2015-07-29 日本电气株式会社 Convolutional neural network training method and target identification method and device
CN104850864A (en) * 2015-06-01 2015-08-19 深圳英智源智能系统有限公司 Unsupervised image recognition method based on convolutional neural network
CN104992179A (en) * 2015-06-23 2015-10-21 浙江大学 Fine-grained convolutional neural network-based clothes recommendation method
CN105224963A (en) * 2014-06-04 2016-01-06 华为技术有限公司 The method of changeable degree of depth learning network structure and terminal
CN105320495A (en) * 2014-07-22 2016-02-10 英特尔公司 Weight-shifting mechanism for convolutional neural network
CN105691381A (en) * 2016-03-10 2016-06-22 大连理工大学 Stability control method and system for electric automobile with four independently driven wheels
WO2016095117A1 (en) * 2014-12-17 2016-06-23 Nokia Technologies Oy Object detection with neural network
CN105975915A (en) * 2016-04-28 2016-09-28 大连理工大学 Front vehicle parameter identification method based on multitask convolution nerve network
WO2016149881A1 (en) * 2015-03-20 2016-09-29 Intel Corporation Object recogntion based on boosting binary convolutional neural network features
CN106205126A (en) * 2016-08-12 2016-12-07 北京航空航天大学 Large-scale Traffic Network based on convolutional neural networks is blocked up Forecasting Methodology and device
CN106407931A (en) * 2016-09-19 2017-02-15 杭州电子科技大学 Novel deep convolution neural network moving vehicle detection method
CN106599832A (en) * 2016-12-09 2017-04-26 重庆邮电大学 Method for detecting and recognizing various types of obstacles based on convolution neural network
CN106874296A (en) * 2015-12-14 2017-06-20 阿里巴巴集团控股有限公司 A kind of style recognition methods of commodity and device
CN107392189A (en) * 2017-09-05 2017-11-24 百度在线网络技术(北京)有限公司 For the method and apparatus for the driving behavior for determining unmanned vehicle
WO2017206066A1 (en) * 2016-05-31 2017-12-07 Nokia Technologies Oy Method and apparatus for detecting small objects with an enhanced deep neural network
CN108122234A (en) * 2016-11-29 2018-06-05 北京市商汤科技开发有限公司 Convolutional neural networks training and method for processing video frequency, device and electronic equipment
CN108496178A (en) * 2016-01-05 2018-09-04 御眼视觉技术有限公司 System and method for estimating Future Path
CN108596115A (en) * 2018-04-27 2018-09-28 济南浪潮高新科技投资发展有限公司 A kind of vehicle checking method, apparatus and system based on convolutional neural networks
TWI638332B (en) * 2016-11-29 2018-10-11 財團法人車輛研究測試中心 Hierarchical object detection system with parallel architecture and method thereof
CN108921003A (en) * 2018-04-26 2018-11-30 东华大学 Unmanned plane obstacle detection method based on convolutional neural networks and morphological image
WO2019001346A1 (en) * 2017-06-26 2019-01-03 Huawei Technologies Co., Ltd. System and methods for object filtering and uniform representation for autonomous systems
WO2019047649A1 (en) * 2017-09-05 2019-03-14 百度在线网络技术(北京)有限公司 Method and device for determining driving behavior of unmanned vehicle
CN109657643A (en) * 2018-12-29 2019-04-19 百度在线网络技术(北京)有限公司 A kind of image processing method and device
CN109726615A (en) * 2017-10-30 2019-05-07 北京京东尚科信息技术有限公司 A kind of recognition methods of road boundary and device
CN109961509A (en) * 2019-03-01 2019-07-02 北京三快在线科技有限公司 Dimensionally map generalization and model training method, device and electronic equipment
CN110084190A (en) * 2019-04-25 2019-08-02 南开大学 Unstructured road detection method in real time under a kind of violent light environment based on ANN
CN110399929A (en) * 2017-11-01 2019-11-01 腾讯科技(深圳)有限公司 Eye fundus image classification method, device and computer readable storage medium
CN111222476A (en) * 2020-01-10 2020-06-02 北京百度网讯科技有限公司 Video time sequence action detection method and device, electronic equipment and storage medium
CN113325948A (en) * 2020-02-28 2021-08-31 华为技术有限公司 Air-isolated gesture adjusting method and terminal

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2007027452A1 (en) * 2005-08-31 2007-03-08 Microsoft Corporation Training convolutional neural networks on graphics processing units
CN101742341A (en) * 2010-01-14 2010-06-16 中山大学 Method and device for image processing
CN102750544A (en) * 2012-06-01 2012-10-24 浙江捷尚视觉科技有限公司 Detection system and detection method of rule-breaking driving that safety belt is not fastened and based on plate number recognition

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2007027452A1 (en) * 2005-08-31 2007-03-08 Microsoft Corporation Training convolutional neural networks on graphics processing units
CN101742341A (en) * 2010-01-14 2010-06-16 中山大学 Method and device for image processing
CN102750544A (en) * 2012-06-01 2012-10-24 浙江捷尚视觉科技有限公司 Detection system and detection method of rule-breaking driving that safety belt is not fastened and based on plate number recognition

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
彭国福: "图像处理中Gamma校正的研究和实现", 《电子工程师》 *

Cited By (52)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103544506A (en) * 2013-10-12 2014-01-29 Tcl集团股份有限公司 Method and device for classifying images on basis of convolutional neural network
CN103544506B (en) * 2013-10-12 2017-08-08 Tcl集团股份有限公司 A kind of image classification method and device based on convolutional neural networks
CN104680508A (en) * 2013-11-29 2015-06-03 华为技术有限公司 Convolutional neural network and target object detection method based on convolutional neural network
CN104809426A (en) * 2014-01-27 2015-07-29 日本电气株式会社 Convolutional neural network training method and target identification method and device
CN104809426B (en) * 2014-01-27 2019-04-05 日本电气株式会社 Training method, target identification method and the device of convolutional neural networks
CN105224963B (en) * 2014-06-04 2019-06-07 华为技术有限公司 The method and terminal of changeable deep learning network structure
CN105224963A (en) * 2014-06-04 2016-01-06 华为技术有限公司 The method of changeable degree of depth learning network structure and terminal
CN104063719A (en) * 2014-06-27 2014-09-24 深圳市赛为智能股份有限公司 Method and device for pedestrian detection based on depth convolutional network
CN104077577A (en) * 2014-07-03 2014-10-01 浙江大学 Trademark detection method based on convolutional neural network
TWI635446B (en) * 2014-07-22 2018-09-11 英特爾股份有限公司 Weight-shifting appratus, method, system and machine accessible storage medium
CN105320495A (en) * 2014-07-22 2016-02-10 英特尔公司 Weight-shifting mechanism for convolutional neural network
CN104182756A (en) * 2014-09-05 2014-12-03 大连理工大学 Method for detecting barriers in front of vehicles on basis of monocular vision
CN104182756B (en) * 2014-09-05 2017-04-12 大连理工大学 Method for detecting barriers in front of vehicles on basis of monocular vision
US10275688B2 (en) 2014-12-17 2019-04-30 Nokia Technologies Oy Object detection with neural network
WO2016095117A1 (en) * 2014-12-17 2016-06-23 Nokia Technologies Oy Object detection with neural network
CN107004138A (en) * 2014-12-17 2017-08-01 诺基亚技术有限公司 Utilize the object detection of neutral net
WO2016149881A1 (en) * 2015-03-20 2016-09-29 Intel Corporation Object recogntion based on boosting binary convolutional neural network features
US10685262B2 (en) 2015-03-20 2020-06-16 Intel Corporation Object recognition based on boosting binary convolutional neural network features
CN104850864A (en) * 2015-06-01 2015-08-19 深圳英智源智能系统有限公司 Unsupervised image recognition method based on convolutional neural network
CN104992179A (en) * 2015-06-23 2015-10-21 浙江大学 Fine-grained convolutional neural network-based clothes recommendation method
CN106874296A (en) * 2015-12-14 2017-06-20 阿里巴巴集团控股有限公司 A kind of style recognition methods of commodity and device
CN108496178B (en) * 2016-01-05 2023-08-08 御眼视觉技术有限公司 System and method for estimating future path
CN108496178A (en) * 2016-01-05 2018-09-04 御眼视觉技术有限公司 System and method for estimating Future Path
CN105691381A (en) * 2016-03-10 2016-06-22 大连理工大学 Stability control method and system for electric automobile with four independently driven wheels
CN105691381B (en) * 2016-03-10 2018-04-27 大连理工大学 A kind of four motorized wheels electric automobile stability control method and system
CN105975915A (en) * 2016-04-28 2016-09-28 大连理工大学 Front vehicle parameter identification method based on multitask convolution nerve network
CN105975915B (en) * 2016-04-28 2019-05-21 大连理工大学 A kind of front vehicles parameter identification method based on multitask convolutional neural networks
US11631005B2 (en) 2016-05-31 2023-04-18 Nokia Technologies Oy Method and apparatus for detecting small objects with an enhanced deep neural network
WO2017206066A1 (en) * 2016-05-31 2017-12-07 Nokia Technologies Oy Method and apparatus for detecting small objects with an enhanced deep neural network
CN106205126A (en) * 2016-08-12 2016-12-07 北京航空航天大学 Large-scale Traffic Network based on convolutional neural networks is blocked up Forecasting Methodology and device
CN106407931A (en) * 2016-09-19 2017-02-15 杭州电子科技大学 Novel deep convolution neural network moving vehicle detection method
CN106407931B (en) * 2016-09-19 2019-11-22 杭州电子科技大学 A kind of depth convolutional neural networks moving vehicle detection method
TWI638332B (en) * 2016-11-29 2018-10-11 財團法人車輛研究測試中心 Hierarchical object detection system with parallel architecture and method thereof
CN108122234A (en) * 2016-11-29 2018-06-05 北京市商汤科技开发有限公司 Convolutional neural networks training and method for processing video frequency, device and electronic equipment
CN106599832A (en) * 2016-12-09 2017-04-26 重庆邮电大学 Method for detecting and recognizing various types of obstacles based on convolution neural network
WO2019001346A1 (en) * 2017-06-26 2019-01-03 Huawei Technologies Co., Ltd. System and methods for object filtering and uniform representation for autonomous systems
WO2019047649A1 (en) * 2017-09-05 2019-03-14 百度在线网络技术(北京)有限公司 Method and device for determining driving behavior of unmanned vehicle
WO2019047655A1 (en) * 2017-09-05 2019-03-14 百度在线网络技术(北京)有限公司 Method and apparatus for use in determining driving behavior of driverless vehicle
CN107392189A (en) * 2017-09-05 2017-11-24 百度在线网络技术(北京)有限公司 For the method and apparatus for the driving behavior for determining unmanned vehicle
CN109726615A (en) * 2017-10-30 2019-05-07 北京京东尚科信息技术有限公司 A kind of recognition methods of road boundary and device
CN110399929A (en) * 2017-11-01 2019-11-01 腾讯科技(深圳)有限公司 Eye fundus image classification method, device and computer readable storage medium
CN110399929B (en) * 2017-11-01 2023-04-28 腾讯科技(深圳)有限公司 Fundus image classification method, fundus image classification apparatus, and computer-readable storage medium
CN108921003A (en) * 2018-04-26 2018-11-30 东华大学 Unmanned plane obstacle detection method based on convolutional neural networks and morphological image
CN108596115A (en) * 2018-04-27 2018-09-28 济南浪潮高新科技投资发展有限公司 A kind of vehicle checking method, apparatus and system based on convolutional neural networks
CN109657643A (en) * 2018-12-29 2019-04-19 百度在线网络技术(北京)有限公司 A kind of image processing method and device
CN109961509A (en) * 2019-03-01 2019-07-02 北京三快在线科技有限公司 Dimensionally map generalization and model training method, device and electronic equipment
CN110084190A (en) * 2019-04-25 2019-08-02 南开大学 Unstructured road detection method in real time under a kind of violent light environment based on ANN
CN110084190B (en) * 2019-04-25 2024-02-06 南开大学 Real-time unstructured road detection method under severe illumination environment based on ANN
CN111222476A (en) * 2020-01-10 2020-06-02 北京百度网讯科技有限公司 Video time sequence action detection method and device, electronic equipment and storage medium
US11600069B2 (en) 2020-01-10 2023-03-07 Beijing Baidu Netcom Science And Technology Co., Ltd. Method and apparatus for detecting temporal action of video, electronic device and storage medium
CN113325948B (en) * 2020-02-28 2023-02-07 华为技术有限公司 Air-isolated gesture adjusting method and terminal
CN113325948A (en) * 2020-02-28 2021-08-31 华为技术有限公司 Air-isolated gesture adjusting method and terminal

Also Published As

Publication number Publication date
CN103279759B (en) 2016-06-01

Similar Documents

Publication Publication Date Title
CN103279759A (en) Vehicle front trafficability analyzing method based on convolution nerve network
CN103034863B (en) The remote sensing image road acquisition methods of a kind of syncaryon Fisher and multiple dimensioned extraction
CN107316064B (en) Asphalt pavement crack classification and identification method based on convolutional neural network
CN103093250B (en) A kind of Adaboost method for detecting human face based on new Haar-like feature
CN104182985B (en) Remote sensing image change detection method
CN103020605B (en) Bridge identification method based on decision-making layer fusion
CN103729853B (en) High score remote sensing image building under three-dimension GIS auxiliary damages detection method
CN105184271A (en) Automatic vehicle detection method based on deep learning
CN102968790B (en) Remote sensing image change detection method based on image fusion
CN103902976A (en) Pedestrian detection method based on infrared image
CN103413151A (en) Hyperspectral image classification method based on image regular low-rank expression dimensionality reduction
CN103679205B (en) Assume based on shade and the Foregut fermenters method of layering HOG symmetrical feature checking
CN104299006A (en) Vehicle license plate recognition method based on deep neural network
CN103218831A (en) Video moving target classification and identification method based on outline constraint
CN104867150A (en) Wave band correction change detection method of remote sensing image fuzzy clustering and system thereof
CN103632363A (en) Object-level high-resolution remote sensing image change detection method based on multi-scale fusion
CN104361351A (en) Synthetic aperture radar (SAR) image classification method on basis of range statistics similarity
CN105005989A (en) Vehicle target segmentation method under weak contrast
CN102867183B (en) Method and device for detecting littered objects of vehicle and intelligent traffic monitoring system
CN102629380B (en) Remote sensing image change detection method based on multi-group filtering and dimension reduction
CN104751166A (en) Spectral angle and Euclidean distance based remote-sensing image classification method
Yuan et al. Learning to count buildings in diverse aerial scenes
CN106096655A (en) A kind of remote sensing image airplane detection method based on convolutional neural networks
CN104103070B (en) Landing point selecting method based on optical images
CN103488993A (en) Crowd abnormal behavior identification method based on FAST

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20160601

Termination date: 20190609

CF01 Termination of patent right due to non-payment of annual fee