CN110097047A - A vehicle detection method using single-line lidar based on deep learning - Google Patents

A vehicle detection method using single-line lidar based on deep learning Download PDF

Info

Publication number
CN110097047A
CN110097047A CN201910206463.6A CN201910206463A CN110097047A CN 110097047 A CN110097047 A CN 110097047A CN 201910206463 A CN201910206463 A CN 201910206463A CN 110097047 A CN110097047 A CN 110097047A
Authority
CN
China
Prior art keywords
convolutional neural
vehicle
deep learning
laser radar
interest
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910206463.6A
Other languages
Chinese (zh)
Other versions
CN110097047B (en
Inventor
瞿三清
卢凡
董金虎
陈广
许仲聪
陈凯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tongji University
Original Assignee
Tongji University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tongji University filed Critical Tongji University
Priority to CN201910206463.6A priority Critical patent/CN110097047B/en
Publication of CN110097047A publication Critical patent/CN110097047A/en
Application granted granted Critical
Publication of CN110097047B publication Critical patent/CN110097047B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/143Sensing or illuminating at different wavelengths
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • G06V10/242Aligning, centring, orientation detection or correction of the image by image rotation, e.g. by 90 degrees
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/584Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Multimedia (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Probability & Statistics with Applications (AREA)
  • Traffic Control Systems (AREA)
  • Image Analysis (AREA)

Abstract

本发明涉及一种基于深度学习采用单线激光雷达的车辆检测方法,包括以下步骤:1)通过单线激光雷达采集车辆的特征点云数据,并进行预处理得到二进制图像;2)对二进制图像中进行标注,获取其中车辆所在的位置,生成训练数据集,并采用聚类方法生成感兴趣区域;3)构建适用于目标车辆检测的深度卷积神经网络及其损失函数;4)对感兴趣区域进行增广后输入深度卷积神经网络中,并根据输出值与训练真值的差异对卷积神经网络中的参数进行训练更新,得到最优的网络参数,并根据训练好的深度卷积神经网络进行车辆位置检测,获取车辆位置。与现有技术相比,本发明具有稳健性高、检测成本低等优点。

The invention relates to a vehicle detection method based on deep learning using a single-line laser radar, comprising the following steps: 1) collecting feature point cloud data of the vehicle through a single-line laser radar, and performing preprocessing to obtain a binary image; 2) performing a binary image on the binary image Marking, obtaining the location of the vehicle, generating a training data set, and using a clustering method to generate a region of interest; 3) Constructing a deep convolutional neural network and its loss function suitable for target vehicle detection; 4) Carrying out the region of interest After augmentation, it is input into the deep convolutional neural network, and the parameters in the convolutional neural network are trained and updated according to the difference between the output value and the training true value to obtain the optimal network parameters, and according to the trained deep convolutional neural network Carry out vehicle position detection and obtain vehicle position. Compared with the prior art, the invention has the advantages of high robustness, low detection cost and the like.

Description

一种基于深度学习采用单线激光雷达的车辆检测方法A vehicle detection method using single-line lidar based on deep learning

技术领域technical field

本发明涉及智能驾驶技术领域,尤其是涉及一种基于深度学习采用单线激光雷达的车辆检测方法。The invention relates to the technical field of intelligent driving, in particular to a vehicle detection method based on deep learning and using single-line laser radar.

背景技术Background technique

在智能驾驶领域,对于车辆的检测是保障无人驾驶车辆安全行驶的关键性任务之一。对于车辆的检测,目前大多利用3D激光雷达或摄像头作为传感器,受限于单线激光雷达较少的数据信息鲜有利用单线激光雷达作为车辆检测传感器。但3D激光雷达的成本高昂,且采集数据得到的点云信息过多,需要耗费巨大的计算资源。利用摄像头作为检测传感器能够获得较高的检测精度,但是摄像头受光线影响较大,特别是在有夜晚、大雾、沙尘暴等情况下受限于采集图像很难完成周围车辆的检测。In the field of intelligent driving, vehicle detection is one of the key tasks to ensure the safe driving of unmanned vehicles. For vehicle detection, 3D lidar or camera is mostly used as the sensor at present, limited by the less data information of single-line lidar, it is rare to use single-line lidar as a vehicle detection sensor. However, the cost of 3D lidar is high, and the point cloud information obtained from the collected data is too much, which requires huge computing resources. Using a camera as a detection sensor can achieve higher detection accuracy, but the camera is greatly affected by light, especially at night, heavy fog, sandstorms, etc. It is difficult to complete the detection of surrounding vehicles due to limited image acquisition.

发明内容Contents of the invention

本发明的目的就是为了克服上述现有技术存在的缺陷而提供一种基于深度学习采用单线激光雷达的车辆检测方法。The purpose of the present invention is to provide a vehicle detection method based on deep learning and using single-line laser radar in order to overcome the above-mentioned defects in the prior art.

本发明的目的可以通过以下技术方案来实现:The purpose of the present invention can be achieved through the following technical solutions:

一种基于深度学习采用单线激光雷达的车辆检测方法,包括以下步骤:A vehicle detection method using single-line laser radar based on deep learning, comprising the following steps:

1)通过单线激光雷达采集车辆的特征点云数据,并进行预处理得到二进制图像;1) Collect the characteristic point cloud data of the vehicle through single-line lidar, and perform preprocessing to obtain a binary image;

2)对二进制图像中进行标注,获取其中车辆所在的位置,生成训练数据集,并采用聚类方法生成感兴趣区域;2) Mark the binary image, obtain the position of the vehicle in it, generate a training data set, and use a clustering method to generate a region of interest;

3)构建适用于目标车辆检测的深度卷积神经网络及其损失函数;3) Construct a deep convolutional neural network and its loss function suitable for target vehicle detection;

4)对感兴趣区域进行增广后输入深度卷积神经网络中,并根据输出值与训练真值的差异对卷积神经网络中的参数进行训练更新,得到最优的网络参数,并根据训练好的深度卷积神经网络进行车辆位置检测,获取车辆位置。4) The region of interest is augmented and then input into the deep convolutional neural network, and the parameters in the convolutional neural network are trained and updated according to the difference between the output value and the training true value to obtain the optimal network parameters, and according to the training A good deep convolutional neural network performs vehicle position detection and obtains the vehicle position.

述的步骤1)具体包括以下步骤:Described step 1) specifically comprises the following steps:

11)将采集到的特征点云数据以单线激光雷达为坐标原点的极坐标系转换为全局统一的笛卡尔坐标系;11) Convert the collected feature point cloud data into a globally unified Cartesian coordinate system with the single-line laser radar as the coordinate origin;

12)对坐标转换后的点云数据栅格网格化,转换为二进制图像。12) Grid the point cloud data after the coordinate transformation, and convert it into a binary image.

所述的步骤11)中,转换表达式为:In described step 11), conversion expression is:

(xj0,yj0)=(xj1,yj1)R+t(x j0 ,y j0 )=(x j1 ,y j1 )R+t

其中,(rjj)为原始点云数据中点j的位置坐标,(xj1,yj1)为点j转换为以激光雷达为坐标原点的笛卡尔坐标系中的位置坐标,(xj0,yj0)为全局统一笛卡尔坐标系中的坐标,R为转换旋转矩阵,t为转换平移向量。Among them, (r j , φ j ) is the position coordinate of point j in the original point cloud data, (x j1 , y j1 ) is the position coordinate of point j transformed into the Cartesian coordinate system with the lidar as the coordinate origin, ( x j0 , y j0 ) are the coordinates in the global unified Cartesian coordinate system, R is the transformation rotation matrix, and t is the transformation translation vector.

所述的步骤2)中,标注内容包括图像的像素级标注和目标车辆的边界框标注。In the step 2), the annotation content includes the pixel-level annotation of the image and the bounding box annotation of the target vehicle.

所述的深度卷积神经网络为Faster R-CNN卷积神经网络,其以设定尺寸的二进制图像作为输入,以与输入二进制图像上对应目标车辆的位置和置信度为输出。The deep convolutional neural network is a Faster R-CNN convolutional neural network, which takes a binary image of a set size as input, and outputs the position and confidence of the target vehicle corresponding to the input binary image.

所述的步骤3)中,深度卷积神经网络的损失函数的表达式为:In described step 3), the expression of the loss function of deep convolutional neural network is:

Loss=Lcls(p,u)+λ[u=1]Lloc(tu,v)Loss=L cls (p,u)+λ[u=1]L loc (t u ,v)

Lcls(p,u)=-log(p)L cls (p,u)=-log(p)

x=(tu-v)x=(t u -v)

其中,Lcls(p,u)为目标分类检测损失子函数,Lloc(tu,v)为距离损失子函数,p为对于目标类别的预测因子,u为对应类别的实际因子,λ为损失函数的加权权重,当u=1时表示在感兴趣区域是目标车辆,当u=0时表示感兴趣区域是背景,tu为预测的位置因子,v为训练样本中真实的位置因子,x为预测值与真实值的偏差。Among them, L cls (p, u) is the target classification detection loss sub-function, L loc (t u , v) is the distance loss sub-function, p is the predictor for the target category, u is the actual factor of the corresponding category, and λ is The weighted weight of the loss function, when u=1, it means that the target vehicle is in the region of interest, when u=0, it means that the region of interest is the background, t u is the predicted position factor, v is the real position factor in the training sample, x is the deviation between the predicted value and the true value.

所述的步骤2)中,所述的聚类方法包括密度聚类、均值聚类和Mean-Shift聚类。In the step 2), the clustering methods include density clustering, mean value clustering and Mean-Shift clustering.

所述的步骤4)中,对感兴趣区域进行增广具体为:In the described step 4), the region of interest is augmented specifically as follows:

将感兴趣区域图像进行随机水平翻转、裁剪并统一缩放至固定的尺寸,并且标注数据也进行相应的翻转、裁剪和缩放。The image of the region of interest is randomly flipped horizontally, cropped, and uniformly scaled to a fixed size, and the annotation data is also flipped, cropped, and scaled accordingly.

所述的步骤4)中,训练深度卷积神经网络具体为:In the described step 4), the training depth convolutional neural network is specifically:

依据损失函数,利用梯度下降反向传播方法,对深度卷积神经网络的参数进行迭代更新,将迭代至最大设定次数后得到的网络参数作为最优的网络参数,完成训练。According to the loss function, the parameters of the deep convolutional neural network are iteratively updated using the gradient descent backpropagation method, and the network parameters obtained after iterating to the maximum set number of times are used as the optimal network parameters to complete the training.

与现有技术相比,本发明具有以下优点:Compared with the prior art, the present invention has the following advantages:

一、稳健性高:由于单线激光雷达能够在各种复杂的工况下都能采集准确的车辆点云数据,辅以具有较高的鲁棒性深度学习神经网络的车辆检测算法,因此,本发明的车辆检测方法具有很高的稳健性,在复杂工况下也能够保证检测结果的相对准确性。1. High robustness: Since the single-line laser radar can collect accurate vehicle point cloud data under various complex working conditions, supplemented by a vehicle detection algorithm with a high robustness deep learning neural network, this paper The invented vehicle detection method has high robustness, and can guarantee the relative accuracy of detection results even under complex working conditions.

二、检测成本低:本发明所提出的车辆检测算法,采用单线激光雷达作为传感器,相比于3D激光雷达,检测成本更低。2. Low detection cost: The vehicle detection algorithm proposed in the present invention uses a single-line laser radar as a sensor, and the detection cost is lower than that of a 3D laser radar.

附图说明Description of drawings

图1为本发明的检测方法的流程图。Fig. 1 is a flowchart of the detection method of the present invention.

图2为本发明的实施例中车辆深度卷积网络结构示意图。Fig. 2 is a schematic diagram of the vehicle deep convolutional network structure in the embodiment of the present invention.

具体实施方式Detailed ways

下面结合附图和具体实施例对本发明进行详细说明。The present invention will be described in detail below in conjunction with the accompanying drawings and specific embodiments.

实施例Example

本发明提供了一种基于深度学习利用单线激光雷达的车辆检测的方法,其通过单线激光雷达作为检测的传感器得到车辆的点云数据,进行预处理后输入到深度卷积神经网络,最后得到车辆的位置与置信度。如图1所示,该方法包括如下步骤:The present invention provides a method for vehicle detection based on deep learning using single-line laser radar, which uses single-line laser radar as a detection sensor to obtain point cloud data of the vehicle, performs preprocessing and then inputs it to a deep convolutional neural network, and finally obtains the vehicle location and confidence. As shown in Figure 1, the method includes the following steps:

(1)利用单线激光雷达采集车辆特征点云数据,并对所采集的点云数据进行预处理;(1) Use single-line lidar to collect vehicle characteristic point cloud data, and preprocess the collected point cloud data;

(2)对采集数据通过人工标注其中车辆所在位置,构建训练所用数据集(2) Manually mark the location of the vehicle in the collected data, and construct the data set used for training

(3)构建适用于车辆检测的深度卷积神经网络与损失函数(3) Construct a deep convolutional neural network and loss function suitable for vehicle detection

(4)对步骤(2)所得训练数据利用聚类方法生成感兴趣区域,同时将训练数据与感兴趣区域增广输入到步骤(3)所构建的深度卷积神经网络中。根据输出值与训练真值的差异对卷积神经网络中的参数进行训练更新,最后得到较为理想的网络参数。(4) Use the clustering method to generate the region of interest for the training data obtained in step (2), and simultaneously input the training data and region of interest augmentation into the deep convolutional neural network constructed in step (3). According to the difference between the output value and the training true value, the parameters in the convolutional neural network are trained and updated, and finally the ideal network parameters are obtained.

在本实施例中,步骤(1)中对点云数据的预处理包括对点云数据的坐标转换与图像转换等两个如下步骤:In the present embodiment, the preprocessing to point cloud data in step (1) includes two following steps such as coordinate conversion and image conversion to point cloud data:

(1-1)本实施例中的单线激光雷达位于实施例车辆的前方,单线激光雷达按一定地帧率采集周围的点云数据,采集结果以激光雷达为坐标原点,并以极坐标形式存储。将采集的点云数据转换到笛卡尔坐标系中。(1-1) The single-line laser radar in this embodiment is located in front of the vehicle in the embodiment, and the single-line laser radar collects surrounding point cloud data at a certain frame rate, and the collection results take the laser radar as the coordinate origin and store them in polar coordinates . Convert the collected point cloud data into Cartesian coordinate system.

转换表达式如下:The conversion expression is as follows:

上式中,(rjj)表示采集的原始点云数据中的某一点,(xj1,yj1)为采集的点云数据中的某一点转换为以激光雷达为坐标原点的笛卡尔坐标系中的表示。In the above formula, (r j , φ j ) represents a certain point in the collected original point cloud data, and (x j1 , y j1 ) is a certain point in the collected point cloud data converted into a flute with the lidar as the coordinate origin. Representation in the Karl coordinate system.

(1-2)对点云数据图像化处理。在本实施例中,激光雷达采集的距离上限设置为25m,单线激光雷达的采集视角为180°,将坐标转换后的点云数据网格化,然后调整尺寸为250×250的二进制图像,如果对应网格中有数据点,则置为1否则置为0。(1-2) Image processing of point cloud data. In this embodiment, the upper limit of the distance collected by the laser radar is set to 25m, and the collection angle of the single-line laser radar is 180°. The point cloud data after coordinate transformation is gridded, and then the binary image with a size of 250×250 is adjusted. If If there is a data point in the corresponding grid, it is set to 1, otherwise it is set to 0.

在本实施例中,步骤(2)为构建深度学习训练所需的数据集。在对采集的激光雷达点云数据进行处理后,需要对训练数据进行人工标注,以形成训练所需数据集。标注的方式包括但不限于图像的像素级标注、目标车辆的边界框标注。标注时需至少包含目标车辆的位置,但可以拓展增加目标车辆的姿态信息等。In this embodiment, step (2) is to construct a data set required for deep learning training. After processing the collected lidar point cloud data, it is necessary to manually label the training data to form the data set required for training. The labeling methods include but are not limited to pixel-level labeling of images and bounding box labeling of target vehicles. When labeling, at least the position of the target vehicle must be included, but the attitude information of the target vehicle can be expanded and added.

在本实施例中,步骤(3)为构建适用于目标车辆检测的深度卷积神经网络与损失函数。深度卷积神经网络的构建与步骤(2)中所准备的训练数据集直接相关,本实施例中,步骤(2)采用目标车辆的边界框标注,因此在本实施例中深度卷积神经网络的结构与FastR-CNN相仿,主体结构的搭建参考Fast R-CNN,卷积神经网络结构如图2所示。In this embodiment, step (3) is to construct a deep convolutional neural network and a loss function suitable for target vehicle detection. The construction of the deep convolutional neural network is directly related to the training data set prepared in step (2). In this embodiment, step (2) uses the bounding box label of the target vehicle, so in this embodiment, the deep convolutional neural network The structure is similar to that of Fast R-CNN, and the construction of the main structure refers to Fast R-CNN. The convolutional neural network structure is shown in Figure 2.

在本实施例中,步骤(3)中,深度卷积神经网络的损失函数以两部分加权构成:In this embodiment, in step (3), the loss function of the deep convolutional neural network is composed of two weighted parts:

Loss=Lcls(p,u)+λ[u=1]Lloc(tu,v)Loss=L cls (p,u)+λ[u=1]L loc (t u ,v)

(3-1)构建目标分类损失函数Lcls(p,u),其中p是对于目标类别的预测因子,而u表示对应类别的实际因子。通常采用log损失函数构建,其中p表示某一类别的预测概率,p越接近1,置信度越高,损失越小。(3-1) Construct the target classification loss function L cls (p,u), where p is the predictor for the target category, and u represents the actual factor of the corresponding category. It is usually constructed using a log loss function, where p represents the predicted probability of a certain category, and the closer p is to 1, the higher the confidence and the smaller the loss.

Lcls(p,u)=-log(p)L cls (p,u)=-log(p)

(3-2)构建目标检测距离Lloc(tu,v)。其中,λ表示损失函数的加权权重,通常可以取λ=1。[u=1]表示在感兴趣区域是目标车辆取1,当感兴趣区域是背景时取0,即如果当前的感兴趣区域是环境无关事物时,不考虑其距离损失。式子中的tu表示代表所预测的位置因子,而v表示训练样本中真实的位置因子。通常所述的距离损失子函数以平滑曼哈顿距离公式smoothL1(x)构建,其表达式为:其中x=(tu-v),表示预测值与真实值的偏差。(3-2) Construct the target detection distance L loc (t u ,v). Among them, λ represents the weighted weight of the loss function, and λ=1 can usually be taken. [u=1] means 1 is taken when the ROI is the target vehicle, and 0 is taken when the ROI is the background, that is, if the current ROI is something irrelevant to the environment, its distance loss is not considered. The t u in the formula represents the predicted position factor, and v represents the real position factor in the training sample. Usually, the distance loss sub-function is constructed with the smooth Manhattan distance formula smooth L1 (x), and its expression is: where x=(t u -v), represents the deviation between the predicted value and the real value.

在本实施例中,步骤(4)利用聚类生成感兴趣区域的方法采用密度聚类算法。具体实现为,将转换前的点云数据对输入到密度聚类函数,通过密度聚类函数对点云数据对的类别划分,然后将点云数据对的类别转换到二进制图像中对应的点,利用最小矩形框包围同一类别的二进制图像点,得到的该矩形框内的图像即为感兴趣区域。在实现密度聚类时,需要指定聚类所生成的簇需要的最小数据点对数,以及每一类簇中的领域距离阈值,在本实施例中,最小数据点对数设置为5,领域距离阈值设置为0.5m。In this embodiment, the method of generating the region of interest by clustering in step (4) adopts a density clustering algorithm. The specific implementation is to input the point cloud data pair before conversion into the density clustering function, classify the point cloud data pair through the density clustering function, and then convert the point cloud data pair to the corresponding point in the binary image, The binary image points of the same category are surrounded by the smallest rectangular frame, and the obtained image within the rectangular frame is the region of interest. When implementing density clustering, it is necessary to specify the minimum logarithm of data points required by the clusters generated by the clustering, and the domain distance threshold in each type of cluster. In this embodiment, the minimum logarithm of data points is set to 5, and the domain distance The distance threshold is set to 0.5m.

在本实施例中,感兴趣区域输入到深度卷积神经网络时,由于感兴趣区域的尺寸取决与聚类结果的大小,因此为保证深度卷积网络的正常运行,先对图像数据进行卷积特征提取,后通过ROI空间池化的方式,将感兴趣区域的特征图像转换为统一尺寸大小。In this embodiment, when the region of interest is input to the deep convolutional neural network, since the size of the region of interest depends on the size of the clustering result, in order to ensure the normal operation of the deep convolutional network, the image data is first convoluted Feature extraction, and then convert the feature image of the region of interest into a uniform size through ROI space pooling.

在本实施例中,步骤(4)对训练数据的增广,主要包括将图像进行随机的水平翻转、裁剪并统一缩放到固定的尺寸,标注数据也进行相应的翻转、裁剪和缩放,在此基础上对得到的图像按通道进行归一化处理,本实施例中采用的固定尺寸为250×250。In this embodiment, the augmentation of the training data in step (4) mainly includes random horizontal flipping, cropping, and uniform scaling of the image to a fixed size, and corresponding flipping, cropping, and scaling of the labeled data. Here Basically, the obtained images are normalized by channel, and the fixed size used in this embodiment is 250×250.

在本实施例中,步骤(4)对训练网络模型的初始化时,先利用在ImageNet或其他图像分类数据集上利用SoftMax损失函数对物体特征提取网络进行预训练,得到的参数值作为网络的初始参数。In this embodiment, when initializing the training network model in step (4), first use the SoftMax loss function on ImageNet or other image classification data sets to pre-train the object feature extraction network, and the obtained parameter values are used as the initial parameters of the network. parameter.

在本实施例中,步骤(4)中对网络的训练时,利用加权的损失函数计算综合损失值,然后进行反向传播计算梯度,并使用Adam等优化器更新网络参数,迭代一定次数得到最终的结果。并将最终的参数结果设置为目标车辆检测器的网络模型参数,以供检测目标车辆使用。In this embodiment, when training the network in step (4), the weighted loss function is used to calculate the comprehensive loss value, and then the gradient is calculated by backpropagation, and the network parameters are updated using an optimizer such as Adam, and a certain number of iterations is used to obtain the final the result of. And the final parameter results are set as the network model parameters of the target vehicle detector for use in detecting the target vehicle.

总之,本发明提供了一种基于深度学习利用单线激光雷达的目标车辆检测的方法,其通过单线激光雷达作为检测的传感器得到目标车辆的点云数据,进行预处理后输入到深度卷积神经网络,最后得到目标车辆的位置与置信度。该检测性能优秀,同时有较高的鲁棒性,实现成本低,容易部署到现有的智能泊车机器人上用于目标车辆的检测。In a word, the present invention provides a method of target vehicle detection based on deep learning using single-line lidar, which uses single-line lidar as a detection sensor to obtain the point cloud data of the target vehicle, and then input it to the deep convolutional neural network after preprocessing , and finally get the position and confidence of the target vehicle. The detection performance is excellent, and at the same time, it has high robustness, low implementation cost, and is easy to deploy to the existing intelligent parking robot for target vehicle detection.

熟悉本领域技术的人员显然可以容易地对这些实施例做出各种修改,并把在此说明的一般原理应用到其他实施例中而不必经过创造性的劳动。因此,本发明不限于这里的实施例,本领域技术人员根据本发明的揭示,不脱离本发明范畴所做出的改进和修改都应该在本发明的保护范围之内。It is obvious that those skilled in the art can easily make various modifications to these embodiments, and apply the general principles described here to other embodiments without creative efforts. Therefore, the present invention is not limited to the embodiments herein. Improvements and modifications made by those skilled in the art according to the disclosure of the present invention without departing from the scope of the present invention should fall within the protection scope of the present invention.

Claims (9)

1. a kind of vehicle checking method for using single line laser radar based on deep learning, which comprises the following steps:
1) the feature point cloud data of vehicle is acquired by single line laser radar, and is pre-processed to obtain binary picture;
2) to being labeled in binary picture, the wherein position where vehicle is obtained, generates training dataset, and using cluster Method generates area-of-interest;
3) building is suitable for the depth convolutional neural networks and its loss function of target vehicle detection;
4) it is inputted in depth convolutional neural networks after carrying out augmentation to area-of-interest, and according to the difference of output valve and training true value The different parameter in convolutional neural networks is trained update, obtains optimal network parameter, and roll up according to trained depth Product neural network carries out vehicle location detection, obtains vehicle location.
2. a kind of vehicle checking method for using single line laser radar based on deep learning according to claim 1, special Sign is, the step 1) stated specifically includes the following steps:
11) collected feature point cloud data is converted to using single line laser radar as the polar coordinate system of coordinate origin global unified Cartesian coordinate system;
12) to the point cloud data grid gridding after coordinate conversion, binary picture is converted to.
3. a kind of vehicle checking method for using single line laser radar based on deep learning according to claim 1, special Sign is, in the step 11), transformed representation are as follows:
(xj0,yj0)=(xj1,yj1)R+t
Wherein, (rjj) be original point cloud data midpoint j position coordinates, (xj1,yj1) it is that point j is converted to and is with laser radar Position coordinates in the cartesian coordinate system of coordinate origin, (xj0,yj0) it is the global coordinate unified in cartesian coordinate system, R is Spin matrix is converted, t is translation vector.
4. a kind of vehicle checking method for using single line laser radar based on deep learning according to claim 1, special Sign is, in the step 2), marked content includes that the Pixel-level mark of image and the bounding box of target vehicle mark.
5. a kind of vehicle checking method for using single line laser radar based on deep learning according to claim 1, special Sign is that the depth convolutional neural networks are Faster R-CNN convolutional neural networks, with the binary system being sized Image is as input, to export with position and the confidence level of target vehicle is corresponded on input binary picture.
6. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described In step 3), the expression formula of the loss function of depth convolutional neural networks are as follows:
Loss=Lcls(p, u)+λ [u=1] Lloc(tu,v)
Lcls(p, u)=- log (p)
X=(tu-v)
Wherein, Lcls(p, u) is target classification Detectability loss subfunction, Lloc(tu, v) and it is range loss subfunction, p is for mesh The predictive factor of classification is marked, u is the practical factor of corresponding classification, and λ is the weighting weight of loss function, indicates feeling as u=1 Interest region is target vehicle, indicates that area-of-interest is background, t as u=0uFor the location factor of prediction, v is training sample True location factor in this, x are the deviation of predicted value and true value.
7. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described In step 2), the clustering method includes Density Clustering, mean cluster and Mean-Shift cluster.
8. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described In step 4), augmentation is carried out to area-of-interest specifically:
Region of interest area image is carried out Random Level overturning, cuts and uniformly zooms to fixed size, and labeled data Also it overturn, cut and is scaled accordingly.
9. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described In step 4), training depth convolutional neural networks specifically:
According to loss function, declines back-propagation method using gradient, the parameter of depth convolutional neural networks is iterated more Newly, the network parameter obtained after iteration to maximum being set number completes training as optimal network parameter.
CN201910206463.6A 2019-03-19 2019-03-19 A vehicle detection method based on deep learning using single-line lidar Active CN110097047B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910206463.6A CN110097047B (en) 2019-03-19 2019-03-19 A vehicle detection method based on deep learning using single-line lidar

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910206463.6A CN110097047B (en) 2019-03-19 2019-03-19 A vehicle detection method based on deep learning using single-line lidar

Publications (2)

Publication Number Publication Date
CN110097047A true CN110097047A (en) 2019-08-06
CN110097047B CN110097047B (en) 2021-10-08

Family

ID=67443320

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910206463.6A Active CN110097047B (en) 2019-03-19 2019-03-19 A vehicle detection method based on deep learning using single-line lidar

Country Status (1)

Country Link
CN (1) CN110097047B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110992731A (en) * 2019-12-12 2020-04-10 苏州智加科技有限公司 Laser radar-based 3D vehicle detection method and device and storage medium
CN112444784A (en) * 2019-08-29 2021-03-05 北京市商汤科技开发有限公司 Three-dimensional target detection and neural network training method, device and equipment
CN112560736A (en) * 2020-12-22 2021-03-26 上海有个机器人有限公司 Random angle laser gate detection method based on convolutional neural network and storage medium
CN113655477A (en) * 2021-06-11 2021-11-16 成都圭目机器人有限公司 Method for automatically detecting geological diseases of land radar by adopting shallow layer
CN114022563A (en) * 2021-10-25 2022-02-08 同济大学 Dynamic obstacle detection method for automatic driving
US11586925B2 (en) * 2017-09-29 2023-02-21 Samsung Electronics Co., Ltd. Neural network recogntion and training method and apparatus

Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170039436A1 (en) * 2015-08-03 2017-02-09 Nokia Technologies Oy Fusion of RGB Images and Lidar Data for Lane Classification
CN107239794A (en) * 2017-05-18 2017-10-10 深圳市速腾聚创科技有限公司 Point cloud data segmentation method and terminal
CN108830188A (en) * 2018-05-30 2018-11-16 西安理工大学 Vehicle checking method based on deep learning
CN109063753A (en) * 2018-07-18 2018-12-21 北方民族大学 A kind of three-dimensional point cloud model classification method based on convolutional neural networks
CN109118500A (en) * 2018-07-16 2019-01-01 重庆大学产业技术研究院 A kind of dividing method of the Point Cloud Data from Three Dimension Laser Scanning based on image
CN109212541A (en) * 2018-09-20 2019-01-15 同济大学 High-precision vehicle detecting system based on vehicle perpendicular type feature and laser radar
CN109270543A (en) * 2018-09-20 2019-01-25 同济大学 A kind of system and method for pair of target vehicle surrounding vehicles location information detection
CN109270544A (en) * 2018-09-20 2019-01-25 同济大学 Mobile robot self-localization system based on shaft identification
CN109324616A (en) * 2018-09-20 2019-02-12 同济大学 Alignment method of unmanned parking and handling robot based on on-board sensors
CN109344786A (en) * 2018-10-11 2019-02-15 深圳步智造科技有限公司 Target identification method, device and computer readable storage medium
CN109389053A (en) * 2018-09-20 2019-02-26 同济大学 High performance vehicle detection system based on vehicle perpendicular type feature
CN109386155A (en) * 2018-09-20 2019-02-26 同济大学 Nobody towards automated parking ground parks the alignment method of transfer robot

Patent Citations (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170039436A1 (en) * 2015-08-03 2017-02-09 Nokia Technologies Oy Fusion of RGB Images and Lidar Data for Lane Classification
CN107239794A (en) * 2017-05-18 2017-10-10 深圳市速腾聚创科技有限公司 Point cloud data segmentation method and terminal
CN108830188A (en) * 2018-05-30 2018-11-16 西安理工大学 Vehicle checking method based on deep learning
CN109118500A (en) * 2018-07-16 2019-01-01 重庆大学产业技术研究院 A kind of dividing method of the Point Cloud Data from Three Dimension Laser Scanning based on image
CN109063753A (en) * 2018-07-18 2018-12-21 北方民族大学 A kind of three-dimensional point cloud model classification method based on convolutional neural networks
CN109212541A (en) * 2018-09-20 2019-01-15 同济大学 High-precision vehicle detecting system based on vehicle perpendicular type feature and laser radar
CN109270543A (en) * 2018-09-20 2019-01-25 同济大学 A kind of system and method for pair of target vehicle surrounding vehicles location information detection
CN109270544A (en) * 2018-09-20 2019-01-25 同济大学 Mobile robot self-localization system based on shaft identification
CN109324616A (en) * 2018-09-20 2019-02-12 同济大学 Alignment method of unmanned parking and handling robot based on on-board sensors
CN109389053A (en) * 2018-09-20 2019-02-26 同济大学 High performance vehicle detection system based on vehicle perpendicular type feature
CN109386155A (en) * 2018-09-20 2019-02-26 同济大学 Nobody towards automated parking ground parks the alignment method of transfer robot
CN109344786A (en) * 2018-10-11 2019-02-15 深圳步智造科技有限公司 Target identification method, device and computer readable storage medium

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
AMIR KOTB 等: "Smart Parking Guidance, Monitoring and Reservations: A Review", 《 IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE》 *
企鹅号 - AI科技大本营: "无人驾驶汽车系统入门:基于深度学习的实时激光雷达点云目标检测及ROS实现", 《HTTPS://CLOUD.TENCENT.COM/DEVELOPER/NEWS/339676》 *
李游: "基于车载激光扫描数据的城市街道信息提取技术研究", 《中国博士学位论文全文数据库 基础科学辑》 *
罗海峰等: "基于DBN的车载激光点云路侧多目标提取", 《测绘学报》 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11586925B2 (en) * 2017-09-29 2023-02-21 Samsung Electronics Co., Ltd. Neural network recogntion and training method and apparatus
CN112444784A (en) * 2019-08-29 2021-03-05 北京市商汤科技开发有限公司 Three-dimensional target detection and neural network training method, device and equipment
CN112444784B (en) * 2019-08-29 2023-11-28 北京市商汤科技开发有限公司 Three-dimensional target detection and neural network training method, device and equipment
CN110992731A (en) * 2019-12-12 2020-04-10 苏州智加科技有限公司 Laser radar-based 3D vehicle detection method and device and storage medium
CN112560736A (en) * 2020-12-22 2021-03-26 上海有个机器人有限公司 Random angle laser gate detection method based on convolutional neural network and storage medium
CN112560736B (en) * 2020-12-22 2025-01-07 上海有个机器人有限公司 A method and storage medium for detecting laser doors at any angle based on convolutional neural network
CN113655477A (en) * 2021-06-11 2021-11-16 成都圭目机器人有限公司 Method for automatically detecting geological diseases of land radar by adopting shallow layer
CN113655477B (en) * 2021-06-11 2023-09-01 成都圭目机器人有限公司 Method for automatically detecting geological diseases by adopting shallow layer ground radar
CN114022563A (en) * 2021-10-25 2022-02-08 同济大学 Dynamic obstacle detection method for automatic driving

Also Published As

Publication number Publication date
CN110097047B (en) 2021-10-08

Similar Documents

Publication Publication Date Title
Yu et al. A real-time detection approach for bridge cracks based on YOLOv4-FPM
CN111507982B (en) A point cloud semantic segmentation method based on deep learning
CN111626128B (en) A Pedestrian Detection Method Based on Improved YOLOv3 in Orchard Environment
CN110097047A (en) A vehicle detection method using single-line lidar based on deep learning
CN109598241B (en) Recognition method of ships at sea based on satellite imagery based on Faster R-CNN
CN110069993B (en) A target vehicle detection method based on deep learning
CN108038846A (en) Transmission line equipment image defect detection method and system based on multilayer convolutional neural networks
CN113536920B (en) A semi-supervised 3D point cloud object detection method
CN114565860A (en) A Multi-dimensional Reinforcement Learning Synthetic Aperture Radar Image Target Detection Method
CN105787501A (en) Vegetation classification method capable of automatically selecting features in power transmission line corridor area
CN110287837A (en) Sea Obstacle Detection Method Based on Prior Estimation Network and Spatial Constraint Hybrid Model
CN113011338B (en) Method and system for detecting lane markings
CN113487610B (en) Herpes image recognition method and device, computer equipment and storage medium
CN113239753A (en) Improved traffic sign detection and identification method based on YOLOv4
WO2024230038A1 (en) Three-dimensional point-cloud semantic segmentation method based on multi-level boundary enhancement for unstructured environment
Zhou et al. Building segmentation from airborne VHR images using Mask R-CNN
Chen et al. Vehicles detection on expressway via deep learning: Single shot multibox object detector
CN108427919A (en) A kind of unsupervised oil tank object detection method guiding conspicuousness model based on shape
CN115115601A (en) A remote sensing ship target detection method based on deformation attention pyramid
CN115115917A (en) 3D point cloud target detection method based on attention mechanism and image feature fusion
CN115100136A (en) Workpiece category and pose estimation method based on YOLOv4-tiny model
Zhang et al. A graph-voxel joint convolution neural network for ALS point cloud segmentation
CN116597270A (en) Road damage target detection method based on integrated learning network of attention mechanism
CN117746252A (en) A landslide detection method based on improved lightweight YOLOv7
CN116912490A (en) Power tower point cloud segmentation method and system based on hierarchical neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant