WO2020083103A1 - Vehicle positioning method based on deep neural network image recognition - Google Patents
Vehicle positioning method based on deep neural network image recognition Download PDFInfo
- Publication number
- WO2020083103A1 WO2020083103A1 PCT/CN2019/111840 CN2019111840W WO2020083103A1 WO 2020083103 A1 WO2020083103 A1 WO 2020083103A1 CN 2019111840 W CN2019111840 W CN 2019111840W WO 2020083103 A1 WO2020083103 A1 WO 2020083103A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- neural network
- deep neural
- road sign
- vehicle
- coordinate system
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/582—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of traffic signs
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C21/00—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00
- G01C21/26—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network
- G01C21/28—Navigation; Navigational instruments not provided for in groups G01C1/00 - G01C19/00 specially adapted for navigation in a road network with correlation of data from several navigational instruments
- G01C21/30—Map- or contour-matching
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/24—Aligning, centring, orientation detection or correction of the image
- G06V10/245—Aligning, centring, orientation detection or correction of the image by locating a pattern; Special marks for positioning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/09—Recognition of logos
Definitions
- the invention relates to the technical field of image recognition and positioning, in particular to a vehicle positioning method based on deep neural network image recognition, and a training method of the deep neural network.
- vehicle positioning technology mainly uses GPS technology and high-precision map matching positioning.
- the GPS technology has the following problems when it is used: when using the ordinary GPS mode for positioning, the positioning error reaches the meter level, which cannot meet the accuracy requirements of the vehicle; if the GPS RTK mode is used, it is necessary to obtain both satellite information and ground reference positioning information It is necessary to install reference positioning communication equipment along the road. The equipment cost and use cost are high. When the vehicle enters the road section with poor satellite signal, such as dense forest or tunnel, the GPS signal is easy to be lost, thereby losing positioning information.
- Map data needs to be established and stored on the vehicle in advance.
- point cloud data or image data of the current environment of the vehicle is obtained through an external laser radar or camera device And match with pre-stored map data.
- the cost of map making and software and hardware cost of matching calculation are relatively high.
- a low-cost, high-precision vehicle positioning method is needed to provide reliable data support for vehicle positioning, pit route planning, and speed control.
- the invention provides a vehicle positioning method based on deep neural network image recognition, and a training method of the deep neural network.
- the invention can increase the training accuracy of the deep neural network by increasing the number of training samples and optimizing the parameters of the neural network, thereby improving the positioning accuracy of the vehicle, and the required equipment cost and use cost are low.
- the first aspect of the present invention provides a deep neural network training method for road sign recognition, including the following steps:
- Road sign graphic setting step set a road sign graphic on the road surface of the station inbound direction, and the distance between the marking point of the road sign graphic and the edge of the station inbound direction is L;
- Steps for setting the camera install the camera on the vehicle, the optical axis of the camera coincides with the longitudinal centerline of the vehicle body, and the distance between the lens center of the camera and the ground is H
- Training sample production step calculate the position coordinates of the identification point of the road sign graphic in the image coordinate system in each of the image samples, make a label set, and combine each of the image samples with the corresponding labels Set pairing to form training samples;
- Steps for building a deep neural network on the basis of the target recognition classification deep neural network, the final classification output layer of the network is modified into an output layer composed of 2 nodes to output the position of the marking point of the road marking graphic coordinate;
- Deep neural network training step input the training samples to the deep neural network for training.
- the shooting time is selected at noon on a sunny day and at night on a sunny day.
- the shooting time is selected at noon on rainy days and night on rainy days.
- the shooting time is selected at noon on a foggy day and at night on a foggy day.
- the photographing device photographs an image sample of the road marking pattern every 5 ° within an angle range of 5 ° to 180 ° between the optical axis and the road surface.
- the lens parameters of the shooting device are selected so that when all the road sign graphics appear in the lens screen, the road sign graphics can occupy more than 20% of the area of the lens screen.
- the photographing device is installed at the front roof position of the vehicle, and points in the direction of the vehicle.
- the road sign graphic adopts a triangle, a rectangle, an arc, or other geometric element combinations that are easy to recognize.
- the road sign graphic is a bar code or a two-dimensional code.
- the identification point of the road identification graphic is its geometric center.
- the deep neural network adopts ResNet50 network, and replaces the last classification output layer of the network with two fully connected layers with 1024 nodes, and the fully connected layer is connected with an output layer with 2 nodes output.
- the deep neural network adopts a ResNet50 network, and replaces the final classification output layer of the network with two fully connected layers with 2048 nodes, and the fully connected layer is connected with an output layer with 2 nodes output.
- the floating point data output by the two nodes belong to the closed interval of [0, 1], and the pixel coordinates can be obtained by multiplying the output floating point data by the corresponding image width and height.
- a second aspect of the present invention provides a vehicle positioning method using the above deep neural network training method, including the following steps:
- Road sign pattern recognition step use the deep neural network after training to identify the road sign pattern photographed during the actual pit stop of the vehicle and obtain its sign point P in the image coordinate system Position coordinates (u, v);
- Vehicle positioning step determine the distance between the shooting device and the edge of the station in the direction of the station according to the distance between the obtained road marking graphic identification point P and the shooting device, and then combine the shooting device on the vehicle To determine the distance between the vehicle and the edge of the station into the station.
- the marking point P of the road marking graphic is on the optical axis of the lens of the shooting device;
- the origin of the coordinate system of the shooting device is set at the position of the imaging aperture of the shooting device, and the horizontal distance between the optical center of the lens of the shooting device and the marking point P of the road marking graphic is Z C ;
- the positive direction of the Z axis of the camera coordinate system is selected as the forward direction of the vehicle, the positive direction of the Y axis of the camera coordinate system is selected as the downward direction of the vehicle, and the positive X axis of the camera coordinate system is positive Select the right direction of the vehicle;
- the world coordinate system coincides with the camera coordinate system
- the origin of the image coordinate system is on the Z axis of the camera coordinate system, and the X axis and Y axis of the image coordinate system are parallel to the X axis and Y axis of the camera coordinate system, respectively;
- the image sample collection process is carried out in multiple periods under different lighting and weather conditions, which reduces the influence of environmental factors on the training results and improves the environmental adaptability of the deep neural network.
- the above method can provide the distance data between the vehicle and the station, provide data support for vehicle positioning, inbound route planning, and speed control, and has the advantages of simple operation, low cost, and high reliability.
- FIG. 1 is a flowchart of a deep neural network training method for road sign recognition
- Figure 2 is a schematic diagram of shooting in the image sample collection step
- FIG. 3 is a flowchart of a vehicle positioning method based on a deep neural network after training is completed
- Figure 4 is a side view of the vehicle during the vehicle entering the station
- 5 is a plan view of the vehicle during the vehicle entering the station
- Fig. 6 is a schematic diagram of calculating the edge distance between the vehicle and the station in the direction of the station.
- FIG. 1 is a flowchart of a deep neural network training method for road sign recognition provided by the present invention, including a road sign graphic setting step 101, a camera setting step 102, an image sample acquisition step 103, a training sample production step 104, Deep neural network construction step 105, deep neural network training step 106.
- Road sign graphic setting step 101 Set a road sign graphic on the road surface of the station inbound direction, and the distance between the marking point of the road sign graphic and the edge of the station inbound direction is L.
- the road marking graphics may be, but not limited to, triangles, rectangles, arcs, or other easily identifiable combinations of geometric elements, or text graphics, or bar codes or two-dimensional codes incorporating station-related information.
- the identification point of the road identification graphic may be a geometric center point, vertex or other geometric feature point of the road identification graphic.
- Shooting device setting step 102 Install a shooting device on the vehicle.
- the lens of the shooting device points to the direction of the vehicle. It is installed on the roof of the front of the vehicle or other position where the road sign can be photographed.
- the longitudinal symmetry centerline of the vehicle body coincides, and the distance H between the lens optical center of the shooting device and the ground is recorded.
- the lens parameters of the shooting device are selected such that, when all the road sign graphics appear in the lens screen, the road sign graphics can occupy more than 20% of the area of the lens screen, and the larger the area, the sign point of the road sign graphics is The more precise the positioning.
- Image sample collection step 103 Under different lighting or weather conditions, such as sunny noon and sunny night, rainy noon and rainy night, foggy noon and foggy night, the road sign pattern is photographed using the above-mentioned shooting device.
- the shooting angle is shown in Figure 2, where the letter A represents the road sign graphic, and the angle between the optical axis of the camera and the road surface is changed in the direction of the vehicle's advance and the direction perpendicular to the direction of the vehicle's direction, so that the camera is in its An image sample of a road marking image is taken every 5 ° within an angle of 5 ° to 180 ° between the optical axis and the road surface.
- Training sample production step 104 Calculate the position coordinates of the road marking graphics in each image sample in the image coordinate system, make a label set, and pair each image sample with the corresponding label set to form a training sample, so that Then input deep neural network for training.
- Deep neural network construction step 105 Use a target recognition classification deep neural network, but modify the final classification output layer of the network to an output layer composed of two nodes, the values output by these two nodes are the identification points of the road marking graphics. Coordinates in the image frame. More specifically, the ResNet50 network can be used, and the final classification output layer is removed. According to the required recognition effect, two fully connected layers with 1024 nodes or 2048 nodes are used instead. After the fully connected layer, there are 2 connected In the output layer output by the node, the floating point data output by these two nodes belongs to the closed interval of [0, 1], and the pixel coordinates can be obtained by multiplying the output floating point data by the corresponding image width and height.
- Deep neural network training step 106 The aforementioned training samples are input to the optimized deep neural network for training. After the training is completed, the deep neural network can be used to identify the road sign graphic and obtain the geometric center position coordinates.
- FIG. 3 is a flowchart of a vehicle positioning method using the above deep neural network training method provided by the present invention, including a road sign pattern recognition step 201, a road sign pattern positioning step 202, and a vehicle positioning step 203.
- Road sign pattern recognition step 201 Using the deep neural network after the training is completed, the road sign graphics photographed during the actual pit stop of the vehicle are recognized and the position coordinates (u, v) of the sign point P in the image coordinate system are obtained .
- Road sign graphic positioning step 202 Calculate the coordinates (X w , Y w , Z w ) of the road sign graphic point P in the world coordinate system through the transformation relationship between the image coordinate system and the world coordinate system, thereby obtaining the road sign graphic The distance between the marking point P and the camera.
- the transformation between the above image coordinate system and the world coordinate system can be described using a small hole imaging model.
- Z C represents the horizontal distance between the road marking graphic marking point P and the optical center of the camera lens
- d x , d y , u 0 , v 0 , f are the internal lens parameters related to the camera lens, specific for:
- d x , d y represent the physical length of the unit pixel in the X and Y directions of the image coordinate system; u 0 , v 0 respectively represent the origin of the image coordinate system and the origin of the camera coordinate system in the X and Y directions Offset; f represents the lens imaging focal length.
- R represents the rotation relationship between the world coordinate system and the camera coordinate system
- formula (2) is used to calculate:
- ⁇ , ⁇ , and ⁇ respectively represent the angles required to rotate around the X axis, Y axis, and Z axis when the world coordinate system and the camera coordinate system coincide.
- T represents the translation relationship between the world coordinate system and the camera coordinate system
- formula (3) is used to calculate:
- t x , t y , and t z represent the translation amounts of the world coordinate system and the camera coordinate system on the X axis, Y axis, and Z axis, respectively.
- the above parameters d x , d y , u 0 , v 0 , f, ⁇ , ⁇ , ⁇ , t x , t y , t z can be calibrated using but not limited to the conditions described below.
- the shooting device is installed at the front roof of the vehicle and points in the forward direction.
- the axis of the lens optical center of the shooting device coincides with the geometric symmetry centerline of the longitudinal axis of the vehicle.
- the marking point P of the road marking on the road surface in front of the vehicle is on the axis of the lens optical center of the shooting device, and the horizontal distance from the lens optical center of the shooting device is Z C .
- Set the origin of the camera coordinate system to the location of the imaging aperture of the camera.
- the world coordinate system coincides with the camera coordinate system, and the forward direction of the vehicle is selected as the positive direction of the Z axis, the downward direction of the vehicle is the positive direction of the Y axis, and the right direction of the vehicle is the positive direction of the X axis.
- the origin of the image coordinate system is on the Z axis of the camera coordinate system, and the X axis and Y axis of the image coordinate system are parallel to the X axis and Y axis of the camera coordinate system, respectively.
- Vehicle positioning step 203 As shown in FIG. 6, after obtaining the horizontal distance Z C of the marking point P of the road marking graphic from the optical center of the lens of the shooting device, and then combining the distance L of the marking point P of the road marking graphic from the edge of the station in the direction of stop , You can calculate the horizontal distance L CZ of the optical center of the camera and the edge of the station in the direction of the station:
- the distance between the vehicle and the edge of the station entering direction is determined, so as to realize vehicle positioning.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- General Health & Medical Sciences (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Evolutionary Biology (AREA)
- Software Systems (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Multimedia (AREA)
- Automation & Control Theory (AREA)
- Traffic Control Systems (AREA)
- Image Analysis (AREA)
Abstract
A vehicle positioning method based on deep neural network image recognition, and a training method for a deep neural network. The training method comprises: road marking graphic configuration (101), photographing device configuration (102), image sample collection (103), training sample making (104), deep neural network building (105), and deep neural network training (106). The image sample collection process is carried out in different periods under different illumination and weather conditions, so that the environmental adaptability of the deep neural network is improved. In addition, by taking sample images every certain angle in a traveling direction of a vehicle and the direction perpendicular to the traveling direction of the vehicle, a large amount of training sample data is obtained, the training precision for a deep neural network is improved, and thus the precision of vehicle positioning is improved.
Description
本发明涉及图像识别及定位技术领域,具体为一种基于深度神经网络图像识别的车辆定位方法,以及该深度神经网络的训练方法。The invention relates to the technical field of image recognition and positioning, in particular to a vehicle positioning method based on deep neural network image recognition, and a training method of the deep neural network.
目前,公共交通车辆在进入站点前,司机仅靠目视来判断车辆与站点之间的距离,无法实现精确的进站路线规划和速度控制。为了能让车辆在进站前实现精确的进站路线规划和速度控制,需要精确定位车辆与站点之间的距离。目前,车辆定位技术主要采用GPS技术和高精地图匹配定位。At present, before a public transportation vehicle enters a station, the driver only judges the distance between the vehicle and the station by visual inspection, and cannot achieve accurate inbound route planning and speed control. In order to enable vehicles to achieve accurate pitting route planning and speed control before entering the station, it is necessary to accurately locate the distance between the vehicle and the station. At present, vehicle positioning technology mainly uses GPS technology and high-precision map matching positioning.
GPS技术在使用时存在以下几个问题:使用普通GPS模式定位时,定位误差达到米级,不能满足车辆靠站精度要求;若使用GPS的RTK模式,需要同时获取卫星信息和地面的参考定位信息,需要在道路沿线布设参考定位通信设备,设备成本以及使用成本较高;当车辆进入卫星信号较差路段例如密林,隧道内时,GPS信号容易丢失,从而失去定位信息。The GPS technology has the following problems when it is used: when using the ordinary GPS mode for positioning, the positioning error reaches the meter level, which cannot meet the accuracy requirements of the vehicle; if the GPS RTK mode is used, it is necessary to obtain both satellite information and ground reference positioning information It is necessary to install reference positioning communication equipment along the road. The equipment cost and use cost are high. When the vehicle enters the road section with poor satellite signal, such as dense forest or tunnel, the GPS signal is easy to be lost, thereby losing positioning information.
高精地图匹配定位一般采用点云数据匹配或立体视觉匹配,需要预先建立地图数据并存储于车辆上,车辆运行时,通过外部激光雷达或摄像装置来获取车辆当前环境的点云数据或图像数据,并与预先存储的地图数据进行匹配。该定位方法地图制作成本以及匹配计算的软硬件成本较高。High-precision map matching and positioning generally use point cloud data matching or stereo vision matching. Map data needs to be established and stored on the vehicle in advance. When the vehicle is running, point cloud data or image data of the current environment of the vehicle is obtained through an external laser radar or camera device And match with pre-stored map data. The cost of map making and software and hardware cost of matching calculation are relatively high.
因此,需要一种低成本、高精度的车辆定位方法,为车辆定位、进站路线规划以及速度控制提供可靠的数据支持。Therefore, a low-cost, high-precision vehicle positioning method is needed to provide reliable data support for vehicle positioning, pit route planning, and speed control.
发明内容Summary of the invention
本发明提供了一种基于深度神经网络图像识别的车辆定位方法,以及该深度神经网络的训练方法。本发明通过增大训练样本量、优化神经网络参数,能够提高深度神经网络的训练精度,从而提高车辆的定位精度,且所需的设备成本和使用成本较低。The invention provides a vehicle positioning method based on deep neural network image recognition, and a training method of the deep neural network. The invention can increase the training accuracy of the deep neural network by increasing the number of training samples and optimizing the parameters of the neural network, thereby improving the positioning accuracy of the vehicle, and the required equipment cost and use cost are low.
本发明第一方面提供了一种用于道路标识识别的深度神经网络训练方法,包括以下步骤:The first aspect of the present invention provides a deep neural network training method for road sign recognition, including the following steps:
(1)道路标识图形设置步骤:在站点进站方向的路面设置道路标识图形,所述道路标识图形的标识点距离所述站点进站方向边缘的距离为L;(1) Road sign graphic setting step: set a road sign graphic on the road surface of the station inbound direction, and the distance between the marking point of the road sign graphic and the edge of the station inbound direction is L;
(2)拍摄装置设置步骤:在车辆上安装拍摄装置,所述拍摄装置的光轴线与所述车辆车身的纵向对称中心线重合,所述拍摄装置的镜头光心到地面的距离为H;(2) Steps for setting the camera: install the camera on the vehicle, the optical axis of the camera coincides with the longitudinal centerline of the vehicle body, and the distance between the lens center of the camera and the ground is H
(3)图像样本采集步骤:在不同光照或天气条件下,使用所述拍摄装置对所述道路标识图形进行拍摄,分别在所述车辆的前进方向以及与所述车辆前进方向垂直的方向上,改变所述拍摄装置的光轴与路面的夹角角度,使得所述拍摄装置在其光轴与路面夹角呈一定范围内每隔一定角度拍摄一张所述道路标识图形的图像样本;(3) Image sample collection step: under different lighting or weather conditions, the road sign graphic is photographed using the shooting device, respectively in the direction of the vehicle's advance and the direction perpendicular to the direction of the vehicle's advance, Changing the angle between the optical axis of the photographing device and the road surface, so that the photographing device takes an image sample of the road marking graphic at a certain angle within a certain range of the angle between the optical axis and the road surface;
(4)训练样本制作步骤:计算每张所述图像样本中所述道路标识图形的标识点在图像坐标系中的位置坐标,制作成标签集,并将每张所述图像样本与相应的标签集配对,形成训练样本;(4) Training sample production step: calculate the position coordinates of the identification point of the road sign graphic in the image coordinate system in each of the image samples, make a label set, and combine each of the image samples with the corresponding labels Set pairing to form training samples;
(5)深度神经网络搭建步骤:在目标识别分类深度神经网络的基础上,将所述网络最后的分类输出层修改为2个节点构成的输出层,以输出所述道路标识图形的标识点位置坐标;(5) Steps for building a deep neural network: on the basis of the target recognition classification deep neural network, the final classification output layer of the network is modified into an output layer composed of 2 nodes to output the position of the marking point of the road marking graphic coordinate;
(6)深度神经网络训练步骤:将所述训练样本输入到所述深度神经网络进行训练。(6) Deep neural network training step: input the training samples to the deep neural network for training.
优选地,所述图像样本采集步骤中,拍摄时间选择在晴天正午和晴天夜晚。Preferably, in the image sample collection step, the shooting time is selected at noon on a sunny day and at night on a sunny day.
优选地,所述图像样本采集步骤中,拍摄时间选择在雨天正午和雨天夜晚。Preferably, in the image sample collection step, the shooting time is selected at noon on rainy days and night on rainy days.
优选地,所述图像样本采集步骤中,拍摄时间选择在雾天正午和雾天夜晚。Preferably, in the image sample collection step, the shooting time is selected at noon on a foggy day and at night on a foggy day.
优选地,所述图像样本采集步骤中,所述拍摄装置在其光轴与路面夹角呈5°到180°的范围内每隔5°拍摄一张所述道路标识图形的图像样本。Preferably, in the image sample collection step, the photographing device photographs an image sample of the road marking pattern every 5 ° within an angle range of 5 ° to 180 ° between the optical axis and the road surface.
优选地,所述拍摄装置的镜头参数选择为,当所述道路标识图形全部出现在镜头画面中时,所述道路标识图形能够占据所述镜头画面20%以上的面积。Preferably, the lens parameters of the shooting device are selected so that when all the road sign graphics appear in the lens screen, the road sign graphics can occupy more than 20% of the area of the lens screen.
优选地,所述拍摄装置安装于所述车辆前部车顶位置,并指向所述车辆前进方向。Preferably, the photographing device is installed at the front roof position of the vehicle, and points in the direction of the vehicle.
优选地,所述道路标识图形采用三角形,或矩形,或圆弧,或其他易于识别的几何元素组合。Preferably, the road sign graphic adopts a triangle, a rectangle, an arc, or other geometric element combinations that are easy to recognize.
优选地,所述道路标识图形采用条形码或二维码。Preferably, the road sign graphic is a bar code or a two-dimensional code.
优选地,所述道路标识图形的标识点为其几何中心。Preferably, the identification point of the road identification graphic is its geometric center.
优选地,所述深度神经网络采用ResNet50网络,将所述网络最后的分类输出层替换为两层具有1024个节点的全连接层,所述全连接层后连接具有2个节点输出的输出层。Preferably, the deep neural network adopts ResNet50 network, and replaces the last classification output layer of the network with two fully connected layers with 1024 nodes, and the fully connected layer is connected with an output layer with 2 nodes output.
优选地,所述深度神经网络采用ResNet50网络,将所述网络最后的分类输出层替换为两层具有2048个节点的全连接层,所述全连接层后连接具有2个节点输出的输出层。Preferably, the deep neural network adopts a ResNet50 network, and replaces the final classification output layer of the network with two fully connected layers with 2048 nodes, and the fully connected layer is connected with an output layer with 2 nodes output.
优选地,所述2个节点输出的浮点数据属于[0,1]的闭区间,将所述输出浮点数据与相应的图像宽度和高度相乘即可获得像素坐标。Preferably, the floating point data output by the two nodes belong to the closed interval of [0, 1], and the pixel coordinates can be obtained by multiplying the output floating point data by the corresponding image width and height.
本发明第二方面提供了一种使用上述深度神经网络训练方法的车辆定位方法,包括以下步骤:A second aspect of the present invention provides a vehicle positioning method using the above deep neural network training method, including the following steps:
(1)道路标识图形识别步骤:利用训练完成后的所述深度神经网络,对所述车辆实际进站过程中拍摄到的所述道路标识图形进行识别并获取其标识点P在图像坐标系中的位置坐标(u,v);(1) Road sign pattern recognition step: use the deep neural network after training to identify the road sign pattern photographed during the actual pit stop of the vehicle and obtain its sign point P in the image coordinate system Position coordinates (u, v);
(2)道路标识图形定位步骤:通过图像坐标系与世界坐标系的变换关系,计算出所述道路标识图形标识点P在世界坐标系中的坐标(X
w,Y
w,Z
w),从而获得所述道路标识图形标识点P与所述拍摄装置的距离;
(2) Road sign graphic positioning step: Through the transformation relationship between the image coordinate system and the world coordinate system, the coordinates (X w , Y w , Z w ) of the road sign graphic identification point P in the world coordinate system are calculated, thereby Obtaining the distance between the road marking graphic marking point P and the shooting device;
(3)车辆定位步骤:根据所获得的道路标识图形标识点P与所述拍摄装置的距离,确定所述拍摄装置与站点进站方向边缘的距离,再结合所述拍摄装置在所述车辆上的安装位置,确定所述车辆与站点进站方向边缘的距离。(3) Vehicle positioning step: determine the distance between the shooting device and the edge of the station in the direction of the station according to the distance between the obtained road marking graphic identification point P and the shooting device, and then combine the shooting device on the vehicle To determine the distance between the vehicle and the edge of the station into the station.
优选地,所述道路标识图形的标识点P在所述拍摄装置镜头光心轴线上;Preferably, the marking point P of the road marking graphic is on the optical axis of the lens of the shooting device;
所述拍摄装置坐标系原点设定在所述拍摄装置成像小孔位置,所述拍摄装置镜头光心与所述道路标识图形标识点P的水平距离为Z
C;
The origin of the coordinate system of the shooting device is set at the position of the imaging aperture of the shooting device, and the horizontal distance between the optical center of the lens of the shooting device and the marking point P of the road marking graphic is Z C ;
所述拍摄装置坐标系的Z轴正向选取为所述车辆前进方向,所述拍摄装置坐标系的Y轴正向选取为所述车辆向下方向,所述拍摄装置坐标系的X轴正向选取为所述车辆向右方向;The positive direction of the Z axis of the camera coordinate system is selected as the forward direction of the vehicle, the positive direction of the Y axis of the camera coordinate system is selected as the downward direction of the vehicle, and the positive X axis of the camera coordinate system is positive Select the right direction of the vehicle;
所述世界坐标系与所述拍摄装置坐标系重合;The world coordinate system coincides with the camera coordinate system;
所述图像坐标系的原点在所述拍摄装置坐标系Z轴上,所述图像坐标系的X轴和Y轴分别与所述拍摄装置坐标系的X轴和Y轴平行;The origin of the image coordinate system is on the Z axis of the camera coordinate system, and the X axis and Y axis of the image coordinate system are parallel to the X axis and Y axis of the camera coordinate system, respectively;
根据公式:According to the formula:
得到所述拍摄装置与所述道路标识图形标识点Pi的水平距离Z
C;
Obtain the horizontal distance Z C between the photographing device and the road identification graphic identification point Pi;
再根据公式:Then according to the formula:
L
cz=Z
c+L
L cz = Z c + L
得到所述拍摄装置与站点进站方向边缘的水平距离L
CZ。
Obtain the horizontal distance L CZ between the shooting device and the edge of the station in the direction of the station.
本发明的优点在于:The advantages of the present invention are:
(1)图像样本的采集过程在不同光照及天气条件下的多个时段内进行,减小了环境因素对训练结果的影响,提高了深度神经网络的环境适应性。(1) The image sample collection process is carried out in multiple periods under different lighting and weather conditions, which reduces the influence of environmental factors on the training results and improves the environmental adaptability of the deep neural network.
(2)在车辆前进方向和与前进方向垂直的方向上每隔一定角度拍摄一张采样图像,获得的训练样本数据量大,提高了深度神经网络的训练精度,从而提高了后续车辆定位的精度。(2) Take a sampled image at a certain angle in the vehicle's direction of travel and the direction perpendicular to the direction of travel. The amount of training samples obtained is large, which improves the training accuracy of the deep neural network and thus improves the accuracy of subsequent vehicle positioning .
(3)利用训练好的深度神经网络来识别设置于站点前方的道路标识图形,并通过图像坐标系与世界坐标系的转换,推算出车载拍摄装置的空间位置进而定位车辆位置。上述方法能够提供车辆与站点之间的距离数据,为车辆定位、进站路线规划以及速度控制提供数据支持,具有操作简单、成本较低、可靠性高的优势。(3) Use the trained deep neural network to identify the road sign graphic set in front of the station, and through the conversion of the image coordinate system and the world coordinate system, calculate the spatial position of the on-board shooting device and then locate the vehicle position. The above method can provide the distance data between the vehicle and the station, provide data support for vehicle positioning, inbound route planning, and speed control, and has the advantages of simple operation, low cost, and high reliability.
本发明的以上内容以及下面的具体实施方式在结合附图阅读时会得到更好的理解。需要说明的是,附图仅作为所请求保护的发明的示例。在附图中,相同的附图标记代表相同或类似的元素。The above content of the present invention and the following specific embodiments will be better understood when read in conjunction with the accompanying drawings. It should be noted that the drawings are only examples of the claimed invention. In the drawings, the same reference numerals represent the same or similar elements.
图1为用于道路标识识别的深度神经网络训练方法的流程图;FIG. 1 is a flowchart of a deep neural network training method for road sign recognition;
图2为图像样本采集步骤中的拍摄示意图;Figure 2 is a schematic diagram of shooting in the image sample collection step;
图3为一种基于训练完成后的深度神经网络的车辆定位方法的流程图;3 is a flowchart of a vehicle positioning method based on a deep neural network after training is completed;
图4为车辆进站过程中的车辆侧视图;Figure 4 is a side view of the vehicle during the vehicle entering the station;
图5为车辆进站过程中的车辆俯视图;5 is a plan view of the vehicle during the vehicle entering the station;
图6为车辆与站点进站方向边缘距离计算示意图。Fig. 6 is a schematic diagram of calculating the edge distance between the vehicle and the station in the direction of the station.
以下结合附图和实施例对本发明作进一步的详细说明。The present invention will be further described in detail below with reference to the drawings and embodiments.
图1为本发明提供的一种用于道路标识识别的深度神经网络训练方法的流程图,包括道路标识图形设置步骤101、拍摄装置设置步骤102、图像样本采集步骤103、训练样本制作步骤104、深度神经网络搭建步骤105、深度神经网络训练步骤106。1 is a flowchart of a deep neural network training method for road sign recognition provided by the present invention, including a road sign graphic setting step 101, a camera setting step 102, an image sample acquisition step 103, a training sample production step 104, Deep neural network construction step 105, deep neural network training step 106.
道路标识图形设置步骤101:在站点进站方向的路面设置道路标识图形,该道路标识图形的标识点距离站点进站方向边缘的距离为L。该道路标识图形可采用但不限于三角形、矩形、圆弧,或其他易于识别的几何元素组合,或文字图形,或融入了车站相关信息的条形码或二维码。该道路标识图形的标识点可以是该道路标识图形的几何中心点、顶点或其他几何特征点。Road sign graphic setting step 101: Set a road sign graphic on the road surface of the station inbound direction, and the distance between the marking point of the road sign graphic and the edge of the station inbound direction is L. The road marking graphics may be, but not limited to, triangles, rectangles, arcs, or other easily identifiable combinations of geometric elements, or text graphics, or bar codes or two-dimensional codes incorporating station-related information. The identification point of the road identification graphic may be a geometric center point, vertex or other geometric feature point of the road identification graphic.
拍摄装置设置步骤102:在车辆上安装拍摄装置,该拍摄装置的镜头指向车辆前进方向,安装在车辆前部车顶位置或其他能够拍摄到道路标识图形的位置,并使拍摄装置的光轴线与车辆车身的纵向对称中心线重合,记录该拍摄装置的镜头光心到地面的距离H。该拍摄装置的镜头参数选择为,当前述道路标识图形全部出现在镜头画面中时,道路标识图形能够占据镜头画面20%以上的面积,占据的面积越大,道路标识图形的标识点在画面中的定位就越精确。Shooting device setting step 102: Install a shooting device on the vehicle. The lens of the shooting device points to the direction of the vehicle. It is installed on the roof of the front of the vehicle or other position where the road sign can be photographed. The longitudinal symmetry centerline of the vehicle body coincides, and the distance H between the lens optical center of the shooting device and the ground is recorded. The lens parameters of the shooting device are selected such that, when all the road sign graphics appear in the lens screen, the road sign graphics can occupy more than 20% of the area of the lens screen, and the larger the area, the sign point of the road sign graphics is The more precise the positioning.
图像样本采集步骤103:在不同光照或天气条件下,例如晴天正午和晴天夜晚,雨天正午和雨天夜晚,雾天正午和雾天夜晚,使用上述拍摄装置对道路标识图形进行拍摄。拍摄角度如图2所示,其中字母A表示道路标识图形,分别在车辆的前进方向以及与车辆前进方向垂直的方向上,改变拍摄装置的光轴与路面的夹角角度,使得拍摄装置在其光轴与路面夹角呈5°到180°的范围内每隔5°拍摄一张道路标识图形的图像样本。Image sample collection step 103: Under different lighting or weather conditions, such as sunny noon and sunny night, rainy noon and rainy night, foggy noon and foggy night, the road sign pattern is photographed using the above-mentioned shooting device. The shooting angle is shown in Figure 2, where the letter A represents the road sign graphic, and the angle between the optical axis of the camera and the road surface is changed in the direction of the vehicle's advance and the direction perpendicular to the direction of the vehicle's direction, so that the camera is in its An image sample of a road marking image is taken every 5 ° within an angle of 5 ° to 180 ° between the optical axis and the road surface.
训练样本制作步骤104:计算每张图像样本中道路标识图形的标识点在图像坐标系中的位置坐标,制作成标签集,并将每张图像样本与相应的标签集配对,形成训练样本,以便后续输入深度神经网络进行训练。Training sample production step 104: Calculate the position coordinates of the road marking graphics in each image sample in the image coordinate system, make a label set, and pair each image sample with the corresponding label set to form a training sample, so that Then input deep neural network for training.
深度神经网络搭建步骤105:使用一目标识别分类深度神经网络,但将网络最后的分类输出层修改成两个节点构成的输出层,这两个节点输出的数值即为道路标识图形的标识点在图像画面中的坐标。更具体地,可采用ResNet50网络,并将最后的分类输出层去除,根据所需的识别效果,使用两层具有1024个节点或2048个节点的全连接层代替,全连接层后连接具有2个节点输出的输出层,这两个节点输出的浮点数据属于[0,1]的闭区间,将输出的浮点数据乘以各自对应的图像宽度和高度即可获得像素坐标。Deep neural network construction step 105: Use a target recognition classification deep neural network, but modify the final classification output layer of the network to an output layer composed of two nodes, the values output by these two nodes are the identification points of the road marking graphics. Coordinates in the image frame. More specifically, the ResNet50 network can be used, and the final classification output layer is removed. According to the required recognition effect, two fully connected layers with 1024 nodes or 2048 nodes are used instead. After the fully connected layer, there are 2 connected In the output layer output by the node, the floating point data output by these two nodes belongs to the closed interval of [0, 1], and the pixel coordinates can be obtained by multiplying the output floating point data by the corresponding image width and height.
深度神经网络训练步骤106:将前述训练样本输入到优化后的深度神经网络进行训练,训练完成后,即可使用该深度神经网络对道路标识图形进行识别并获取其几何中心位置坐标。Deep neural network training step 106: The aforementioned training samples are input to the optimized deep neural network for training. After the training is completed, the deep neural network can be used to identify the road sign graphic and obtain the geometric center position coordinates.
图3为本发明提供的一种使用上述深度神经网络训练方法的车辆定位方法的流程图,包括道路标识图形识别步骤201、道路标识图形定位步骤202、车辆定位步骤203。3 is a flowchart of a vehicle positioning method using the above deep neural network training method provided by the present invention, including a road sign pattern recognition step 201, a road sign pattern positioning step 202, and a vehicle positioning step 203.
道路标识图形识别步骤201:利用训练完成后的深度神经网络,对车辆实际进站过程中拍摄到的道路标识图形进行识别并获取其标识点P在图像坐标系中的位置坐标(u,v)。Road sign pattern recognition step 201: Using the deep neural network after the training is completed, the road sign graphics photographed during the actual pit stop of the vehicle are recognized and the position coordinates (u, v) of the sign point P in the image coordinate system are obtained .
道路标识图形定位步骤202:通过图像坐标系与世界坐标系的变换关系,计算出道路标识图形标识点P在世界坐标系中的坐标(X
w,Y
w,Z
w),从而获得道路标识图形标识点P与拍摄装置的距离。上述图像坐标系与世界坐标系的变换可采用小孔成像模型来描述。世界坐标系中的一点P
w,其在世界坐标系中的坐标为(X
w,Y
w,Z
w),通过镜头成像到二维图像坐标系中的P
i点,其坐标为(u,v),则P
w与P
i点的坐标可以使用公式(1)进行换算:
Road sign graphic positioning step 202: Calculate the coordinates (X w , Y w , Z w ) of the road sign graphic point P in the world coordinate system through the transformation relationship between the image coordinate system and the world coordinate system, thereby obtaining the road sign graphic The distance between the marking point P and the camera. The transformation between the above image coordinate system and the world coordinate system can be described using a small hole imaging model. A point P w in the world coordinate system, whose coordinates in the world coordinate system are (X w , Y w , Z w ), and the point P i in the two-dimensional image coordinate system imaged by the lens, the coordinates are (u, v), the coordinates of P w and P i points can be converted using formula (1):
(1)式中,Z
C表示道路标识图形标识点P与拍摄装置镜头光心的水平距离;d
x,d
y,u
0,v
0,f是与拍摄装置镜头有关的镜头内部参数,具体为:
(1) In the formula, Z C represents the horizontal distance between the road marking graphic marking point P and the optical center of the camera lens; d x , d y , u 0 , v 0 , f are the internal lens parameters related to the camera lens, specific for:
d
x,d
y分别表示图像坐标系X方向和Y方向单位像素的物理长度;u
0,v
0分别表示图像坐标系中,图像坐标系原点与拍摄装置坐标系原点在X方向和Y方向的偏移;f表示镜头成像焦距。
d x , d y represent the physical length of the unit pixel in the X and Y directions of the image coordinate system; u 0 , v 0 respectively represent the origin of the image coordinate system and the origin of the camera coordinate system in the X and Y directions Offset; f represents the lens imaging focal length.
(1)式中,R表示世界坐标系与拍摄装置坐标系的旋转关系,采用(2)式计算:In the formula (1), R represents the rotation relationship between the world coordinate system and the camera coordinate system, and the formula (2) is used to calculate:
其中α、β、γ分别表示世界坐标系与拍摄装置坐标系重合时,需要围绕X轴、Y轴、Z轴转动的角度。Where α, β, and γ respectively represent the angles required to rotate around the X axis, Y axis, and Z axis when the world coordinate system and the camera coordinate system coincide.
(1)式中,T表示世界坐标系与拍摄装置坐标系的平移关系,采用(3)式计算:In the formula (1), T represents the translation relationship between the world coordinate system and the camera coordinate system, and the formula (3) is used to calculate:
T=[t
x t
y t
z]
T (3)
T = [t x t y t z ] T (3)
其中t
x、t
y、t
z分别表示世界坐标系与拍摄装置坐标系在X轴、Y轴、Z轴的平移量。
Where t x , t y , and t z represent the translation amounts of the world coordinate system and the camera coordinate system on the X axis, Y axis, and Z axis, respectively.
具体实施时,以上参数d
x,d
y,u
0,v
0,f,α,β,γ,t
x,t
y,t
z可以采用但不限于以下所描述的情况进行标定。
During specific implementation, the above parameters d x , d y , u 0 , v 0 , f, α, β, γ, t x , t y , t z can be calibrated using but not limited to the conditions described below.
如图4、图5所示,拍摄装置安装于车辆前部车顶位置,并指向前进方向。拍摄装置的镜头光心轴线与车辆纵轴几何对称中心线重合,车辆前方路面的道路标识图形的标识点P在拍摄装置镜头光心轴线上,其与拍摄装置镜头光心的水平距离为Z
C。将拍摄装置坐标系原点设定在拍摄装置成像小孔位置。为简化计算,假设世界坐标系与拍摄装置坐标系重合,且选取车辆前进方向为Z轴正向,车辆向下为Y轴正方向,车辆向右为X轴正向。图像坐标系的原点在拍摄装置坐标系的Z轴上,图像坐标系的X轴和Y轴分别与拍摄装置坐标系的X轴和Y轴平行。
As shown in FIGS. 4 and 5, the shooting device is installed at the front roof of the vehicle and points in the forward direction. The axis of the lens optical center of the shooting device coincides with the geometric symmetry centerline of the longitudinal axis of the vehicle. The marking point P of the road marking on the road surface in front of the vehicle is on the axis of the lens optical center of the shooting device, and the horizontal distance from the lens optical center of the shooting device is Z C . Set the origin of the camera coordinate system to the location of the imaging aperture of the camera. To simplify the calculation, it is assumed that the world coordinate system coincides with the camera coordinate system, and the forward direction of the vehicle is selected as the positive direction of the Z axis, the downward direction of the vehicle is the positive direction of the Y axis, and the right direction of the vehicle is the positive direction of the X axis. The origin of the image coordinate system is on the Z axis of the camera coordinate system, and the X axis and Y axis of the image coordinate system are parallel to the X axis and Y axis of the camera coordinate system, respectively.
根据以上条件,可知P点在世界坐标系中的坐标为(X
w,Y
w,Z
w),在拍摄装置坐标系中的坐标为(X
c,Y
c,Z
c),并且X
w=X
c=0,Y
w=H,Z
w=Z
c。P点经过拍摄装置小孔成像后,在图像坐标系中的坐标为(u,v),且u=0,u
0=0,v
0=0。拍摄装置坐标系与世界坐标系的平移参数t
x=t
y=t
z=0。从而式(1)可以化简为:
According to the above conditions, the coordinates of point P in the world coordinate system are (X w , Y w , Z w ), the coordinates in the camera coordinate system are (X c , Y c , Z c ), and X w = X c = 0, Y w = H, Z w = Z c . After the point P is imaged through the aperture of the camera, the coordinates in the image coordinate system are (u, v), and u = 0, u 0 = 0, v 0 = 0. The translation parameter t x = t y = t z = 0 of the camera coordinate system and the world coordinate system. Thus equation (1) can be simplified as:
即which is
车辆定位步骤203:如图6所示,获得道路标识图形的标识点P距离拍摄装置镜头光心的水平距离Z
C后,再结合道路标识图形的标识点P距离站点进站方向边缘的距离L,可以计算得到拍摄装置镜头光心与站点进站方向边缘的水平距离L
CZ:
Vehicle positioning step 203: As shown in FIG. 6, after obtaining the horizontal distance Z C of the marking point P of the road marking graphic from the optical center of the lens of the shooting device, and then combining the distance L of the marking point P of the road marking graphic from the edge of the station in the direction of stop , You can calculate the horizontal distance L CZ of the optical center of the camera and the edge of the station in the direction of the station:
L
cz=Z
c+L (6)
L cz = Z c + L (6)
再结合该拍摄装置在车辆上的安装位置,确定车辆与站点进站方向边缘的距离,从而实现车辆定位。Combined with the installation position of the photographing device on the vehicle, the distance between the vehicle and the edge of the station entering direction is determined, so as to realize vehicle positioning.
这里基于的术语和表述方式只是用于描述,本发明并不应局限于这些术语和表述。使用这些术语和表述并不意味着排除任何示意和描述(或其中部分)的等效特征,应认 识到可能存在的各种修改也应包含在权利要求范围内。其他修改、变化和替换也可能存在。相应的,权利要求应视为覆盖所有这些等效物。The terms and expressions based here are for description only, and the present invention should not be limited to these terms and expressions. The use of these terms and expressions is not meant to exclude any schematic and descriptive (or part of) equivalent features, and it should be recognized that various modifications that may exist should also be included within the scope of the claims. Other modifications, changes and substitutions may also exist. Accordingly, the claims should be considered to cover all such equivalents.
同样,需要指出的是,虽然本发明已参照当前的具体实施例来描述,但是本技术领域中的普通技术人员应当认识到,以上的实施例仅是用来说明本发明,在没有脱离本发明精神的情况下还可做出各种等效的变化或替换,因此,只要在本发明的实质精神范围内对上述实施例的变化、变型都将落在本申请的权利要求书的范围内。Similarly, it should be pointed out that although the present invention has been described with reference to the current specific embodiments, those of ordinary skill in the art should realize that the above embodiments are only used to illustrate the present invention without departing from the present invention. Various equivalent changes or substitutions can also be made under the spirit, so as long as the changes and modifications to the above-mentioned embodiments are within the spirit of the present invention, they will fall within the scope of the claims of the present application.
Claims (15)
- 一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述方法包括:A deep neural network training method for road sign recognition, characterized in that the method includes:道路标识图形设置步骤:在站点进站方向的路面设置道路标识图形,所述道路标识图形的标识点距离所述站点进站方向边缘的距离为L;The road sign graphic setting step: set a road sign graphic on the road surface of the station inbound direction, and the distance between the mark point of the road sign graphic and the edge of the station inbound direction is L;拍摄装置设置步骤:在车辆上安装拍摄装置,所述拍摄装置的光轴线与所述车辆车身的纵向对称中心线重合,所述拍摄装置的镜头光心到地面的距离为H;The setting step of the shooting device: install the shooting device on the vehicle, the optical axis of the shooting device coincides with the longitudinal symmetry centerline of the vehicle body, and the distance from the lens optical center of the shooting device to the ground is H;图像样本采集步骤:在不同光照或天气条件下,使用所述拍摄装置对所述道路标识图形进行拍摄,分别在所述车辆的前进方向以及与所述车辆前进方向垂直的方向上,改变所述拍摄装置的光轴与路面的夹角角度,使得所述拍摄装置在其光轴与路面夹角呈一定范围内每隔一定角度拍摄一张所述道路标识图形的图像样本;Image sample collection step: under different lighting or weather conditions, use the shooting device to shoot the road sign graphic, and change the direction of the vehicle and the direction perpendicular to the direction of the vehicle respectively The angle between the optical axis of the photographing device and the road surface makes the photographing device take an image sample of the road marking graphic at a certain angle within a certain range of the angle between the optical axis and the road surface;训练样本制作步骤:计算每张所述图像样本中所述道路标识图形的标识点在图像坐标系中的位置坐标,制作成标签集,并将每张所述图像样本与相应的标签集配对,形成训练样本;Training sample production step: Calculate the position coordinates of the identification point of the road sign graphic in the image coordinate system in each image sample, make a label set, and pair each image sample with the corresponding label set, Form training samples;深度神经网络搭建步骤:在目标识别分类深度神经网络的基础上,将所述网络最后的分类输出层修改为2个节点构成的输出层,以输出所述道路标识图形的标识点位置坐标;Steps for building a deep neural network: on the basis of a deep neural network for target recognition classification, the final classification output layer of the network is modified to an output layer composed of 2 nodes to output the position coordinates of the marking points of the road marking graphics;深度神经网络训练步骤:将所述训练样本输入到所述深度神经网络进行训练。Deep neural network training step: input the training samples to the deep neural network for training.
- 根据权利要求1所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述图像样本采集步骤中,拍摄时间选择在晴天正午和晴天夜晚。The deep neural network training method for road sign recognition according to claim 1, characterized in that, in the image sample collection step, the shooting time is selected to be at noon on a sunny day and at night on a sunny day.
- 根据权利要求1所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述图像样本采集步骤中,拍摄时间选择在雨天正午和雨天夜晚。The deep neural network training method for road sign recognition according to claim 1, characterized in that, in the image sample collection step, the shooting time is selected to be noon on rainy days and night on rainy days.
- 根据权利要求1所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述图像样本采集步骤中,拍摄时间选择在雾天正午和雾天夜晚。The deep neural network training method for road sign recognition according to claim 1, characterized in that, in the image sample collection step, the shooting time is selected at noon on foggy days and at night on foggy days.
- 根据权利要求1~4所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述图像样本采集步骤中,所述拍摄装置在其光轴与路面夹角呈5°到180°的范围内每隔5°拍摄一张所述道路标识图形的图像样本。A deep neural network training method for road sign recognition according to claims 1 to 4, characterized in that in the image sample collection step, the angle between the optical axis of the shooting device and the road surface is 5 ° An image sample of the road sign graphic is taken every 5 ° within the range of 180 °.
- 根据权利要求1所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述拍摄装置的镜头参数选择为,当所述道路标识图形全部出现在镜头画面中时,所述道路标识图形能够占据所述镜头画面20%以上的面积。A deep neural network training method for road sign recognition according to claim 1, characterized in that the lens parameters of the shooting device are selected when all of the road sign graphics appear in the lens screen, The road sign graphic can occupy more than 20% of the area of the lens screen.
- 根据权利要求6所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述拍摄装置安装于所述车辆前部车顶位置,并指向所述车辆前进方向。A deep neural network training method for road sign recognition according to claim 6, characterized in that the photographing device is installed at a position on the roof of the front of the vehicle and points in the direction of the vehicle.
- 根据权利要求1所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述道路标识图形采用三角形,或矩形,或圆弧,或其他易于识别的几何元素组合。A deep neural network training method for road sign recognition according to claim 1, characterized in that the road sign graphics use triangles, or rectangles, or arcs, or other geometric element combinations that are easy to recognize.
- 根据权利要求1所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述道路标识图形采用条形码或二维码。The deep neural network training method for road sign recognition according to claim 1, characterized in that the road sign graphic is a bar code or a two-dimensional code.
- 根据权利要求1所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述道路标识图形的标识点为其几何中心。The deep neural network training method for road sign recognition according to claim 1, wherein the sign point of the road sign graphic is its geometric center.
- 根据权利要求1所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述深度神经网络采用ResNet50网络,将所述网络最后的分类输出层替换为两层具有1024个节点的全连接层,所述全连接层后连接具有2个节点输出的输出层。A deep neural network training method for road sign recognition according to claim 1, wherein the deep neural network uses a ResNet50 network, and the last classification output layer of the network is replaced with two layers with 1024 A fully connected layer of nodes, which is connected behind an output layer with two nodes output.
- 根据权利要求1所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述深度神经网络采用ResNet50网络,将所述网络最后的分类输出层替换为两层具有2048个节点的全连接层,所述全连接层后连接具有2个节点输出的输出层。A deep neural network training method for road sign recognition according to claim 1, characterized in that the deep neural network uses a ResNet50 network, and the last classification output layer of the network is replaced with two layers with 2048 A fully connected layer of nodes, which is connected behind an output layer with two nodes output.
- 根据权利要求11~12所述的一种用于道路标识识别的深度神经网络训练方法,其特征在于,所述2个节点输出的浮点数据属于[0,1]的闭区间,将所述输出浮点数据与相应的图像宽度和高度相乘即可获得像素坐标。A deep neural network training method for road sign recognition according to claims 11-12, characterized in that the floating point data output by the two nodes belongs to the closed interval of [0, 1], and the The pixel coordinates can be obtained by multiplying the output floating point data with the corresponding image width and height.
- 一种使用权利要求1所述的深度神经网络训练方法的车辆定位方法,其特征在于,所述方法包括:A vehicle positioning method using the deep neural network training method of claim 1, wherein the method comprises:道路标识图形识别步骤:利用训练完成后的所述深度神经网络,对所述车辆实际进站过程中拍摄到的所述道路标识图形进行识别并获取其标识点P在图像坐标系中的位置坐标(u,v);Road sign pattern recognition step: use the deep neural network after training to identify the road sign pattern photographed during the actual pit stop of the vehicle and obtain the position coordinates of its sign point P in the image coordinate system (u, v);道路标识图形定位步骤:通过图像坐标系与世界坐标系的变换关系,计算出所述道路标识图形标识点P在世界坐标系中的坐标(X w,Y w,Z w),从而获得所述道路标识图形标识点P与所述拍摄装置的距离; The road sign graphic positioning step: calculate the coordinates (X w , Y w , Z w ) of the road sign graphic point P in the world coordinate system through the transformation relationship between the image coordinate system and the world coordinate system, thereby obtaining the The distance between the road marking graphic marking point P and the shooting device;车辆定位步骤:根据所获得的道路标识图形标识点P与所述拍摄装置的距离,确定所述拍摄装置与站点进站方向边缘的距离,再结合所述拍摄装置在所述车辆上的安装位置,确定所述车辆与站点进站方向边缘的距离。Vehicle positioning step: determine the distance between the shooting device and the edge of the station's entry direction according to the distance between the obtained road marking graphic identification point P and the shooting device, and then combine the installation position of the shooting device on the vehicle To determine the distance between the vehicle and the edge of the station's inbound direction.
- 根据权利要求14所述的一种车辆定位方法,其特征在于,所述道路标识图形的标识点P在所述拍摄装置镜头光心轴线上;A vehicle positioning method according to claim 14, wherein the marking point P of the road marking graphic is on the optical axis of the lens of the shooting device;所述拍摄装置坐标系原点设定在所述拍摄装置成像小孔位置,所述拍摄装置镜头光心与所述道路标识图形标识点P的水平距离为Z C; The origin of the coordinate system of the shooting device is set at the position of the imaging aperture of the shooting device, and the horizontal distance between the optical center of the lens of the shooting device and the marking point P of the road marking graphic is Z C ;所述拍摄装置坐标系的Z轴正向选取为所述车辆前进方向,所述拍摄装置坐标系的Y轴正向选取为所述车辆向下方向,所述拍摄装置坐标系的X轴正向选取为所述车辆向右方向;The positive direction of the Z axis of the camera coordinate system is selected as the forward direction of the vehicle, the positive direction of the Y axis of the camera coordinate system is selected as the downward direction of the vehicle, and the positive X axis of the camera coordinate system is positive Select the right direction of the vehicle;所述世界坐标系与所述拍摄装置坐标系重合;The world coordinate system coincides with the camera coordinate system;所述图像坐标系的原点在所述拍摄装置坐标系Z轴上,所述图像坐标系的X轴和Y轴分别与所述拍摄装置坐标系的X轴和Y轴平行;The origin of the image coordinate system is on the Z axis of the camera coordinate system, and the X axis and Y axis of the image coordinate system are parallel to the X axis and Y axis of the camera coordinate system, respectively;根据公式:According to the formula:得到所述拍摄装置与所述道路标识图形标识点P的水平距离Z C; Obtain the horizontal distance Z C between the photographing device and the road marking graphic marking point P;再根据公式:Then according to the formula:L cz=Z c+L L cz = Z c + L得到所述拍摄装置与站点进站方向边缘的水平距离L CZ。 Obtain the horizontal distance L CZ between the shooting device and the edge of the station in the direction of the station.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SG11202103814PA SG11202103814PA (en) | 2018-10-24 | 2019-10-18 | Vehicle positioning method based on deep neural network image recognition |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811245274.1A CN109446973B (en) | 2018-10-24 | 2018-10-24 | Vehicle positioning method based on deep neural network image recognition |
CN201811245274.1 | 2018-10-24 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020083103A1 true WO2020083103A1 (en) | 2020-04-30 |
Family
ID=65547888
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/111840 WO2020083103A1 (en) | 2018-10-24 | 2019-10-18 | Vehicle positioning method based on deep neural network image recognition |
Country Status (3)
Country | Link |
---|---|
CN (1) | CN109446973B (en) |
SG (1) | SG11202103814PA (en) |
WO (1) | WO2020083103A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111914691A (en) * | 2020-07-15 | 2020-11-10 | 北京埃福瑞科技有限公司 | Rail transit vehicle positioning method and system |
CN113378735A (en) * | 2021-06-18 | 2021-09-10 | 北京东土科技股份有限公司 | Road marking line identification method and device, electronic equipment and storage medium |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109446973B (en) * | 2018-10-24 | 2021-01-22 | 中车株洲电力机车研究所有限公司 | Vehicle positioning method based on deep neural network image recognition |
CN110726414B (en) * | 2019-10-25 | 2021-07-27 | 百度在线网络技术(北京)有限公司 | Method and apparatus for outputting information |
CN111161227B (en) * | 2019-12-20 | 2022-09-06 | 成都数之联科技股份有限公司 | Target positioning method and system based on deep neural network |
CN113496594A (en) * | 2020-04-03 | 2021-10-12 | 郑州宇通客车股份有限公司 | Bus arrival control method, device and system |
CN112699823A (en) * | 2021-01-05 | 2021-04-23 | 浙江得图网络有限公司 | Fixed-point returning method for sharing electric vehicle |
CN112950922B (en) * | 2021-01-26 | 2022-06-10 | 浙江得图网络有限公司 | Fixed-point returning method for sharing electric vehicle |
WO2023019509A1 (en) * | 2021-08-19 | 2023-02-23 | 浙江吉利控股集团有限公司 | Environment matching-based vehicle localization method and apparatus, vehicle, and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN202350794U (en) * | 2011-11-29 | 2012-07-25 | 高德软件有限公司 | Navigation data acquisition device |
CN103925927A (en) * | 2014-04-18 | 2014-07-16 | 中国科学院软件研究所 | Traffic sign positioning method based on vehicle-mounted video |
CN108009518A (en) * | 2017-12-19 | 2018-05-08 | 大连理工大学 | A kind of stratification traffic mark recognition methods based on quick two points of convolutional neural networks |
US20180211120A1 (en) * | 2017-01-25 | 2018-07-26 | Ford Global Technologies, Llc | Training An Automatic Traffic Light Detection Model Using Simulated Images |
CN109446973A (en) * | 2018-10-24 | 2019-03-08 | 中车株洲电力机车研究所有限公司 | A kind of vehicle positioning method based on deep neural network image recognition |
Family Cites Families (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9940553B2 (en) * | 2013-02-22 | 2018-04-10 | Microsoft Technology Licensing, Llc | Camera/object pose from predicted coordinates |
CN105718860B (en) * | 2016-01-15 | 2019-09-10 | 武汉光庭科技有限公司 | Localization method and system based on driving safety map and binocular Traffic Sign Recognition |
US9773196B2 (en) * | 2016-01-25 | 2017-09-26 | Adobe Systems Incorporated | Utilizing deep learning for automatic digital image segmentation and stylization |
CN106326858A (en) * | 2016-08-23 | 2017-01-11 | 北京航空航天大学 | Road traffic sign automatic identification and management system based on deep learning |
CN106403926B (en) * | 2016-08-30 | 2020-09-11 | 上海擎朗智能科技有限公司 | Positioning method and system |
CN106845547B (en) * | 2017-01-23 | 2018-08-14 | 重庆邮电大学 | A kind of intelligent automobile positioning and road markings identifying system and method based on camera |
CN107563419B (en) * | 2017-08-22 | 2020-09-04 | 交控科技股份有限公司 | Train positioning method combining image matching and two-dimensional code |
CN107703936A (en) * | 2017-09-22 | 2018-02-16 | 南京轻力舟智能科技有限公司 | Automatic Guided Vehicle system and dolly localization method based on convolutional neural networks |
-
2018
- 2018-10-24 CN CN201811245274.1A patent/CN109446973B/en active Active
-
2019
- 2019-10-18 SG SG11202103814PA patent/SG11202103814PA/en unknown
- 2019-10-18 WO PCT/CN2019/111840 patent/WO2020083103A1/en active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN202350794U (en) * | 2011-11-29 | 2012-07-25 | 高德软件有限公司 | Navigation data acquisition device |
CN103925927A (en) * | 2014-04-18 | 2014-07-16 | 中国科学院软件研究所 | Traffic sign positioning method based on vehicle-mounted video |
US20180211120A1 (en) * | 2017-01-25 | 2018-07-26 | Ford Global Technologies, Llc | Training An Automatic Traffic Light Detection Model Using Simulated Images |
CN108009518A (en) * | 2017-12-19 | 2018-05-08 | 大连理工大学 | A kind of stratification traffic mark recognition methods based on quick two points of convolutional neural networks |
CN109446973A (en) * | 2018-10-24 | 2019-03-08 | 中车株洲电力机车研究所有限公司 | A kind of vehicle positioning method based on deep neural network image recognition |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111914691A (en) * | 2020-07-15 | 2020-11-10 | 北京埃福瑞科技有限公司 | Rail transit vehicle positioning method and system |
CN111914691B (en) * | 2020-07-15 | 2024-03-19 | 北京埃福瑞科技有限公司 | Rail transit vehicle positioning method and system |
CN113378735A (en) * | 2021-06-18 | 2021-09-10 | 北京东土科技股份有限公司 | Road marking line identification method and device, electronic equipment and storage medium |
CN113378735B (en) * | 2021-06-18 | 2023-04-07 | 北京东土科技股份有限公司 | Road marking line identification method and device, electronic equipment and storage medium |
Also Published As
Publication number | Publication date |
---|---|
SG11202103814PA (en) | 2021-05-28 |
CN109446973A (en) | 2019-03-08 |
CN109446973B (en) | 2021-01-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020083103A1 (en) | Vehicle positioning method based on deep neural network image recognition | |
CN106651953B (en) | A kind of vehicle position and orientation estimation method based on traffic sign | |
CN108802785B (en) | Vehicle self-positioning method based on high-precision vector map and monocular vision sensor | |
CN106441319B (en) | A kind of generation system and method for automatic driving vehicle lane grade navigation map | |
CN108256413B (en) | Passable area detection method and device, storage medium and electronic equipment | |
CN106525057A (en) | Generation system for high-precision road map | |
US11625851B2 (en) | Geographic object detection apparatus and geographic object detection method | |
CN109815300B (en) | Vehicle positioning method | |
JP2021508815A (en) | Systems and methods for correcting high-definition maps based on the detection of obstructing objects | |
CN110146910A (en) | A kind of localization method and device merged based on GPS with laser radar data | |
CN109212545A (en) | Multiple source target following measuring system and tracking based on active vision | |
CN109583409A (en) | A kind of intelligent vehicle localization method and system towards cognitive map | |
CN109767637A (en) | The method and apparatus of the identification of countdown signal lamp and processing | |
CN112740225B (en) | Method and device for determining road surface elements | |
WO2022041706A1 (en) | Positioning method, positioning system, and vehicle | |
CN109515439A (en) | Automatic Pilot control method, device, system and storage medium | |
CN113673386B (en) | Marking method for traffic signal lamp in prior inspection map | |
CN112446915B (en) | Picture construction method and device based on image group | |
CN112444251B (en) | Vehicle driving position determining method and device, storage medium and computer equipment | |
CN112424568B (en) | System and method for constructing high-definition map | |
CN110135387B (en) | Image rapid identification method based on sensor fusion | |
CN115127547B (en) | Tunnel detection vehicle positioning method based on strapdown inertial navigation system and image positioning | |
CN112446234B (en) | Position determining method and device based on data association | |
US20240353842A1 (en) | Position determination via encoded indicators in a physical environment | |
CN116630559A (en) | Construction method of lightweight road semantic map |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19876299 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19876299 Country of ref document: EP Kind code of ref document: A1 |