CN110069993A - A kind of target vehicle detection method based on deep learning - Google Patents

A kind of target vehicle detection method based on deep learning Download PDF

Info

Publication number
CN110069993A
CN110069993A CN201910206458.5A CN201910206458A CN110069993A CN 110069993 A CN110069993 A CN 110069993A CN 201910206458 A CN201910206458 A CN 201910206458A CN 110069993 A CN110069993 A CN 110069993A
Authority
CN
China
Prior art keywords
target vehicle
convolutional neural
neural networks
vehicle detection
deep learning
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910206458.5A
Other languages
Chinese (zh)
Other versions
CN110069993B (en
Inventor
瞿三清
许仲聪
卢凡
陈广
董金虎
陈凯
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tongji University
Original Assignee
Tongji University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tongji University filed Critical Tongji University
Priority to CN201910206458.5A priority Critical patent/CN110069993B/en
Publication of CN110069993A publication Critical patent/CN110069993A/en
Application granted granted Critical
Publication of CN110069993B publication Critical patent/CN110069993B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition
    • G06V10/12Details of acquisition arrangements; Constructional details thereof
    • G06V10/14Optical characteristics of the device performing the acquisition or on the illumination arrangements
    • G06V10/143Sensing or illuminating at different wavelengths
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • G06V10/242Aligning, centring, orientation detection or correction of the image by image rotation, e.g. by 90 degrees
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • G06V20/58Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
    • G06V20/584Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Image Analysis (AREA)
  • Traffic Control Systems (AREA)

Abstract

The target vehicle detection method based on deep learning that the present invention relates to a kind of, the following steps are included: 1) acquire the tail feature point cloud data of target vehicle by the way that two single line laser radars of parking robot tail portion are arranged in, and pre-processed to obtain binary picture;2) binary picture is labeled, obtains the wherein position where target vehicle tail portion, training dataset is generated with this;3) building is suitable for the depth convolutional neural networks and its loss function of target vehicle detection;4) it is inputted in depth convolutional neural networks after training dataset being carried out augmentation, and update is trained to the parameter in convolutional neural networks according to the difference of output valve and training true value, optimal network parameter is obtained, and is detected according to trained depth convolutional neural networks.Compared with prior art, the present invention has the advantages that robustness is high, it is low etc. not depend on manual feature, testing cost.

Description

A kind of target vehicle detection method based on deep learning
Technical field
The present invention relates to intelligent parking technical fields, more particularly, to a kind of target vehicle detection side based on deep learning Method.
Background technique
In intelligent driving field, the detection for target vehicle is key for ensureing automatic driving vehicle safety traffic One of business.Equally in intelligent parking technical field, the position of target vehicle is detected, is to realize parking robot to target vehicle essence The committed step of quasi- contraposition.Since laser radar is affected by environment small, target vehicle accurately point cloud data, laser can be acquired Radar has become the sensor of vehicle detection and positioning mostly important in intelligent parking field.
Currently, in intelligent parking technical field, it is also main using based on target vehicle to the detection method of target vehicle The traditional detection algorithm of manual feature.Traditional detection algorithm, although calculation amount is small, fast speed, under many scenes The feature of target vehicle can not match well with manual feature, cause that conventional target vehicle detecting algorithm recall rate is low, Shandong Stick is poor.
Summary of the invention
It is an object of the present invention to overcome the above-mentioned drawbacks of the prior art and provide one kind to be based on deep learning Target vehicle detection method.
The purpose of the present invention can be achieved through the following technical solutions:
A kind of target vehicle detection method based on deep learning, to realize intelligent parking, comprising the following steps:
1) the tail feature point of target vehicle is acquired by the way that two single line laser radars of parking robot tail portion are arranged in Cloud data, and pre-processed to obtain binary picture;
2) binary picture is labeled, obtains the wherein position where target vehicle tail portion, trained number is generated with this According to collection;
3) building is suitable for the depth convolutional neural networks and its loss function of target vehicle detection;
4) it is inputted in depth convolutional neural networks after training dataset being carried out augmentation, and according to output valve and training true value Difference update is trained to the parameter in convolutional neural networks, obtain optimal network parameter, and according to trained depth Degree convolutional neural networks are detected.
The step 1) specifically includes the following steps:
11) collected point cloud data is converted to using single line laser radar as the polar coordinate system of coordinate origin global unified Cartesian coordinate system;
12) to the meshing point cloud data after coordinate conversion, binary picture is converted to.
In the step 11), transformed representation are as follows:
(xj0,yj0)=(xj1,yj1)R+t
Wherein, (rjj) be original point cloud data midpoint j position coordinates, (xj1,yj1) it is that point j is converted to laser thunder Up to the position coordinates in the cartesian coordinate system for coordinate origin, (xj0,yj0) it is the global seat unified in cartesian coordinate system Mark, R are conversion spin matrix, and t is translation vector.
In the step 2), marked content includes that the Pixel-level mark of image and the bounding box of target vehicle mark.
The depth convolutional neural networks are Faster R-CNN convolutional neural networks, with the binary system being sized Image is as input, to export with position and the confidence level of target vehicle is corresponded on input binary picture.
The expression formula of the loss function of depth convolutional neural networks are as follows:
Loss=Lcls(p, u)+λ [u=1] Lloc(tu,v)
Lcls(p, u)=- log (p)
X=(tu-v)
Wherein, Lcls(p, u) is target classification Detectability loss subfunction, Lloc(tu, v) and it is range loss subfunction, p is pair In the predictive factor of target category, u is the practical factor of corresponding classification, and λ is the weighting weight of loss function, is indicated as u=1 It is target vehicle in area-of-interest, indicates that area-of-interest is background, t as u=0uFor the location factor of prediction, v is instruction Practice true location factor in sample, x is the deviation of predicted value and true value.
Augmentation is carried out to training dataset specifically:
Image is carried out Random Level overturning, cuts and uniformly zooms to fixed size, and labeled data also carries out Corresponding overturning cuts and scales.
Training depth convolutional neural networks specifically:
According to loss function, declines back-propagation method using gradient, change to the parameter of depth convolutional neural networks In generation, updates, and the network parameter obtained after iteration to maximum is set number completes training as optimal network parameter.
Compared with prior art, the invention has the following advantages that
One, robustness is high: since laser radar can acquire accurate target vehicle under the operating condition of various complexity Point cloud data is aided with the target vehicle detection algorithm of robustness with higher, and therefore, vehicle checking method of the invention has Very high robustness can also guarantee the relative precision of testing result under complex working condition.
Two, do not depend on manual feature: the present invention passes through depth mind compared to traditional target vehicle detection algorithm, this method Target signature is learnt through network, does not depend on manual feature, the recall rate of testing result is high.
Three, testing cost is low: the present invention uses two single line laser radars as sensor, compared to multi-thread laser thunder It reaches, testing cost is low.
Detailed description of the invention
Fig. 1 is the flow chart of detection method of the invention.
Fig. 2 is intelligent parking robot architecture's schematic diagram in the embodiment of the present invention.
Fig. 3 is target vehicle depth convolutional network structural schematic diagram in the embodiment of the present invention.
Specific embodiment
The present invention is described in detail with specific embodiment below in conjunction with the accompanying drawings.
The present invention provides a kind of based on deep learning using the method for the target vehicle detection of single line laser radar, leads to The sensor that single line laser radar is crossed as detection obtains the point cloud data of target vehicle, and depth volume is input to after being pre-processed Product neural network, finally obtains position and the confidence level of target vehicle.As shown in Figure 1, this method comprises the following steps:
(1) using the tail feature point cloud data of 2 single line laser radar acquisition target vehicles, and to collected cloud Data are pre-processed;
(2) acquisition data are passed through with the tail portion position for manually marking wherein target vehicle, constructs training data used Collection
(3) building is suitable for the depth convolutional neural networks and loss function of target vehicle detection
(4) data used in the training of step (2) are subjected to augmentation, are input to depth convolutional Neural constructed by step (3) In network.Update is trained to the parameter in convolutional neural networks according to the difference of output valve and training true value, is finally obtained Ideal network parameter.
It in the present embodiment, include to the coordinate conversion of point cloud data and figure to the pretreatment of point cloud data in step (1) As two following steps such as conversions:
2 single line laser radars in (1-1) the present embodiment are located at the rear two sides of intelligent parking robot, intelligent parking Robot architecture is as shown in Fig. 2, two single line laser radars are by the point cloud data around the acquisition of definitely frame per second, and collection result is to swash Optical radar is coordinate origin, and is stored with polar form.The point cloud data of acquisition is converted to the overall situation system of parking robot In one cartesian coordinate system.
Transformed representation is as follows:
(xj0,yj0)=(xj1,yj1)R+t
In above formula, (rjj) indicate acquisition original point cloud data in certain point, (xj1,yj1) it is the point cloud number acquired Certain point in is converted to using laser radar to indicate in the cartesian coordinate system of coordinate origin.(xj0,yj0) it is then correspondence Point cloud data point converts the expression into the unified cartesian coordinate system of the overall situation of parking robot.R is conversion spin matrix, and t is Translation vector.
(1-2), to point cloud data imageization processing.In the present embodiment, what laser radar acquired is set as apart from the upper limit Then 10m, the meshing point cloud data after coordinate is converted adjust the binary picture having a size of 250 × 250, if corresponding There is data point in grid, is then set to 1 and is otherwise set to 0.
In the present embodiment, step (2) is to construct the required data set of deep learning training.In the laser radar to acquisition It after point cloud data is handled, needs manually to mark training data, to form the required data set of training.The mode of mark The including but not limited to bounding box mark of the Pixel-level mark of image, target vehicle.Target vehicle need to be included at least when mark Position, but the posture information etc. for increasing target vehicle can be expanded.
In the present embodiment, step (3) is depth convolutional neural networks and the loss that building is suitable for target vehicle detection Function.The training dataset prepared in the building of depth convolutional neural networks and step (2) is directly related, in the present embodiment, Step (2) using target vehicle bounding box mark, therefore in the present embodiment the structure of depth convolutional neural networks with Faster R-CNN is similar, and main structure is built with reference to Faster R-CNN, and convolutional neural networks structure is as shown in Figure 3.
In the present embodiment, in step (3), the loss function of depth convolutional neural networks is constituted with two parts weighting:
Loss=Lcls(p, u)+λ [u=1] Lloc(tu,v)
(3-1) constructs target classification loss function Lcls(p, u), wherein p is the predictive factor for target category, and u table Show the practical factor of corresponding classification.The building of log loss function is generallyd use, wherein p indicates that the prediction probability of a certain classification, p are got over Close to 1, confidence level is higher, loses smaller.
Lcls(p, u)=- log (p)
(3-2) constructs target detection range Lloc(tu,v).Wherein λ indicates the weighting weight of loss function, can usually take λ=1.[u=1] indicates in area-of-interest to be that target vehicle takes 1, takes 0 when area-of-interest is background, i.e., if current When area-of-interest is environment unrelated things, its range loss is not considered.T in formulauIndicate to represent predicted position because Son, and v indicates true location factor in training sample.Usually the range loss subfunction is public with smooth manhatton distance FormulaBuilding, expression formula are as follows: wherein x=(tu- v), indicate the deviation of predicted value and true value.
In the present embodiment, step (4) mainly includes turning over image into row stochastic level to the augmentation of training data Turn, cut and uniformly zoom to fixed size, labeled data is also overturn accordingly, cut and scaled, on this basis Obtained image is normalized by channel, the fixed dimension used in the present embodiment is 250 × 250.
In the present embodiment, when step (4) is to the initialization for training network model, first in ImageNet or other figures Pre-training is carried out as extracting network to object features using SoftMax loss function on categorized data set, obtained parameter value is made For the initial parameter of network.
In the present embodiment, when in step (4) to the training of network, comprehensive loss is calculated using the loss function of weighting Then value carries out backpropagation and calculates gradient, and updates network parameter using optimizers such as Adam, the certain number of iteration obtains most Whole result.And set final parametric results to the network model parameter of target vehicle detection device, for detecting target carriage Use.
The present invention provides a kind of based on deep learning using the method for the target vehicle detection of single line laser radar, leads to The sensor that single line laser radar is crossed as detection obtains the point cloud data of target vehicle, and depth volume is input to after being pre-processed Product neural network, finally obtains position and the confidence level of target vehicle.The detection performance is outstanding, while having higher robustness, Cost of implementation is low, is easy to be deployed to the detection that target vehicle is used in existing intelligent parking robot.
Person skilled in the art obviously easily can make various modifications to these embodiments, and saying herein Bright General Principle is applied in other embodiments without having to go through creative labor.Therefore, the present invention is not limited to here Embodiment, those skilled in the art's announcement according to the present invention, improvement and modification made without departing from the scope of the present invention are all answered This is within protection scope of the present invention.

Claims (8)

1. a kind of target vehicle detection method based on deep learning, to realize intelligent parking, which is characterized in that including following Step:
1) the tail feature point cloud number of target vehicle is acquired by the way that two single line laser radars of parking robot tail portion are arranged in According to, and pre-processed to obtain binary picture;
2) binary picture is labeled, obtains the wherein position where target vehicle tail portion, training dataset is generated with this;
3) building is suitable for the depth convolutional neural networks and its loss function of target vehicle detection;
4) it is inputted in depth convolutional neural networks after training dataset being carried out augmentation, and according to the difference of output valve and training true value The different parameter in convolutional neural networks is trained update, obtains optimal network parameter, and roll up according to trained depth Product neural network is detected.
2. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described Step 1) specifically includes the following steps:
11) collected point cloud data is converted into global unified flute by the polar coordinate system of coordinate origin of single line laser radar Karr coordinate system;
12) to the meshing point cloud data after coordinate conversion, binary picture is converted to.
3. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described In step 11), transformed representation are as follows:
(xj0,yj0)=(xj1,yj1)R+t
Wherein, (rjj) be original point cloud data midpoint j position coordinates, (xj1,yj1) it is that point j is converted to and is with laser radar Position coordinates in the cartesian coordinate system of coordinate origin, (xj0,yj0) it is the global coordinate unified in cartesian coordinate system, R is Spin matrix is converted, t is translation vector.
4. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described In step 2), marked content includes that the Pixel-level mark of image and the bounding box of target vehicle mark.
5. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described Depth convolutional neural networks are Faster R-CNN convolutional neural networks, using the binary picture that is sized as inputting, To be exported with position and the confidence level of target vehicle is corresponded on input binary picture.
6. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described In step 3), the expression formula of the loss function of depth convolutional neural networks are as follows:
Loss=Lcls(p, u)+λ [u=1] Lloc(tu,v)
Lcls(p, u)=- log (p)
X=(tu-v)
Wherein, Lcls(p, u) is target classification Detectability loss subfunction, Lloc(tu, v) and it is range loss subfunction, p is for mesh The predictive factor of classification is marked, u is the practical factor of corresponding classification, and λ is the weighting weight of loss function, indicates feeling as u=1 Interest region is target vehicle, indicates that area-of-interest is background, t as u=0uFor the location factor of prediction, v is training sample True location factor in this, x are the deviation of predicted value and true value.
7. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described In step 4), augmentation is carried out to training dataset specifically:
Image is carried out Random Level overturning, cuts and uniformly zooms to fixed size, and labeled data also carries out accordingly Overturning, cutting and scaling.
8. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described In step 4), training depth convolutional neural networks specifically:
According to loss function, declines back-propagation method using gradient, the parameter of depth convolutional neural networks is iterated more Newly, the network parameter obtained after iteration to maximum being set number completes training as optimal network parameter.
CN201910206458.5A 2019-03-19 2019-03-19 Target vehicle detection method based on deep learning Active CN110069993B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910206458.5A CN110069993B (en) 2019-03-19 2019-03-19 Target vehicle detection method based on deep learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910206458.5A CN110069993B (en) 2019-03-19 2019-03-19 Target vehicle detection method based on deep learning

Publications (2)

Publication Number Publication Date
CN110069993A true CN110069993A (en) 2019-07-30
CN110069993B CN110069993B (en) 2021-10-08

Family

ID=67366360

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910206458.5A Active CN110069993B (en) 2019-03-19 2019-03-19 Target vehicle detection method based on deep learning

Country Status (1)

Country Link
CN (1) CN110069993B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111178213A (en) * 2019-12-23 2020-05-19 大连理工大学 Aerial photography vehicle detection method based on deep learning
CN111177297A (en) * 2019-12-31 2020-05-19 信阳师范学院 Dynamic target speed calculation optimization method based on video and GIS
CN111523403A (en) * 2020-04-03 2020-08-11 咪咕文化科技有限公司 Method and device for acquiring target area in picture and computer readable storage medium
CN111539347A (en) * 2020-04-27 2020-08-14 北京百度网讯科技有限公司 Method and apparatus for detecting target
CN111653103A (en) * 2020-05-07 2020-09-11 浙江大华技术股份有限公司 Target object identification method and device
CN111783844A (en) * 2020-06-10 2020-10-16 东莞正扬电子机械有限公司 Target detection model training method and device based on deep learning and storage medium
CN112560834A (en) * 2019-09-26 2021-03-26 武汉金山办公软件有限公司 Coordinate prediction model generation method and device and graph recognition method and device
CN113313201A (en) * 2021-06-21 2021-08-27 南京挥戈智能科技有限公司 Multi-target detection and distance measurement method based on Swin transducer and ZED camera
US20220204018A1 (en) * 2020-12-29 2022-06-30 GM Global Technology Operations LLC Vehicle control using neural network controller in combination with model-based controller
CN114692720A (en) * 2022-02-25 2022-07-01 广州文远知行科技有限公司 Image classification method, device, equipment and storage medium based on aerial view

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050240323A1 (en) * 2004-03-31 2005-10-27 Honda Motor Co., Ltd. Parking lot attendant robot system
CN106355194A (en) * 2016-08-22 2017-01-25 广东华中科技大学工业技术研究院 Treatment method for surface target of unmanned ship based on laser imaging radar
US20170039436A1 (en) * 2015-08-03 2017-02-09 Nokia Technologies Oy Fusion of RGB Images and Lidar Data for Lane Classification
CN106650809A (en) * 2016-12-20 2017-05-10 福州大学 Method and system for classifying vehicle-borne laser-point cloud targets
CN107239794A (en) * 2017-05-18 2017-10-10 深圳市速腾聚创科技有限公司 Point cloud data segmentation method and terminal
CN108009509A (en) * 2017-12-12 2018-05-08 河南工业大学 Vehicle target detection method
CN108830188A (en) * 2018-05-30 2018-11-16 西安理工大学 Vehicle checking method based on deep learning
CN109063753A (en) * 2018-07-18 2018-12-21 北方民族大学 A kind of three-dimensional point cloud model classification method based on convolutional neural networks
CN109118500A (en) * 2018-07-16 2019-01-01 重庆大学产业技术研究院 A kind of dividing method of the Point Cloud Data from Three Dimension Laser Scanning based on image
US20190004508A1 (en) * 2017-07-03 2019-01-03 Volvo Car Corporation Method and system for automatic parking of a vehicle
CN109270543A (en) * 2018-09-20 2019-01-25 同济大学 A kind of system and method for pair of target vehicle surrounding vehicles location information detection
CN109270544A (en) * 2018-09-20 2019-01-25 同济大学 Mobile robot self-localization system based on shaft identification
CN109324616A (en) * 2018-09-20 2019-02-12 同济大学 Nobody based on onboard sensor parks the alignment method of transfer robot
CN109344786A (en) * 2018-10-11 2019-02-15 深圳步智造科技有限公司 Target identification method, device and computer readable storage medium
CN109386155A (en) * 2018-09-20 2019-02-26 同济大学 Nobody towards automated parking ground parks the alignment method of transfer robot

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20050240323A1 (en) * 2004-03-31 2005-10-27 Honda Motor Co., Ltd. Parking lot attendant robot system
US20170039436A1 (en) * 2015-08-03 2017-02-09 Nokia Technologies Oy Fusion of RGB Images and Lidar Data for Lane Classification
CN106355194A (en) * 2016-08-22 2017-01-25 广东华中科技大学工业技术研究院 Treatment method for surface target of unmanned ship based on laser imaging radar
CN106650809A (en) * 2016-12-20 2017-05-10 福州大学 Method and system for classifying vehicle-borne laser-point cloud targets
CN107239794A (en) * 2017-05-18 2017-10-10 深圳市速腾聚创科技有限公司 Point cloud data segmentation method and terminal
US20190004508A1 (en) * 2017-07-03 2019-01-03 Volvo Car Corporation Method and system for automatic parking of a vehicle
CN108009509A (en) * 2017-12-12 2018-05-08 河南工业大学 Vehicle target detection method
CN108830188A (en) * 2018-05-30 2018-11-16 西安理工大学 Vehicle checking method based on deep learning
CN109118500A (en) * 2018-07-16 2019-01-01 重庆大学产业技术研究院 A kind of dividing method of the Point Cloud Data from Three Dimension Laser Scanning based on image
CN109063753A (en) * 2018-07-18 2018-12-21 北方民族大学 A kind of three-dimensional point cloud model classification method based on convolutional neural networks
CN109270543A (en) * 2018-09-20 2019-01-25 同济大学 A kind of system and method for pair of target vehicle surrounding vehicles location information detection
CN109270544A (en) * 2018-09-20 2019-01-25 同济大学 Mobile robot self-localization system based on shaft identification
CN109324616A (en) * 2018-09-20 2019-02-12 同济大学 Nobody based on onboard sensor parks the alignment method of transfer robot
CN109386155A (en) * 2018-09-20 2019-02-26 同济大学 Nobody towards automated parking ground parks the alignment method of transfer robot
CN109344786A (en) * 2018-10-11 2019-02-15 深圳步智造科技有限公司 Target identification method, device and computer readable storage medium

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
AMIR KOTB 等: "Smart Parking Guidance, Monitoring and Reservations: A Review", 《 IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE》 *
企鹅号 - AI科技大本营: "无人驾驶汽车系统入门:基于深度学习的实时激光雷达点云目标检测及ROS实现", 《HTTPS://CLOUD.TENCENT.COM/DEVELOPER/NEWS/339676》 *
李游: "基于车载激光扫描数据的城市街道信息提取技术研究", 《中国博士学位论文全文数据库 基础科学辑》 *
罗海峰等: "基于DBN的车载激光点云路侧多目标提取", 《测绘学报》 *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112560834B (en) * 2019-09-26 2024-05-10 武汉金山办公软件有限公司 Coordinate prediction model generation method and device and pattern recognition method and device
CN112560834A (en) * 2019-09-26 2021-03-26 武汉金山办公软件有限公司 Coordinate prediction model generation method and device and graph recognition method and device
CN111178213A (en) * 2019-12-23 2020-05-19 大连理工大学 Aerial photography vehicle detection method based on deep learning
CN111178213B (en) * 2019-12-23 2022-11-18 大连理工大学 Aerial photography vehicle detection method based on deep learning
CN111177297B (en) * 2019-12-31 2022-09-02 信阳师范学院 Dynamic target speed calculation optimization method based on video and GIS
CN111177297A (en) * 2019-12-31 2020-05-19 信阳师范学院 Dynamic target speed calculation optimization method based on video and GIS
CN111523403A (en) * 2020-04-03 2020-08-11 咪咕文化科技有限公司 Method and device for acquiring target area in picture and computer readable storage medium
CN111523403B (en) * 2020-04-03 2023-10-20 咪咕文化科技有限公司 Method and device for acquiring target area in picture and computer readable storage medium
CN111539347A (en) * 2020-04-27 2020-08-14 北京百度网讯科技有限公司 Method and apparatus for detecting target
CN111539347B (en) * 2020-04-27 2023-08-08 北京百度网讯科技有限公司 Method and device for detecting target
CN111653103A (en) * 2020-05-07 2020-09-11 浙江大华技术股份有限公司 Target object identification method and device
CN111783844A (en) * 2020-06-10 2020-10-16 东莞正扬电子机械有限公司 Target detection model training method and device based on deep learning and storage medium
CN111783844B (en) * 2020-06-10 2024-05-28 广东正扬传感科技股份有限公司 Deep learning-based target detection model training method, device and storage medium
US20220204018A1 (en) * 2020-12-29 2022-06-30 GM Global Technology Operations LLC Vehicle control using neural network controller in combination with model-based controller
US11834066B2 (en) * 2020-12-29 2023-12-05 GM Global Technology Operations LLC Vehicle control using neural network controller in combination with model-based controller
CN113313201A (en) * 2021-06-21 2021-08-27 南京挥戈智能科技有限公司 Multi-target detection and distance measurement method based on Swin transducer and ZED camera
CN114692720A (en) * 2022-02-25 2022-07-01 广州文远知行科技有限公司 Image classification method, device, equipment and storage medium based on aerial view

Also Published As

Publication number Publication date
CN110069993B (en) 2021-10-08

Similar Documents

Publication Publication Date Title
CN110069993A (en) A kind of target vehicle detection method based on deep learning
CN110163187B (en) F-RCNN-based remote traffic sign detection and identification method
CN109977918B (en) Target detection positioning optimization method based on unsupervised domain adaptation
CN112766087A (en) Optical remote sensing image ship detection method based on knowledge distillation
CN108629288B (en) Gesture recognition model training method, gesture recognition method and system
CN111444760A (en) Traffic sign detection and identification method based on pruning and knowledge distillation
CN111242037A (en) Lane line detection method based on structural information
CN110097047A (en) A kind of vehicle checking method using single line laser radar based on deep learning
CN108550259A (en) Congestion in road judgment method, terminal device and computer readable storage medium
CN111950453A (en) Optional-shape text recognition method based on selective attention mechanism
CN107818302A (en) Non-rigid multiple dimensioned object detecting method based on convolutional neural networks
CN113486764B (en) Pothole detection method based on improved YOLOv3
CN112364931B (en) Few-sample target detection method and network system based on meta-feature and weight adjustment
CN111898432B (en) Pedestrian detection system and method based on improved YOLOv3 algorithm
CN111368671A (en) SAR image ship target detection and identification integrated method based on deep learning
CN108921030A (en) A kind of SAR automatic target recognition method of Fast Learning
CN112418330A (en) Improved SSD (solid State drive) -based high-precision detection method for small target object
CN111160407A (en) Deep learning target detection method and system
CN112950780B (en) Intelligent network map generation method and system based on remote sensing image
CN113177560A (en) Universal lightweight deep learning vehicle detection method
CN113487600B (en) Feature enhancement scale self-adaptive perception ship detection method
CN111882620A (en) Road drivable area segmentation method based on multi-scale information
CN115965862A (en) SAR ship target detection method based on mask network fusion image characteristics
CN113205103A (en) Lightweight tattoo detection method
Wang et al. Accurate real-time ship target detection using Yolov4

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant