CN110069993A - A kind of target vehicle detection method based on deep learning - Google Patents
A kind of target vehicle detection method based on deep learning Download PDFInfo
- Publication number
- CN110069993A CN110069993A CN201910206458.5A CN201910206458A CN110069993A CN 110069993 A CN110069993 A CN 110069993A CN 201910206458 A CN201910206458 A CN 201910206458A CN 110069993 A CN110069993 A CN 110069993A
- Authority
- CN
- China
- Prior art keywords
- target vehicle
- convolutional neural
- neural networks
- vehicle detection
- deep learning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
- G06V10/12—Details of acquisition arrangements; Constructional details thereof
- G06V10/14—Optical characteristics of the device performing the acquisition or on the illumination arrangements
- G06V10/143—Sensing or illuminating at different wavelengths
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/24—Aligning, centring, orientation detection or correction of the image
- G06V10/242—Aligning, centring, orientation detection or correction of the image by image rotation, e.g. by 90 degrees
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/25—Determination of region of interest [ROI] or a volume of interest [VOI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/58—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads
- G06V20/584—Recognition of moving objects or obstacles, e.g. vehicles or pedestrians; Recognition of traffic objects, e.g. traffic signs, traffic lights or roads of vehicle lights or traffic lights
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Life Sciences & Earth Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- General Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Molecular Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Evolutionary Biology (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
- Traffic Control Systems (AREA)
Abstract
The target vehicle detection method based on deep learning that the present invention relates to a kind of, the following steps are included: 1) acquire the tail feature point cloud data of target vehicle by the way that two single line laser radars of parking robot tail portion are arranged in, and pre-processed to obtain binary picture;2) binary picture is labeled, obtains the wherein position where target vehicle tail portion, training dataset is generated with this;3) building is suitable for the depth convolutional neural networks and its loss function of target vehicle detection;4) it is inputted in depth convolutional neural networks after training dataset being carried out augmentation, and update is trained to the parameter in convolutional neural networks according to the difference of output valve and training true value, optimal network parameter is obtained, and is detected according to trained depth convolutional neural networks.Compared with prior art, the present invention has the advantages that robustness is high, it is low etc. not depend on manual feature, testing cost.
Description
Technical field
The present invention relates to intelligent parking technical fields, more particularly, to a kind of target vehicle detection side based on deep learning
Method.
Background technique
In intelligent driving field, the detection for target vehicle is key for ensureing automatic driving vehicle safety traffic
One of business.Equally in intelligent parking technical field, the position of target vehicle is detected, is to realize parking robot to target vehicle essence
The committed step of quasi- contraposition.Since laser radar is affected by environment small, target vehicle accurately point cloud data, laser can be acquired
Radar has become the sensor of vehicle detection and positioning mostly important in intelligent parking field.
Currently, in intelligent parking technical field, it is also main using based on target vehicle to the detection method of target vehicle
The traditional detection algorithm of manual feature.Traditional detection algorithm, although calculation amount is small, fast speed, under many scenes
The feature of target vehicle can not match well with manual feature, cause that conventional target vehicle detecting algorithm recall rate is low, Shandong
Stick is poor.
Summary of the invention
It is an object of the present invention to overcome the above-mentioned drawbacks of the prior art and provide one kind to be based on deep learning
Target vehicle detection method.
The purpose of the present invention can be achieved through the following technical solutions:
A kind of target vehicle detection method based on deep learning, to realize intelligent parking, comprising the following steps:
1) the tail feature point of target vehicle is acquired by the way that two single line laser radars of parking robot tail portion are arranged in
Cloud data, and pre-processed to obtain binary picture;
2) binary picture is labeled, obtains the wherein position where target vehicle tail portion, trained number is generated with this
According to collection;
3) building is suitable for the depth convolutional neural networks and its loss function of target vehicle detection;
4) it is inputted in depth convolutional neural networks after training dataset being carried out augmentation, and according to output valve and training true value
Difference update is trained to the parameter in convolutional neural networks, obtain optimal network parameter, and according to trained depth
Degree convolutional neural networks are detected.
The step 1) specifically includes the following steps:
11) collected point cloud data is converted to using single line laser radar as the polar coordinate system of coordinate origin global unified
Cartesian coordinate system;
12) to the meshing point cloud data after coordinate conversion, binary picture is converted to.
In the step 11), transformed representation are as follows:
(xj0,yj0)=(xj1,yj1)R+t
Wherein, (rj,φj) be original point cloud data midpoint j position coordinates, (xj1,yj1) it is that point j is converted to laser thunder
Up to the position coordinates in the cartesian coordinate system for coordinate origin, (xj0,yj0) it is the global seat unified in cartesian coordinate system
Mark, R are conversion spin matrix, and t is translation vector.
In the step 2), marked content includes that the Pixel-level mark of image and the bounding box of target vehicle mark.
The depth convolutional neural networks are Faster R-CNN convolutional neural networks, with the binary system being sized
Image is as input, to export with position and the confidence level of target vehicle is corresponded on input binary picture.
The expression formula of the loss function of depth convolutional neural networks are as follows:
Loss=Lcls(p, u)+λ [u=1] Lloc(tu,v)
Lcls(p, u)=- log (p)
X=(tu-v)
Wherein, Lcls(p, u) is target classification Detectability loss subfunction, Lloc(tu, v) and it is range loss subfunction, p is pair
In the predictive factor of target category, u is the practical factor of corresponding classification, and λ is the weighting weight of loss function, is indicated as u=1
It is target vehicle in area-of-interest, indicates that area-of-interest is background, t as u=0uFor the location factor of prediction, v is instruction
Practice true location factor in sample, x is the deviation of predicted value and true value.
Augmentation is carried out to training dataset specifically:
Image is carried out Random Level overturning, cuts and uniformly zooms to fixed size, and labeled data also carries out
Corresponding overturning cuts and scales.
Training depth convolutional neural networks specifically:
According to loss function, declines back-propagation method using gradient, change to the parameter of depth convolutional neural networks
In generation, updates, and the network parameter obtained after iteration to maximum is set number completes training as optimal network parameter.
Compared with prior art, the invention has the following advantages that
One, robustness is high: since laser radar can acquire accurate target vehicle under the operating condition of various complexity
Point cloud data is aided with the target vehicle detection algorithm of robustness with higher, and therefore, vehicle checking method of the invention has
Very high robustness can also guarantee the relative precision of testing result under complex working condition.
Two, do not depend on manual feature: the present invention passes through depth mind compared to traditional target vehicle detection algorithm, this method
Target signature is learnt through network, does not depend on manual feature, the recall rate of testing result is high.
Three, testing cost is low: the present invention uses two single line laser radars as sensor, compared to multi-thread laser thunder
It reaches, testing cost is low.
Detailed description of the invention
Fig. 1 is the flow chart of detection method of the invention.
Fig. 2 is intelligent parking robot architecture's schematic diagram in the embodiment of the present invention.
Fig. 3 is target vehicle depth convolutional network structural schematic diagram in the embodiment of the present invention.
Specific embodiment
The present invention is described in detail with specific embodiment below in conjunction with the accompanying drawings.
The present invention provides a kind of based on deep learning using the method for the target vehicle detection of single line laser radar, leads to
The sensor that single line laser radar is crossed as detection obtains the point cloud data of target vehicle, and depth volume is input to after being pre-processed
Product neural network, finally obtains position and the confidence level of target vehicle.As shown in Figure 1, this method comprises the following steps:
(1) using the tail feature point cloud data of 2 single line laser radar acquisition target vehicles, and to collected cloud
Data are pre-processed;
(2) acquisition data are passed through with the tail portion position for manually marking wherein target vehicle, constructs training data used
Collection
(3) building is suitable for the depth convolutional neural networks and loss function of target vehicle detection
(4) data used in the training of step (2) are subjected to augmentation, are input to depth convolutional Neural constructed by step (3)
In network.Update is trained to the parameter in convolutional neural networks according to the difference of output valve and training true value, is finally obtained
Ideal network parameter.
It in the present embodiment, include to the coordinate conversion of point cloud data and figure to the pretreatment of point cloud data in step (1)
As two following steps such as conversions:
2 single line laser radars in (1-1) the present embodiment are located at the rear two sides of intelligent parking robot, intelligent parking
Robot architecture is as shown in Fig. 2, two single line laser radars are by the point cloud data around the acquisition of definitely frame per second, and collection result is to swash
Optical radar is coordinate origin, and is stored with polar form.The point cloud data of acquisition is converted to the overall situation system of parking robot
In one cartesian coordinate system.
Transformed representation is as follows:
(xj0,yj0)=(xj1,yj1)R+t
In above formula, (rj,φj) indicate acquisition original point cloud data in certain point, (xj1,yj1) it is the point cloud number acquired
Certain point in is converted to using laser radar to indicate in the cartesian coordinate system of coordinate origin.(xj0,yj0) it is then correspondence
Point cloud data point converts the expression into the unified cartesian coordinate system of the overall situation of parking robot.R is conversion spin matrix, and t is
Translation vector.
(1-2), to point cloud data imageization processing.In the present embodiment, what laser radar acquired is set as apart from the upper limit
Then 10m, the meshing point cloud data after coordinate is converted adjust the binary picture having a size of 250 × 250, if corresponding
There is data point in grid, is then set to 1 and is otherwise set to 0.
In the present embodiment, step (2) is to construct the required data set of deep learning training.In the laser radar to acquisition
It after point cloud data is handled, needs manually to mark training data, to form the required data set of training.The mode of mark
The including but not limited to bounding box mark of the Pixel-level mark of image, target vehicle.Target vehicle need to be included at least when mark
Position, but the posture information etc. for increasing target vehicle can be expanded.
In the present embodiment, step (3) is depth convolutional neural networks and the loss that building is suitable for target vehicle detection
Function.The training dataset prepared in the building of depth convolutional neural networks and step (2) is directly related, in the present embodiment,
Step (2) using target vehicle bounding box mark, therefore in the present embodiment the structure of depth convolutional neural networks with
Faster R-CNN is similar, and main structure is built with reference to Faster R-CNN, and convolutional neural networks structure is as shown in Figure 3.
In the present embodiment, in step (3), the loss function of depth convolutional neural networks is constituted with two parts weighting:
Loss=Lcls(p, u)+λ [u=1] Lloc(tu,v)
(3-1) constructs target classification loss function Lcls(p, u), wherein p is the predictive factor for target category, and u table
Show the practical factor of corresponding classification.The building of log loss function is generallyd use, wherein p indicates that the prediction probability of a certain classification, p are got over
Close to 1, confidence level is higher, loses smaller.
Lcls(p, u)=- log (p)
(3-2) constructs target detection range Lloc(tu,v).Wherein λ indicates the weighting weight of loss function, can usually take
λ=1.[u=1] indicates in area-of-interest to be that target vehicle takes 1, takes 0 when area-of-interest is background, i.e., if current
When area-of-interest is environment unrelated things, its range loss is not considered.T in formulauIndicate to represent predicted position because
Son, and v indicates true location factor in training sample.Usually the range loss subfunction is public with smooth manhatton distance
FormulaBuilding, expression formula are as follows: wherein x=(tu- v), indicate the deviation of predicted value and true value.
In the present embodiment, step (4) mainly includes turning over image into row stochastic level to the augmentation of training data
Turn, cut and uniformly zoom to fixed size, labeled data is also overturn accordingly, cut and scaled, on this basis
Obtained image is normalized by channel, the fixed dimension used in the present embodiment is 250 × 250.
In the present embodiment, when step (4) is to the initialization for training network model, first in ImageNet or other figures
Pre-training is carried out as extracting network to object features using SoftMax loss function on categorized data set, obtained parameter value is made
For the initial parameter of network.
In the present embodiment, when in step (4) to the training of network, comprehensive loss is calculated using the loss function of weighting
Then value carries out backpropagation and calculates gradient, and updates network parameter using optimizers such as Adam, the certain number of iteration obtains most
Whole result.And set final parametric results to the network model parameter of target vehicle detection device, for detecting target carriage
Use.
The present invention provides a kind of based on deep learning using the method for the target vehicle detection of single line laser radar, leads to
The sensor that single line laser radar is crossed as detection obtains the point cloud data of target vehicle, and depth volume is input to after being pre-processed
Product neural network, finally obtains position and the confidence level of target vehicle.The detection performance is outstanding, while having higher robustness,
Cost of implementation is low, is easy to be deployed to the detection that target vehicle is used in existing intelligent parking robot.
Person skilled in the art obviously easily can make various modifications to these embodiments, and saying herein
Bright General Principle is applied in other embodiments without having to go through creative labor.Therefore, the present invention is not limited to here
Embodiment, those skilled in the art's announcement according to the present invention, improvement and modification made without departing from the scope of the present invention are all answered
This is within protection scope of the present invention.
Claims (8)
1. a kind of target vehicle detection method based on deep learning, to realize intelligent parking, which is characterized in that including following
Step:
1) the tail feature point cloud number of target vehicle is acquired by the way that two single line laser radars of parking robot tail portion are arranged in
According to, and pre-processed to obtain binary picture;
2) binary picture is labeled, obtains the wherein position where target vehicle tail portion, training dataset is generated with this;
3) building is suitable for the depth convolutional neural networks and its loss function of target vehicle detection;
4) it is inputted in depth convolutional neural networks after training dataset being carried out augmentation, and according to the difference of output valve and training true value
The different parameter in convolutional neural networks is trained update, obtains optimal network parameter, and roll up according to trained depth
Product neural network is detected.
2. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described
Step 1) specifically includes the following steps:
11) collected point cloud data is converted into global unified flute by the polar coordinate system of coordinate origin of single line laser radar
Karr coordinate system;
12) to the meshing point cloud data after coordinate conversion, binary picture is converted to.
3. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described
In step 11), transformed representation are as follows:
(xj0,yj0)=(xj1,yj1)R+t
Wherein, (rj,φj) be original point cloud data midpoint j position coordinates, (xj1,yj1) it is that point j is converted to and is with laser radar
Position coordinates in the cartesian coordinate system of coordinate origin, (xj0,yj0) it is the global coordinate unified in cartesian coordinate system, R is
Spin matrix is converted, t is translation vector.
4. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described
In step 2), marked content includes that the Pixel-level mark of image and the bounding box of target vehicle mark.
5. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described
Depth convolutional neural networks are Faster R-CNN convolutional neural networks, using the binary picture that is sized as inputting,
To be exported with position and the confidence level of target vehicle is corresponded on input binary picture.
6. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described
In step 3), the expression formula of the loss function of depth convolutional neural networks are as follows:
Loss=Lcls(p, u)+λ [u=1] Lloc(tu,v)
Lcls(p, u)=- log (p)
X=(tu-v)
Wherein, Lcls(p, u) is target classification Detectability loss subfunction, Lloc(tu, v) and it is range loss subfunction, p is for mesh
The predictive factor of classification is marked, u is the practical factor of corresponding classification, and λ is the weighting weight of loss function, indicates feeling as u=1
Interest region is target vehicle, indicates that area-of-interest is background, t as u=0uFor the location factor of prediction, v is training sample
True location factor in this, x are the deviation of predicted value and true value.
7. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described
In step 4), augmentation is carried out to training dataset specifically:
Image is carried out Random Level overturning, cuts and uniformly zooms to fixed size, and labeled data also carries out accordingly
Overturning, cutting and scaling.
8. a kind of target vehicle detection method based on deep learning according to claim 1, which is characterized in that described
In step 4), training depth convolutional neural networks specifically:
According to loss function, declines back-propagation method using gradient, the parameter of depth convolutional neural networks is iterated more
Newly, the network parameter obtained after iteration to maximum being set number completes training as optimal network parameter.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910206458.5A CN110069993B (en) | 2019-03-19 | 2019-03-19 | Target vehicle detection method based on deep learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910206458.5A CN110069993B (en) | 2019-03-19 | 2019-03-19 | Target vehicle detection method based on deep learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110069993A true CN110069993A (en) | 2019-07-30 |
CN110069993B CN110069993B (en) | 2021-10-08 |
Family
ID=67366360
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910206458.5A Active CN110069993B (en) | 2019-03-19 | 2019-03-19 | Target vehicle detection method based on deep learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110069993B (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111178213A (en) * | 2019-12-23 | 2020-05-19 | 大连理工大学 | Aerial photography vehicle detection method based on deep learning |
CN111177297A (en) * | 2019-12-31 | 2020-05-19 | 信阳师范学院 | Dynamic target speed calculation optimization method based on video and GIS |
CN111523403A (en) * | 2020-04-03 | 2020-08-11 | 咪咕文化科技有限公司 | Method and device for acquiring target area in picture and computer readable storage medium |
CN111539347A (en) * | 2020-04-27 | 2020-08-14 | 北京百度网讯科技有限公司 | Method and apparatus for detecting target |
CN111653103A (en) * | 2020-05-07 | 2020-09-11 | 浙江大华技术股份有限公司 | Target object identification method and device |
CN111783844A (en) * | 2020-06-10 | 2020-10-16 | 东莞正扬电子机械有限公司 | Target detection model training method and device based on deep learning and storage medium |
CN112560834A (en) * | 2019-09-26 | 2021-03-26 | 武汉金山办公软件有限公司 | Coordinate prediction model generation method and device and graph recognition method and device |
CN113313201A (en) * | 2021-06-21 | 2021-08-27 | 南京挥戈智能科技有限公司 | Multi-target detection and distance measurement method based on Swin transducer and ZED camera |
US20220204018A1 (en) * | 2020-12-29 | 2022-06-30 | GM Global Technology Operations LLC | Vehicle control using neural network controller in combination with model-based controller |
CN114692720A (en) * | 2022-02-25 | 2022-07-01 | 广州文远知行科技有限公司 | Image classification method, device, equipment and storage medium based on aerial view |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050240323A1 (en) * | 2004-03-31 | 2005-10-27 | Honda Motor Co., Ltd. | Parking lot attendant robot system |
CN106355194A (en) * | 2016-08-22 | 2017-01-25 | 广东华中科技大学工业技术研究院 | Treatment method for surface target of unmanned ship based on laser imaging radar |
US20170039436A1 (en) * | 2015-08-03 | 2017-02-09 | Nokia Technologies Oy | Fusion of RGB Images and Lidar Data for Lane Classification |
CN106650809A (en) * | 2016-12-20 | 2017-05-10 | 福州大学 | Method and system for classifying vehicle-borne laser-point cloud targets |
CN107239794A (en) * | 2017-05-18 | 2017-10-10 | 深圳市速腾聚创科技有限公司 | Point cloud data segmentation method and terminal |
CN108009509A (en) * | 2017-12-12 | 2018-05-08 | 河南工业大学 | Vehicle target detection method |
CN108830188A (en) * | 2018-05-30 | 2018-11-16 | 西安理工大学 | Vehicle checking method based on deep learning |
CN109063753A (en) * | 2018-07-18 | 2018-12-21 | 北方民族大学 | A kind of three-dimensional point cloud model classification method based on convolutional neural networks |
CN109118500A (en) * | 2018-07-16 | 2019-01-01 | 重庆大学产业技术研究院 | A kind of dividing method of the Point Cloud Data from Three Dimension Laser Scanning based on image |
US20190004508A1 (en) * | 2017-07-03 | 2019-01-03 | Volvo Car Corporation | Method and system for automatic parking of a vehicle |
CN109270543A (en) * | 2018-09-20 | 2019-01-25 | 同济大学 | A kind of system and method for pair of target vehicle surrounding vehicles location information detection |
CN109270544A (en) * | 2018-09-20 | 2019-01-25 | 同济大学 | Mobile robot self-localization system based on shaft identification |
CN109324616A (en) * | 2018-09-20 | 2019-02-12 | 同济大学 | Nobody based on onboard sensor parks the alignment method of transfer robot |
CN109344786A (en) * | 2018-10-11 | 2019-02-15 | 深圳步智造科技有限公司 | Target identification method, device and computer readable storage medium |
CN109386155A (en) * | 2018-09-20 | 2019-02-26 | 同济大学 | Nobody towards automated parking ground parks the alignment method of transfer robot |
-
2019
- 2019-03-19 CN CN201910206458.5A patent/CN110069993B/en active Active
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050240323A1 (en) * | 2004-03-31 | 2005-10-27 | Honda Motor Co., Ltd. | Parking lot attendant robot system |
US20170039436A1 (en) * | 2015-08-03 | 2017-02-09 | Nokia Technologies Oy | Fusion of RGB Images and Lidar Data for Lane Classification |
CN106355194A (en) * | 2016-08-22 | 2017-01-25 | 广东华中科技大学工业技术研究院 | Treatment method for surface target of unmanned ship based on laser imaging radar |
CN106650809A (en) * | 2016-12-20 | 2017-05-10 | 福州大学 | Method and system for classifying vehicle-borne laser-point cloud targets |
CN107239794A (en) * | 2017-05-18 | 2017-10-10 | 深圳市速腾聚创科技有限公司 | Point cloud data segmentation method and terminal |
US20190004508A1 (en) * | 2017-07-03 | 2019-01-03 | Volvo Car Corporation | Method and system for automatic parking of a vehicle |
CN108009509A (en) * | 2017-12-12 | 2018-05-08 | 河南工业大学 | Vehicle target detection method |
CN108830188A (en) * | 2018-05-30 | 2018-11-16 | 西安理工大学 | Vehicle checking method based on deep learning |
CN109118500A (en) * | 2018-07-16 | 2019-01-01 | 重庆大学产业技术研究院 | A kind of dividing method of the Point Cloud Data from Three Dimension Laser Scanning based on image |
CN109063753A (en) * | 2018-07-18 | 2018-12-21 | 北方民族大学 | A kind of three-dimensional point cloud model classification method based on convolutional neural networks |
CN109270543A (en) * | 2018-09-20 | 2019-01-25 | 同济大学 | A kind of system and method for pair of target vehicle surrounding vehicles location information detection |
CN109270544A (en) * | 2018-09-20 | 2019-01-25 | 同济大学 | Mobile robot self-localization system based on shaft identification |
CN109324616A (en) * | 2018-09-20 | 2019-02-12 | 同济大学 | Nobody based on onboard sensor parks the alignment method of transfer robot |
CN109386155A (en) * | 2018-09-20 | 2019-02-26 | 同济大学 | Nobody towards automated parking ground parks the alignment method of transfer robot |
CN109344786A (en) * | 2018-10-11 | 2019-02-15 | 深圳步智造科技有限公司 | Target identification method, device and computer readable storage medium |
Non-Patent Citations (4)
Title |
---|
AMIR KOTB 等: "Smart Parking Guidance, Monitoring and Reservations: A Review", 《 IEEE INTELLIGENT TRANSPORTATION SYSTEMS MAGAZINE》 * |
企鹅号 - AI科技大本营: "无人驾驶汽车系统入门:基于深度学习的实时激光雷达点云目标检测及ROS实现", 《HTTPS://CLOUD.TENCENT.COM/DEVELOPER/NEWS/339676》 * |
李游: "基于车载激光扫描数据的城市街道信息提取技术研究", 《中国博士学位论文全文数据库 基础科学辑》 * |
罗海峰等: "基于DBN的车载激光点云路侧多目标提取", 《测绘学报》 * |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112560834B (en) * | 2019-09-26 | 2024-05-10 | 武汉金山办公软件有限公司 | Coordinate prediction model generation method and device and pattern recognition method and device |
CN112560834A (en) * | 2019-09-26 | 2021-03-26 | 武汉金山办公软件有限公司 | Coordinate prediction model generation method and device and graph recognition method and device |
CN111178213A (en) * | 2019-12-23 | 2020-05-19 | 大连理工大学 | Aerial photography vehicle detection method based on deep learning |
CN111178213B (en) * | 2019-12-23 | 2022-11-18 | 大连理工大学 | Aerial photography vehicle detection method based on deep learning |
CN111177297B (en) * | 2019-12-31 | 2022-09-02 | 信阳师范学院 | Dynamic target speed calculation optimization method based on video and GIS |
CN111177297A (en) * | 2019-12-31 | 2020-05-19 | 信阳师范学院 | Dynamic target speed calculation optimization method based on video and GIS |
CN111523403A (en) * | 2020-04-03 | 2020-08-11 | 咪咕文化科技有限公司 | Method and device for acquiring target area in picture and computer readable storage medium |
CN111523403B (en) * | 2020-04-03 | 2023-10-20 | 咪咕文化科技有限公司 | Method and device for acquiring target area in picture and computer readable storage medium |
CN111539347A (en) * | 2020-04-27 | 2020-08-14 | 北京百度网讯科技有限公司 | Method and apparatus for detecting target |
CN111539347B (en) * | 2020-04-27 | 2023-08-08 | 北京百度网讯科技有限公司 | Method and device for detecting target |
CN111653103A (en) * | 2020-05-07 | 2020-09-11 | 浙江大华技术股份有限公司 | Target object identification method and device |
CN111783844A (en) * | 2020-06-10 | 2020-10-16 | 东莞正扬电子机械有限公司 | Target detection model training method and device based on deep learning and storage medium |
CN111783844B (en) * | 2020-06-10 | 2024-05-28 | 广东正扬传感科技股份有限公司 | Deep learning-based target detection model training method, device and storage medium |
US20220204018A1 (en) * | 2020-12-29 | 2022-06-30 | GM Global Technology Operations LLC | Vehicle control using neural network controller in combination with model-based controller |
US11834066B2 (en) * | 2020-12-29 | 2023-12-05 | GM Global Technology Operations LLC | Vehicle control using neural network controller in combination with model-based controller |
CN113313201A (en) * | 2021-06-21 | 2021-08-27 | 南京挥戈智能科技有限公司 | Multi-target detection and distance measurement method based on Swin transducer and ZED camera |
CN114692720A (en) * | 2022-02-25 | 2022-07-01 | 广州文远知行科技有限公司 | Image classification method, device, equipment and storage medium based on aerial view |
Also Published As
Publication number | Publication date |
---|---|
CN110069993B (en) | 2021-10-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110069993A (en) | A kind of target vehicle detection method based on deep learning | |
CN110163187B (en) | F-RCNN-based remote traffic sign detection and identification method | |
CN109977918B (en) | Target detection positioning optimization method based on unsupervised domain adaptation | |
CN112766087A (en) | Optical remote sensing image ship detection method based on knowledge distillation | |
CN108629288B (en) | Gesture recognition model training method, gesture recognition method and system | |
CN111444760A (en) | Traffic sign detection and identification method based on pruning and knowledge distillation | |
CN111242037A (en) | Lane line detection method based on structural information | |
CN110097047A (en) | A kind of vehicle checking method using single line laser radar based on deep learning | |
CN108550259A (en) | Congestion in road judgment method, terminal device and computer readable storage medium | |
CN111950453A (en) | Optional-shape text recognition method based on selective attention mechanism | |
CN107818302A (en) | Non-rigid multiple dimensioned object detecting method based on convolutional neural networks | |
CN113486764B (en) | Pothole detection method based on improved YOLOv3 | |
CN112364931B (en) | Few-sample target detection method and network system based on meta-feature and weight adjustment | |
CN111898432B (en) | Pedestrian detection system and method based on improved YOLOv3 algorithm | |
CN111368671A (en) | SAR image ship target detection and identification integrated method based on deep learning | |
CN108921030A (en) | A kind of SAR automatic target recognition method of Fast Learning | |
CN112418330A (en) | Improved SSD (solid State drive) -based high-precision detection method for small target object | |
CN111160407A (en) | Deep learning target detection method and system | |
CN112950780B (en) | Intelligent network map generation method and system based on remote sensing image | |
CN113177560A (en) | Universal lightweight deep learning vehicle detection method | |
CN113487600B (en) | Feature enhancement scale self-adaptive perception ship detection method | |
CN111882620A (en) | Road drivable area segmentation method based on multi-scale information | |
CN115965862A (en) | SAR ship target detection method based on mask network fusion image characteristics | |
CN113205103A (en) | Lightweight tattoo detection method | |
Wang et al. | Accurate real-time ship target detection using Yolov4 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |