CN113673484A - Road condition identification and decision-making method in unmanned driving scene - Google Patents

Road condition identification and decision-making method in unmanned driving scene Download PDF

Info

Publication number
CN113673484A
CN113673484A CN202111058219.3A CN202111058219A CN113673484A CN 113673484 A CN113673484 A CN 113673484A CN 202111058219 A CN202111058219 A CN 202111058219A CN 113673484 A CN113673484 A CN 113673484A
Authority
CN
China
Prior art keywords
vehicle
map
environment
decision
road condition
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111058219.3A
Other languages
Chinese (zh)
Inventor
皮从明
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Rongjin E Commerce Co ltd
Original Assignee
Shanghai Rongjin E Commerce Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Rongjin E Commerce Co ltd filed Critical Shanghai Rongjin E Commerce Co ltd
Priority to CN202111058219.3A priority Critical patent/CN113673484A/en
Publication of CN113673484A publication Critical patent/CN113673484A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/05Geographic models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/32Indexing scheme for image data processing or generation, in general involving image mosaicing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10028Range image; Depth image; 3D point clouds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30248Vehicle exterior or interior
    • G06T2207/30252Vehicle exterior; Vicinity of vehicle

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Geometry (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Molecular Biology (AREA)
  • Mathematical Physics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Computer Graphics (AREA)
  • Computing Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Remote Sensing (AREA)
  • Evolutionary Biology (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)

Abstract

The invention discloses a road condition identification and decision-making method in an unmanned driving scene, which comprises the following steps: firstly, a vehicle trains a picture expansion model by building a deep convolution neural network, and obtains a trained model; secondly, the vehicle realizes the relative positioning of the position of the vehicle in an unknown environment, and pictures are shot through a sensor; thirdly, the vehicle identifies and classifies the environment through a model which is matched and trained by pictures shot by the sensor; by adopting the technology of the invention, under the condition of adopting the same data set of the road condition picture in the unmanned scene, the road condition types in the picture can be identified, less storage space is occupied, the algorithm has better precision in a dynamic environment, decision opinions for guiding unmanned driving can be generated and returned to an unmanned control system, and the global semantic information obtained by semantic segmentation can help the robot to navigate and plan the path, thereby obviously improving the intelligent level of the mobile vehicle.

Description

Road condition identification and decision-making method in unmanned driving scene
Technical Field
The invention belongs to the technical field of computer image processing, and particularly relates to a road condition identification and decision-making method in an unmanned driving scene.
Background
In the field of computer image processing, a map expression mode based on a feature descriptor can complete a visual positioning task of a vehicle, but when the method is applied to road condition recognition and exploration in an unmanned scene, the method has the following defects: the traditional map based on the feature descriptor occupies more storage space and has larger estimation error, so the traditional map based on the feature descriptor cannot be directly used in the development of an unmanned control system.
Disclosure of Invention
The present invention is directed to a method for road condition identification and decision-making in an unmanned driving scene, so as to solve the problems mentioned in the background art.
In order to achieve the purpose, the invention provides the following technical scheme: a road condition identification and decision-making method in an unmanned driving scene is characterized in that: the method comprises the following steps:
firstly, a vehicle trains a picture expansion model by building a deep convolution neural network, and obtains a trained model;
secondly, the vehicle realizes the relative positioning of the position of the vehicle in an unknown environment, and pictures are shot through a sensor;
thirdly, the vehicle identifies and classifies the environment through a model which is matched and trained by pictures shot by the sensor;
step four, starting moving exploration of the vehicle through an unknown map path exploration algorithm, carrying out self-positioning through observation of the environment, and simultaneously building an incremental map;
fifthly, the vehicle realizes real-time construction and updating of the map during exploration through multi-sensor fusion data until the map construction is completed;
and step six, the vehicle control system makes a decision according to the sensed information.
Preferably, in the third step, the pose estimation of the visual SLAM algorithm is adopted.
Preferably, in the fourth step, local map splicing and global map model construction are realized through point cloud splicing and filtering technology.
Preferably, the semantic map is constructed in a dynamic environment by combining the semantic information with the SLAM algorithm adopted by the invention in the fifth step.
Compared with the prior art, the invention has the beneficial effects that: by adopting the technology of the invention, under the condition of adopting the same data set of the road condition picture in the unmanned scene, the road condition types in the picture can be identified, less storage space is occupied, the algorithm has better precision in a dynamic environment, decision opinions for guiding unmanned driving can be generated and returned to an unmanned control system, and the global semantic information obtained by semantic segmentation can help the robot to navigate and plan the path, thereby obviously improving the intelligent level of the mobile vehicle.
Drawings
FIG. 1 is a schematic diagram of the principles of the present invention;
FIG. 2 is a schematic view of the workflow of pose estimation of the visual SLAM algorithm of the present invention;
FIG. 3 is a schematic diagram of a method for local map stitching and global map model construction according to the present invention;
FIG. 4 is a schematic diagram of a single frame semantic point generation algorithm of the present invention;
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Example 1
Referring to fig. 1 to 4, the present invention provides a technical solution: a method for road condition identification and decision-making in an unmanned driving scene, comprising 1) convolutional neural network-based image semantic segmentation: building a full convolution neural network model for semantic segmentation, and training through a data set, wherein the data set comprises 11335 semantic segmentation images, the semantic segmentation images are divided into 20 types of objects, and the resolution ratio is about 500 x 300; for semantic segmentation tasks, a training set and a verification set of the data set both have corresponding labels, and the labels store corresponding semantic categories of each pixel in an image; if other categories need to be identified, the trained network can be finely adjusted by using an additional data stage, pictures are cut, and the resolution of the input image is unified to 360 x 480; adopting a multi-distribution learning rate, taking a larger learning rate at the early stage of neural network training, and attenuating the learning rate to be 0.1 time of the original learning rate when reaching a certain set training frequency along with the training; the method can enable the neural network to reach the vicinity of the optimal value quickly without fluctuating back and forth around the optimal value.
2) Pose estimation of visual SLAM algorithm: referring to the work flow of fig. 2, the position and orientation of the camera are estimated by using the SLAM algorithm of the feature point method, the projection of the same space point in different images is obtained by matching the features, the relation between the position and the posture of the images is calculated, and the motion track of the moving vehicle of the camera is obtained.
3) Local map splicing and global map model construction are realized through point cloud splicing and filtering technology: after the SLAM algorithm is used for processing a moving object in a dynamic scene and a pose graph with the pose as a variable is optimized, the optimized globally consistent mobile robot key frame pose can be obtained; accordingly, the point clouds of the key frames are spliced to obtain a local map of the environment, and the method is as shown in fig. 3; converting the point cloud map corresponding to each key frame into a world coordinate system by using the pose of the key frame, and connecting to obtain a local global point cloud map; filtering outliers in the point cloud map by adopting a filtering technology to realize map optimization and complete map construction; the mathematical expression is as follows:
Figure BDA0003255395880000031
in the formula:
Ck-key(s)A single frame point cloud formed by the frame images,
Tkthe pose of the camera at the moment of the frame,
and (4) local point cloud under the m-camera coordinate system.
4) Combining the semantic information with the SLAM algorithm adopted by the invention to complete the construction of the semantic map under the dynamic environment: adding semantic information to the acquired point cloud information, identifying the class information of an object in the image by semantic segmentation of the image, fusing the segmentation result of the two-dimensional image into the point cloud, and giving semantic information to each point in the point cloud so as to obtain a semantic point cloud map with rich information; adopting different colors to correspond to objects of a specific category, namely semantic information of scenes; the single frame semantic point generation algorithm is shown in fig. 4.
5) And generating the unmanned decision-making opinions according to the obtained road condition types, returning the unmanned decision-making opinions to the unmanned control system, and generating the unmanned decision-making opinions according to the road condition types.
Although embodiments of the present invention have been shown and described, it will be appreciated by those skilled in the art that changes, modifications, substitutions and alterations can be made in these embodiments without departing from the principles and spirit of the invention, the scope of which is defined in the appended claims and their equivalents.

Claims (4)

1. A road condition identification and decision-making method in an unmanned driving scene is characterized in that: the method comprises the following steps:
firstly, a vehicle trains a picture expansion model by building a deep convolution neural network, and obtains a trained model;
secondly, the vehicle realizes the relative positioning of the position of the vehicle in an unknown environment, and pictures are shot through a sensor;
thirdly, the vehicle identifies and classifies the environment through a model which is matched and trained by pictures shot by the sensor;
step four, starting moving exploration of the vehicle through an unknown map path exploration algorithm, carrying out self-positioning through observation of the environment, and simultaneously building an incremental map;
fifthly, the vehicle realizes real-time construction and updating of the map during exploration through multi-sensor fusion data until the map construction is completed;
and step six, the vehicle control system makes a decision according to the sensed information.
2. The method of claim 1, wherein the method comprises the steps of: and in the third step, the pose estimation of the visual SLAM algorithm is adopted.
3. The method of claim 1, wherein the method comprises the steps of: and in the fourth step, local map splicing and global map model construction are realized through point cloud splicing and filtering technology.
4. The method of claim 1, wherein the method comprises the steps of: and fifthly, building the semantic map in the dynamic environment by combining the semantic information with the SLAM algorithm adopted by the invention.
CN202111058219.3A 2021-09-09 2021-09-09 Road condition identification and decision-making method in unmanned driving scene Pending CN113673484A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111058219.3A CN113673484A (en) 2021-09-09 2021-09-09 Road condition identification and decision-making method in unmanned driving scene

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111058219.3A CN113673484A (en) 2021-09-09 2021-09-09 Road condition identification and decision-making method in unmanned driving scene

Publications (1)

Publication Number Publication Date
CN113673484A true CN113673484A (en) 2021-11-19

Family

ID=78549224

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111058219.3A Pending CN113673484A (en) 2021-09-09 2021-09-09 Road condition identification and decision-making method in unmanned driving scene

Country Status (1)

Country Link
CN (1) CN113673484A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114445593A (en) * 2022-01-30 2022-05-06 重庆长安汽车股份有限公司 Aerial view semantic segmentation label generation method based on multi-frame semantic point cloud splicing

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109117718A (en) * 2018-07-02 2019-01-01 东南大学 A kind of semantic map structuring of three-dimensional towards road scene and storage method
CN110363816A (en) * 2019-06-25 2019-10-22 广东工业大学 A kind of mobile robot environment semanteme based on deep learning builds drawing method
CN111368759A (en) * 2020-03-09 2020-07-03 河海大学常州校区 Monocular vision-based semantic map construction system for mobile robot
US20200364554A1 (en) * 2018-02-09 2020-11-19 Baidu Usa Llc Systems and methods for deep localization and segmentation with a 3d semantic map

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200364554A1 (en) * 2018-02-09 2020-11-19 Baidu Usa Llc Systems and methods for deep localization and segmentation with a 3d semantic map
CN109117718A (en) * 2018-07-02 2019-01-01 东南大学 A kind of semantic map structuring of three-dimensional towards road scene and storage method
CN110363816A (en) * 2019-06-25 2019-10-22 广东工业大学 A kind of mobile robot environment semanteme based on deep learning builds drawing method
CN111368759A (en) * 2020-03-09 2020-07-03 河海大学常州校区 Monocular vision-based semantic map construction system for mobile robot

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
何松 等: "基于激光SLAM和深度学习的语义地图构建", 《计算机技术与发展》, vol. 30, no. 9, pages 88 - 94 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114445593A (en) * 2022-01-30 2022-05-06 重庆长安汽车股份有限公司 Aerial view semantic segmentation label generation method based on multi-frame semantic point cloud splicing
CN114445593B (en) * 2022-01-30 2024-05-10 重庆长安汽车股份有限公司 Bird's eye view semantic segmentation label generation method based on multi-frame semantic point cloud splicing

Similar Documents

Publication Publication Date Title
Toft et al. Long-term visual localization revisited
Sahu et al. Artificial intelligence (AI) in augmented reality (AR)-assisted manufacturing applications: a review
US10854006B2 (en) AR-enabled labeling using aligned CAD models
CN110135249B (en) Human behavior identification method based on time attention mechanism and LSTM (least Square TM)
CN110119148B (en) Six-degree-of-freedom attitude estimation method and device and computer readable storage medium
CN109815847B (en) Visual SLAM method based on semantic constraint
Liu et al. RGB-D inertial odometry for a resource-restricted robot in dynamic environments
CN107967457A (en) A kind of place identification for adapting to visual signature change and relative positioning method and system
Wen et al. Hybrid semi-dense 3D semantic-topological mapping from stereo visual-inertial odometry SLAM with loop closure detection
Li et al. Overview of deep learning application on visual SLAM
WO2019175532A1 (en) Urban environment labelling
Cheng et al. Improving dense mapping for mobile robots in dynamic environments based on semantic information
Liu et al. D-lc-nets: Robust denoising and loop closing networks for lidar slam in complicated circumstances with noisy point clouds
Civera et al. Rgb-d odometry and slam
Li et al. Multi-modal neural feature fusion for automatic driving through perception-aware path planning
Dai et al. A review of common techniques for visual simultaneous localization and mapping
CN113673484A (en) Road condition identification and decision-making method in unmanned driving scene
CN114399515A (en) Language description-based class-level target object 6D pose acquisition method and storage medium
CN117788544A (en) Image depth estimation method based on lightweight attention mechanism
Zhang et al. Diff-tracker: Text-to-image diffusion models are unsupervised trackers
CN112348854A (en) Visual inertial mileage detection method based on deep learning
Liedtke et al. Use of explicit knowledge for the reconstruction of 3-D object geometry
CN116721139A (en) Generating depth images of image data
CN114943785A (en) Map construction method, map construction device, map construction equipment and storage medium
Liu et al. An efficient edge-feature constraint visual SLAM

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination