CN117253138A - Deep learning position recognition algorithm based on laser radar - Google Patents
Deep learning position recognition algorithm based on laser radar Download PDFInfo
- Publication number
- CN117253138A CN117253138A CN202210656594.6A CN202210656594A CN117253138A CN 117253138 A CN117253138 A CN 117253138A CN 202210656594 A CN202210656594 A CN 202210656594A CN 117253138 A CN117253138 A CN 117253138A
- Authority
- CN
- China
- Prior art keywords
- point
- point cloud
- laser radar
- descriptors
- value
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000004422 calculation algorithm Methods 0.000 title claims description 7
- 238000013135 deep learning Methods 0.000 title claims description 6
- 239000011159 matrix material Substances 0.000 claims description 7
- 230000006870 function Effects 0.000 claims description 3
- 230000004931 aggregating effect Effects 0.000 claims description 2
- 238000000034 method Methods 0.000 abstract description 8
- 238000001514 detection method Methods 0.000 abstract description 7
- 230000007613 environmental effect Effects 0.000 abstract description 4
- 238000005516 engineering process Methods 0.000 abstract description 3
- 238000000605 extraction Methods 0.000 abstract description 3
- 238000012821 model calculation Methods 0.000 abstract description 2
- 230000001737 promoting effect Effects 0.000 abstract description 2
- 238000004364 calculation method Methods 0.000 abstract 1
- 230000007547 defect Effects 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/10—Terrestrial scenes
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/4802—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/22—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/74—Image or video pattern matching; Proximity measures in feature spaces
- G06V10/761—Proximity, similarity or dissimilarity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/77—Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
- G06V10/774—Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A90/00—Technologies having an indirect contribution to adaptation to climate change
- Y02A90/10—Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computing Systems (AREA)
- Artificial Intelligence (AREA)
- Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Software Systems (AREA)
- Computer Networks & Wireless Communication (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Image Analysis (AREA)
Abstract
The point cloud position recognition technology in the dynamic environment has the advantages of high detection precision, high detection speed and small model calculation amount. The method adopts a characteristic point extraction module to greatly reduce the scale of the point cloud and retain the data characteristics so as to lighten the influence of environmental change on data acquisition, and designs a Point Transformer module which introduces Multi Head Attention capable of promoting local information exchange to control the calculated amount and extract local descriptors; feature Attention is designed to aggregate recognizable global descriptors; thereby improving the accuracy and speed of the point cloud position identification and reducing the network calculation amount.
Description
Technical Field
The deep learning position recognition algorithm based on the laser radar is used for improving the accuracy and speed of position recognition.
Technical Field
In recent years, some vision-based position recognition methods are proposed, but the vision sensor has high sensitivity to light, and is difficult to apply to dynamic environments; as a reliable alternative, 3D lidar can acquire accurate and detailed three-dimensional information that is inherently immune to illumination variations; however, the position recognition method based on the 3D lidar cannot obtain an effective and robust detection effect due to environmental factors such as sensor shielding and viewpoint change.
From the recent academic research, the accuracy and the robustness of the position recognition technology based on the point cloud in the dynamic environment are difficult to be ensured due to environmental factors such as sensor shielding, viewpoint change and the like; the existing method with high detection precision has the defects that a network model is too large, the detection speed of the method with small network model is not fast enough, and the method is difficult to be applied to a real scene; in order to solve the above problems, we propose an efficient and lightweight point cloud location recognition learning network combining feature point extraction and transformation.
Disclosure of Invention
The point cloud position recognition technology in the dynamic environment has the advantages of high detection precision, high detection speed and small model calculation amount.
The characteristic point extraction module is adopted to greatly reduce the scale of the point cloud and preserve the data characteristics so as to reduce the influence of environmental change on data acquisition, the Point Transformer module is designed to introduce a multi-head attention capable of promoting local information exchange to control the calculated amount, the expression capacity of the model is enhanced, and the accuracy and the speed of point cloud position identification are improved and the network calculated amount is reduced.
A deep learning position recognition algorithm based on a laser radar comprises the following steps:
step one: firstly, calculating a horizontal index and a scan index of each point in point cloud data acquired by a laser radar; the expression is as follows:
wherein x, y and z are three-dimensional coordinate values of each point; delta is the absolute value of the angle which can be scanned under the horizontal axis of the laser radar;
step two: each point is cycled, and the point r is used i Calculating the curvature value of each point by using a sub-point cloud S consisting of five points which are positioned in front of and behind the same scan index and closest to the horizontal index; discarding if the curvature value of the point is smaller than a certain threshold value; finally obtaining characteristic points; the formula is expressed as:
C i <σ (4)
step three: the feature points are copied into 3 parts of data K, Q and V with equal size, and the data K, Q and V are processed by Multi Head Attention to obtain local descriptors; the formula is expressed as follows:
MultiHead(Q,K,V)=Concat(head 1 ,...,head h )W O (4)
wherein W is i Q ,W i K And W is i V ∈R D×(D/h) ,W O ∈R D×D Are all a matrix of learnable parameters;
step four: aggregating global descriptors by assigning local descriptor weights to Feature Attention; the formula is expressed as follows:
wherein N is the number of local descriptors, u is the number of local descriptors, and J is a learnable matrix parameter;
step five: obtaining corresponding scene similarity by comparing the global descriptor similarity; the formula is described as follows:
wherein omega [1:s] ∈R D×D×S ,α∈R 2D×S Is a learnable matrix parameter b epsilon R S Is a learnable vector parameter, S is the dimension size of the global descriptor.
Step six: in the training stage, the model has the function of learning by obtaining scene similarity output, returning to the input position to compare the true value, comparing the difference value between the scene similarity output and the true value, and modifying the weight parameters in the next training again.
The method is mainly characterized in that redundant original point clouds firstly extract characteristic points, meaningless points are removed, the scale of the point clouds is greatly reduced, and the calculated amount is further reduced; the Point Transformer module learns the local features of the feature points well and aggregates the discernable global descriptors.
Drawings
Description of the drawings figure 1 is a general flow chart of the invention
FIG. 2 is a flow chart of the Muti-Head addition
Description of the drawings figure 3 is a flow chart of Feature Attention
Description of the drawings figure 4 is a flow chart of Feature Similarity Network
FIG. 5 is a partial experimental result diagram
Detailed description of the preferred embodiments
As shown in the overall flow of FIG. 1, the invention herein provides a deep learning position based on laser radar
An identification algorithm.
Firstly, extracting point cloud data from two paired scenes through a Feature Point Extrarction module to obtain sub point clouds with rich characteristics. Then extracting the local descriptor F through the Muti-Head attribute layer in the Point Transformer module L Feature Attention layer aggregates distinct global descriptors. Finally, two global descriptors e to be paired 1 And e 2 And the incoming Feature similarity network module obtains the scene similarity. The method specifically comprises the following steps:
an origin point cloud is acquired in a scene using a lidar sensor.
The original point cloud is first projected as a Range image and then the scan index and the horizontal index for each point are calculated.
Each point in the original point cloud is traversed circularly, the curvature value of the point is calculated through a sub point cloud S formed by five points which are positioned in the same scan index and closest to the point, and if the curvature value of the point is smaller than a certain threshold value, the curvature value of the point is omitted; the final rest is the feature points;
as shown in fig. 2, the feature points are mapped into three equal-sized parts, denoted by K, Q and V, using a linear projection layer; and then dividing K, Q and V into h parts respectively and transmitting the h parts into Multi Head Attention for extracting local descriptors.
As shown in fig. 3, the global context is constructed by multiplying the learning matrix J with the average value of all the local descriptors to provide global structure and feature information; and multiplying the local descriptor by using the global context, ensuring the obtained weight value to be 0,1 by using a sigmoid function, and finally giving the weight to the local descriptor to obtain the final global descriptor through summation.
As shown in fig. 4, feature Similarity Network is used to obtain the similarity of two global descriptors, so as to obtain the similarity of the scene.
As shown in fig. 5, our algorithm can well distinguish subtle differences in scenes, giving a suitable score for the similarity of two scenes; scene 1 and the compared scene are very similar in terms of both whole and detail, so that a higher similarity score is obtained, and scene 2 is completely opposite; scene 3, while similar to the compared scene as a whole, is not identical in detail, and therefore results in a lower similarity score.
Claims (1)
1. A deep learning position recognition algorithm based on laser radar is characterized by comprising the following steps,
step one: firstly, calculating a horizontal index and a scan index of each point in point cloud data acquired by a laser radar; the expression is as follows:
wherein x, y and z are three-dimensional coordinate values of each point; delta is the absolute value of the angle which can be scanned under the horizontal axis of the laser radar;
step two: each point is cycled, and the point r is used i Calculating the curvature value of each point by using a sub-point cloud S consisting of five points which are positioned in front of and behind the same scan index and closest to the horizontal index; discarding if the curvature value of the point is smaller than a certain threshold value; finally obtaining characteristic points; the formula is expressed as:
C i <σ (4)
step three: the feature points are copied into 3 parts of data K, Q and V with equal size, and the data K, Q and V are processed by Multi Head Attention to obtain local descriptors; the formula is expressed as follows:
MultiHead(Q,K,V)=Concat(head 1 ,...,head h )W O (4)
wherein W is i Q ,W i K And W is i V ∈R D×(D/h) ,W O ∈R D×D Are all a matrix of learnable parameters;
step four: aggregating global descriptors by assigning local descriptor weights to Feature Attention; the formula is expressed as follows:
wherein N is the number of local descriptors, u is the number of local descriptors, and J is a learnable matrix parameter;
step five: obtaining corresponding scene similarity by comparing the global descriptor similarity; the formula is described as follows:
wherein omega [1:s] ∈R D×D×S ,α∈R 2D×S Is a learnable matrix parameter b epsilon R S Is a learnable vector parameter, S is the dimension size of the global descriptor.
Step six: in the training stage, the model has the function of learning by obtaining scene similarity output, returning to the input position to compare the true value, comparing the difference value between the scene similarity output and the true value, and modifying the weight parameters in the next training again.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210656594.6A CN117253138A (en) | 2022-06-10 | 2022-06-10 | Deep learning position recognition algorithm based on laser radar |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210656594.6A CN117253138A (en) | 2022-06-10 | 2022-06-10 | Deep learning position recognition algorithm based on laser radar |
Publications (1)
Publication Number | Publication Date |
---|---|
CN117253138A true CN117253138A (en) | 2023-12-19 |
Family
ID=89131900
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210656594.6A Pending CN117253138A (en) | 2022-06-10 | 2022-06-10 | Deep learning position recognition algorithm based on laser radar |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN117253138A (en) |
-
2022
- 2022-06-10 CN CN202210656594.6A patent/CN117253138A/en active Pending
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110533722B (en) | Robot rapid repositioning method and system based on visual dictionary | |
CN110472627B (en) | End-to-end SAR image recognition method, device and storage medium | |
CN110322511B (en) | Semantic SLAM method and system based on object and plane features | |
CN108648161B (en) | Binocular vision obstacle detection system and method of asymmetric kernel convolution neural network | |
CN110009674B (en) | Monocular image depth of field real-time calculation method based on unsupervised depth learning | |
CN110853075B (en) | Visual tracking positioning method based on dense point cloud and synthetic view | |
CN110097553A (en) | The semanteme for building figure and three-dimensional semantic segmentation based on instant positioning builds drawing system | |
CN110689562A (en) | Trajectory loop detection optimization method based on generation of countermeasure network | |
CN111998862B (en) | BNN-based dense binocular SLAM method | |
CN114140527B (en) | Dynamic environment binocular vision SLAM method based on semantic segmentation | |
Yang et al. | Visual SLAM based on semantic segmentation and geometric constraints for dynamic indoor environments | |
Shi et al. | An improved lightweight deep neural network with knowledge distillation for local feature extraction and visual localization using images and LiDAR point clouds | |
CN114067075A (en) | Point cloud completion method and device based on generation of countermeasure network | |
CN116703996A (en) | Monocular three-dimensional target detection algorithm based on instance-level self-adaptive depth estimation | |
Gao et al. | Efficient view-based 3-D object retrieval via hypergraph learning | |
Konishi et al. | Detection of target persons using deep learning and training data generation for Tsukuba challenge | |
CN117253138A (en) | Deep learning position recognition algorithm based on laser radar | |
CN115496859A (en) | Three-dimensional scene motion trend estimation method based on scattered point cloud cross attention learning | |
CN115718303A (en) | Dynamic environment-oriented camera and solid-state laser radar fusion repositioning method | |
Wada et al. | Dataset Genreratoin for Semantic Segmentation from 3D Scanned Data Considering Domain Gap | |
Deng et al. | Fusion Scan Context: A Global Descriptor Fusing Altitude, Intensity and Density for Place Recognition | |
CN114549917B (en) | Point cloud classification method with enhanced data characterization | |
Su | Vanishing points in road recognition: A review | |
Liu et al. | Deep learning of volumetric representation for 3D object recognition | |
CN117593618B (en) | Point cloud generation method based on nerve radiation field and depth map |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |