CN117253138A - Deep learning position recognition algorithm based on laser radar - Google Patents

Deep learning position recognition algorithm based on laser radar Download PDF

Info

Publication number
CN117253138A
CN117253138A CN202210656594.6A CN202210656594A CN117253138A CN 117253138 A CN117253138 A CN 117253138A CN 202210656594 A CN202210656594 A CN 202210656594A CN 117253138 A CN117253138 A CN 117253138A
Authority
CN
China
Prior art keywords
point
point cloud
laser radar
descriptors
value
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202210656594.6A
Other languages
Chinese (zh)
Inventor
叶涛
严翔明
汪寿安
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China University of Mining and Technology Beijing CUMTB
Original Assignee
China University of Mining and Technology Beijing CUMTB
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China University of Mining and Technology Beijing CUMTB filed Critical China University of Mining and Technology Beijing CUMTB
Priority to CN202210656594.6A priority Critical patent/CN117253138A/en
Publication of CN117253138A publication Critical patent/CN117253138A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S7/00Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
    • G01S7/48Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
    • G01S7/4802Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00 using analysis of echo signal for target characterisation; Target signature; Target cross-section
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/22Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/761Proximity, similarity or dissimilarity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/77Processing image or video features in feature spaces; using data integration or data reduction, e.g. principal component analysis [PCA] or independent component analysis [ICA] or self-organising maps [SOM]; Blind source separation
    • G06V10/774Generating sets of training patterns; Bootstrap methods, e.g. bagging or boosting
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Image Analysis (AREA)

Abstract

The point cloud position recognition technology in the dynamic environment has the advantages of high detection precision, high detection speed and small model calculation amount. The method adopts a characteristic point extraction module to greatly reduce the scale of the point cloud and retain the data characteristics so as to lighten the influence of environmental change on data acquisition, and designs a Point Transformer module which introduces Multi Head Attention capable of promoting local information exchange to control the calculated amount and extract local descriptors; feature Attention is designed to aggregate recognizable global descriptors; thereby improving the accuracy and speed of the point cloud position identification and reducing the network calculation amount.

Description

Deep learning position recognition algorithm based on laser radar
Technical Field
The deep learning position recognition algorithm based on the laser radar is used for improving the accuracy and speed of position recognition.
Technical Field
In recent years, some vision-based position recognition methods are proposed, but the vision sensor has high sensitivity to light, and is difficult to apply to dynamic environments; as a reliable alternative, 3D lidar can acquire accurate and detailed three-dimensional information that is inherently immune to illumination variations; however, the position recognition method based on the 3D lidar cannot obtain an effective and robust detection effect due to environmental factors such as sensor shielding and viewpoint change.
From the recent academic research, the accuracy and the robustness of the position recognition technology based on the point cloud in the dynamic environment are difficult to be ensured due to environmental factors such as sensor shielding, viewpoint change and the like; the existing method with high detection precision has the defects that a network model is too large, the detection speed of the method with small network model is not fast enough, and the method is difficult to be applied to a real scene; in order to solve the above problems, we propose an efficient and lightweight point cloud location recognition learning network combining feature point extraction and transformation.
Disclosure of Invention
The point cloud position recognition technology in the dynamic environment has the advantages of high detection precision, high detection speed and small model calculation amount.
The characteristic point extraction module is adopted to greatly reduce the scale of the point cloud and preserve the data characteristics so as to reduce the influence of environmental change on data acquisition, the Point Transformer module is designed to introduce a multi-head attention capable of promoting local information exchange to control the calculated amount, the expression capacity of the model is enhanced, and the accuracy and the speed of point cloud position identification are improved and the network calculated amount is reduced.
A deep learning position recognition algorithm based on a laser radar comprises the following steps:
step one: firstly, calculating a horizontal index and a scan index of each point in point cloud data acquired by a laser radar; the expression is as follows:
wherein x, y and z are three-dimensional coordinate values of each point; delta is the absolute value of the angle which can be scanned under the horizontal axis of the laser radar;
step two: each point is cycled, and the point r is used i Calculating the curvature value of each point by using a sub-point cloud S consisting of five points which are positioned in front of and behind the same scan index and closest to the horizontal index; discarding if the curvature value of the point is smaller than a certain threshold value; finally obtaining characteristic points; the formula is expressed as:
C i <σ (4)
step three: the feature points are copied into 3 parts of data K, Q and V with equal size, and the data K, Q and V are processed by Multi Head Attention to obtain local descriptors; the formula is expressed as follows:
MultiHead(Q,K,V)=Concat(head 1 ,...,head h )W O (4)
wherein W is i Q ,W i K And W is i V ∈R D×(D/h) ,W O ∈R D×D Are all a matrix of learnable parameters;
step four: aggregating global descriptors by assigning local descriptor weights to Feature Attention; the formula is expressed as follows:
wherein N is the number of local descriptors, u is the number of local descriptors, and J is a learnable matrix parameter;
step five: obtaining corresponding scene similarity by comparing the global descriptor similarity; the formula is described as follows:
wherein omega [1:s] ∈R D×D×S ,α∈R 2D×S Is a learnable matrix parameter b epsilon R S Is a learnable vector parameter, S is the dimension size of the global descriptor.
Step six: in the training stage, the model has the function of learning by obtaining scene similarity output, returning to the input position to compare the true value, comparing the difference value between the scene similarity output and the true value, and modifying the weight parameters in the next training again.
The method is mainly characterized in that redundant original point clouds firstly extract characteristic points, meaningless points are removed, the scale of the point clouds is greatly reduced, and the calculated amount is further reduced; the Point Transformer module learns the local features of the feature points well and aggregates the discernable global descriptors.
Drawings
Description of the drawings figure 1 is a general flow chart of the invention
FIG. 2 is a flow chart of the Muti-Head addition
Description of the drawings figure 3 is a flow chart of Feature Attention
Description of the drawings figure 4 is a flow chart of Feature Similarity Network
FIG. 5 is a partial experimental result diagram
Detailed description of the preferred embodiments
As shown in the overall flow of FIG. 1, the invention herein provides a deep learning position based on laser radar
An identification algorithm.
Firstly, extracting point cloud data from two paired scenes through a Feature Point Extrarction module to obtain sub point clouds with rich characteristics. Then extracting the local descriptor F through the Muti-Head attribute layer in the Point Transformer module L Feature Attention layer aggregates distinct global descriptors. Finally, two global descriptors e to be paired 1 And e 2 And the incoming Feature similarity network module obtains the scene similarity. The method specifically comprises the following steps:
an origin point cloud is acquired in a scene using a lidar sensor.
The original point cloud is first projected as a Range image and then the scan index and the horizontal index for each point are calculated.
Each point in the original point cloud is traversed circularly, the curvature value of the point is calculated through a sub point cloud S formed by five points which are positioned in the same scan index and closest to the point, and if the curvature value of the point is smaller than a certain threshold value, the curvature value of the point is omitted; the final rest is the feature points;
as shown in fig. 2, the feature points are mapped into three equal-sized parts, denoted by K, Q and V, using a linear projection layer; and then dividing K, Q and V into h parts respectively and transmitting the h parts into Multi Head Attention for extracting local descriptors.
As shown in fig. 3, the global context is constructed by multiplying the learning matrix J with the average value of all the local descriptors to provide global structure and feature information; and multiplying the local descriptor by using the global context, ensuring the obtained weight value to be 0,1 by using a sigmoid function, and finally giving the weight to the local descriptor to obtain the final global descriptor through summation.
As shown in fig. 4, feature Similarity Network is used to obtain the similarity of two global descriptors, so as to obtain the similarity of the scene.
As shown in fig. 5, our algorithm can well distinguish subtle differences in scenes, giving a suitable score for the similarity of two scenes; scene 1 and the compared scene are very similar in terms of both whole and detail, so that a higher similarity score is obtained, and scene 2 is completely opposite; scene 3, while similar to the compared scene as a whole, is not identical in detail, and therefore results in a lower similarity score.

Claims (1)

1. A deep learning position recognition algorithm based on laser radar is characterized by comprising the following steps,
step one: firstly, calculating a horizontal index and a scan index of each point in point cloud data acquired by a laser radar; the expression is as follows:
wherein x, y and z are three-dimensional coordinate values of each point; delta is the absolute value of the angle which can be scanned under the horizontal axis of the laser radar;
step two: each point is cycled, and the point r is used i Calculating the curvature value of each point by using a sub-point cloud S consisting of five points which are positioned in front of and behind the same scan index and closest to the horizontal index; discarding if the curvature value of the point is smaller than a certain threshold value; finally obtaining characteristic points; the formula is expressed as:
C i <σ (4)
step three: the feature points are copied into 3 parts of data K, Q and V with equal size, and the data K, Q and V are processed by Multi Head Attention to obtain local descriptors; the formula is expressed as follows:
MultiHead(Q,K,V)=Concat(head 1 ,...,head h )W O (4)
wherein W is i Q ,W i K And W is i V ∈R D×(D/h) ,W O ∈R D×D Are all a matrix of learnable parameters;
step four: aggregating global descriptors by assigning local descriptor weights to Feature Attention; the formula is expressed as follows:
wherein N is the number of local descriptors, u is the number of local descriptors, and J is a learnable matrix parameter;
step five: obtaining corresponding scene similarity by comparing the global descriptor similarity; the formula is described as follows:
wherein omega [1:s] ∈R D×D×S ,α∈R 2D×S Is a learnable matrix parameter b epsilon R S Is a learnable vector parameter, S is the dimension size of the global descriptor.
Step six: in the training stage, the model has the function of learning by obtaining scene similarity output, returning to the input position to compare the true value, comparing the difference value between the scene similarity output and the true value, and modifying the weight parameters in the next training again.
CN202210656594.6A 2022-06-10 2022-06-10 Deep learning position recognition algorithm based on laser radar Pending CN117253138A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210656594.6A CN117253138A (en) 2022-06-10 2022-06-10 Deep learning position recognition algorithm based on laser radar

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210656594.6A CN117253138A (en) 2022-06-10 2022-06-10 Deep learning position recognition algorithm based on laser radar

Publications (1)

Publication Number Publication Date
CN117253138A true CN117253138A (en) 2023-12-19

Family

ID=89131900

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210656594.6A Pending CN117253138A (en) 2022-06-10 2022-06-10 Deep learning position recognition algorithm based on laser radar

Country Status (1)

Country Link
CN (1) CN117253138A (en)

Similar Documents

Publication Publication Date Title
CN110533722B (en) Robot rapid repositioning method and system based on visual dictionary
CN110472627B (en) End-to-end SAR image recognition method, device and storage medium
CN110322511B (en) Semantic SLAM method and system based on object and plane features
CN108648161B (en) Binocular vision obstacle detection system and method of asymmetric kernel convolution neural network
CN110009674B (en) Monocular image depth of field real-time calculation method based on unsupervised depth learning
CN110853075B (en) Visual tracking positioning method based on dense point cloud and synthetic view
CN110097553A (en) The semanteme for building figure and three-dimensional semantic segmentation based on instant positioning builds drawing system
CN110689562A (en) Trajectory loop detection optimization method based on generation of countermeasure network
CN111998862B (en) BNN-based dense binocular SLAM method
CN114140527B (en) Dynamic environment binocular vision SLAM method based on semantic segmentation
Yang et al. Visual SLAM based on semantic segmentation and geometric constraints for dynamic indoor environments
Shi et al. An improved lightweight deep neural network with knowledge distillation for local feature extraction and visual localization using images and LiDAR point clouds
CN114067075A (en) Point cloud completion method and device based on generation of countermeasure network
CN116703996A (en) Monocular three-dimensional target detection algorithm based on instance-level self-adaptive depth estimation
Gao et al. Efficient view-based 3-D object retrieval via hypergraph learning
Konishi et al. Detection of target persons using deep learning and training data generation for Tsukuba challenge
CN117253138A (en) Deep learning position recognition algorithm based on laser radar
CN115496859A (en) Three-dimensional scene motion trend estimation method based on scattered point cloud cross attention learning
CN115718303A (en) Dynamic environment-oriented camera and solid-state laser radar fusion repositioning method
Wada et al. Dataset Genreratoin for Semantic Segmentation from 3D Scanned Data Considering Domain Gap
Deng et al. Fusion Scan Context: A Global Descriptor Fusing Altitude, Intensity and Density for Place Recognition
CN114549917B (en) Point cloud classification method with enhanced data characterization
Su Vanishing points in road recognition: A review
Liu et al. Deep learning of volumetric representation for 3D object recognition
CN117593618B (en) Point cloud generation method based on nerve radiation field and depth map

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination