WO2022252346A1 - 3d地图的检索方法和装置 - Google Patents
3d地图的检索方法和装置 Download PDFInfo
- Publication number
- WO2022252346A1 WO2022252346A1 PCT/CN2021/105624 CN2021105624W WO2022252346A1 WO 2022252346 A1 WO2022252346 A1 WO 2022252346A1 CN 2021105624 W CN2021105624 W CN 2021105624W WO 2022252346 A1 WO2022252346 A1 WO 2022252346A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- map
- data
- descriptors
- retrieval
- descriptor
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 230
- 230000006837 decompression Effects 0.000 claims abstract description 163
- 230000000007 visual effect Effects 0.000 claims description 187
- 238000012545 processing Methods 0.000 claims description 94
- 230000008569 process Effects 0.000 claims description 36
- 238000013139 quantization Methods 0.000 claims description 23
- 238000000605 extraction Methods 0.000 claims description 20
- 230000004044 response Effects 0.000 claims description 12
- 238000004590 computer program Methods 0.000 claims description 10
- 230000001960 triggered effect Effects 0.000 claims description 3
- 230000006872 improvement Effects 0.000 abstract description 6
- 238000012216 screening Methods 0.000 abstract description 6
- 238000013461 design Methods 0.000 description 73
- 239000000284 extract Substances 0.000 description 38
- 238000004891 communication Methods 0.000 description 22
- 238000010586 diagram Methods 0.000 description 21
- 239000012634 fragment Substances 0.000 description 11
- 238000004422 calculation algorithm Methods 0.000 description 8
- 230000006835 compression Effects 0.000 description 8
- 238000007906 compression Methods 0.000 description 8
- 238000005516 engineering process Methods 0.000 description 7
- 230000005540 biological transmission Effects 0.000 description 5
- 238000001514 detection method Methods 0.000 description 5
- 230000006870 function Effects 0.000 description 5
- 229920001621 AMOLED Polymers 0.000 description 4
- 230000007423 decrease Effects 0.000 description 4
- 230000004807 localization Effects 0.000 description 4
- 238000005259 measurement Methods 0.000 description 4
- 230000009471 action Effects 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 3
- 238000013500 data storage Methods 0.000 description 3
- 239000011521 glass Substances 0.000 description 3
- 230000033001 locomotion Effects 0.000 description 3
- 238000007726 management method Methods 0.000 description 3
- 230000003287 optical effect Effects 0.000 description 3
- 238000003672 processing method Methods 0.000 description 3
- 230000003595 spectral effect Effects 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 2
- 230000008859 change Effects 0.000 description 2
- 238000005056 compaction Methods 0.000 description 2
- 239000000835 fiber Substances 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000002452 interceptive effect Effects 0.000 description 2
- 239000004973 liquid crystal related substance Substances 0.000 description 2
- 239000002096 quantum dot Substances 0.000 description 2
- 210000001525 retina Anatomy 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 238000003491 array Methods 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000003795 chemical substances by application Substances 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012549 training Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/29—Geographical information databases
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S19/00—Satellite radio beacon positioning systems; Determining position, velocity or attitude using signals transmitted by such systems
- G01S19/38—Determining a navigation solution using signals transmitted by a satellite radio beacon positioning system
- G01S19/39—Determining a navigation solution using signals transmitted by a satellite radio beacon positioning system the satellite radio beacon positioning system transmitting time-stamped messages, e.g. GPS [Global Positioning System], GLONASS [Global Orbiting Navigation Satellite System] or GALILEO
- G01S19/42—Determining position
- G01S19/45—Determining position by combining measurements of signals from the satellite radio beacon positioning system with a supplementary measurement
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/40—Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
- G06F16/48—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/487—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using geographical or spatial information, e.g. location
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/05—Geographic models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T9/00—Image coding
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B29/00—Maps; Plans; Charts; Diagrams, e.g. route diagram
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/189—Recording image signals; Reproducing recorded image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/10—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
- H04N19/102—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or selection affected or controlled by the adaptive coding
- H04N19/13—Adaptive entropy coding, e.g. adaptive variable length coding [AVLC] or context adaptive binary arithmetic coding [CABAC]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/90—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using coding techniques not provided for in groups H04N19/10-H04N19/85, e.g. fractals
- H04N19/96—Tree coding, e.g. quad-tree coding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
Definitions
- the present application relates to positioning technology, in particular to a method and device for retrieving 3D maps.
- Virtual reality Virtual Reality
- AR Augmented Reality
- MR Mixed reality
- This type of technology can create virtual reality and superimpose virtual reality and real world to generate new visual environment and interactive experience.
- electronic devices need to determine their own pose information in the current environment, so as to accurately realize the fusion of virtual objects and real scenes.
- a typical solution is: the electronic device receives a three-dimensional (3-dimension, 3D) map of its environment from a server or other devices, and then passes The local sensor collects visual information in the environment, and searches the downloaded 3D map according to the collected visual information to determine the current pose of the electronic device.
- 3D three-dimensional
- the original 3D map usually contains a huge amount of data, and the amount of calculation for retrieval is usually huge, which consumes a lot of computing resources and takes a long time, which affects the user experience.
- Embodiments of the present application provide a 3D map retrieval method and device, which can improve retrieval performance.
- the embodiment of the present application provides a method for retrieving a 3D map
- the method may include: extracting the binary data of S 3D map descriptors from the compressed data of S 3D map descriptors, and the S 3D map
- the descriptor corresponds to multiple 3D map points in the 3D map.
- the i-th level retrieval is performed in the binarized data of the S 3D map descriptors to obtain P 3D map descriptors.
- the retrieval descriptor is a feature corresponding to the real environment extracted from the visual information collected by the sensor of the electronic device.
- the decompression process (full decompression process) of the compressed data of the P 3D map descriptors includes at least the mth level of decompression.
- the i-th level search is performed in the binarized data of the S 3D map descriptors, and P 3D map descriptors are obtained by screening.
- P 3D map The compressed data of the descriptor is decompressed at the mth level to obtain the reconstructed data of the P 3D map descriptors.
- the jth step is performed in the reconstructed data of the P 3D map descriptors Level retrieval to filter out fewer 3D map descriptors.
- the first-level retrieval of at least two-level retrieval uses binary data in the compressed data, which can increase the retrieval speed
- the other-level retrieval uses the reconstructed data obtained by decompression, which can ensure retrieval accuracy.
- the retrieval method for the 3D map provided in the embodiment of the present application can improve the retrieval performance.
- the 3D map retrieval method provided by the embodiment of the present application can improve retrieval performance.
- the above-mentioned m-th level decompression may be staged decompression, which may be understood as a sub-process of complete decompression, or may be understood as partial decompression.
- staged decompression may be understood as a sub-process of complete decompression, or may be understood as partial decompression.
- the above-mentioned mth stage decompression may be full decompression.
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are matched with the 3D map points corresponding to the retrieval descriptors, and N is a positive integer, 0 ⁇ N ⁇ Q, and the above method may further include: performing positioning according to the 3D map points corresponding to the N 3D map descriptors, so as to obtain the pose information of the electronic device.
- positioning can be performed according to the 3D map points corresponding to the N 3D map descriptors, which is beneficial to improve positioning performance.
- the retrieval method adopted by the i-level retrieval is a retrieval method based on the first distance
- the retrieval method adopted by the j-level retrieval is a retrieval method based on the second distance
- the first distance may include a distance obtained by using binarized data, such as a Hamming distance.
- the first distance may be the Hamming distance.
- the first distance may be a distance obtained by taking the absolute value of the difference between corresponding bits of the two binarized data and adding all the absolute values of the difference.
- the retrieval based on the first distance refers to calculating the first distance to determine the degree of correlation or similarity to filter the 3D map descriptors.
- the second distance may include, but is not limited to, Euclidean distance, inner product distance, cosine distance, Manhattan distance, and the like.
- the retrieval based on the second distance refers to calculating the second distance to determine the degree of correlation or similarity to filter the 3D map descriptors.
- performing retrieval based on the first distance can increase retrieval speed, and performing retrieval based on the second distance can improve retrieval accuracy.
- the method may further include: receiving the retrieval descriptor, and performing binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor.
- the execution subject (which may be an electronic device or a server (such as a cloud server)) for executing the retrieval method described in the first aspect may receive retrieval descriptors sent by other devices.
- the execution subject for executing the retrieval method described in the first aspect receives retrieval descriptors collected and extracted by other electronic devices. Afterwards, the retrieval descriptor is binarized to obtain the binarized data of the retrieval descriptor.
- the method may further include: receiving the visual information, extracting a retrieval descriptor from the visual information, and performing binarization processing on the retrieval descriptor to obtain binary data of the retrieval descriptor;
- the execution subject (which may be an electronic device or a server) for executing the retrieval method described in the first aspect may receive visual information sent by other devices, and extract retrieval descriptors therefrom. Afterwards, the retrieval descriptor is binarized to obtain the binarized data of the retrieval descriptor.
- the method may further include: responding to the operation of collecting visual information input by the user, triggering the sensor to collect visual information from the real environment, obtaining the visual information, and obtaining the visual information from the visual information
- the retrieval descriptor is extracted, and the retrieval descriptor is binarized to obtain binarized data of the retrieval descriptor.
- the execution subject (which may be an electronic device or a server) for executing the retrieval method described in the first aspect may collect the visual information and extract the retrieval descriptor from it. Afterwards, the retrieval descriptor is binarized to obtain the binarized data of the retrieval descriptor.
- the 3D map retrieval method in the embodiment of the present application can retrieve retrieval descriptors in different application scenarios in the 3D map, and has wide applicability.
- the method further includes: decompressing the compressed data of the Q 3D map descriptors at the kth level to obtain the reconstructed data of the Q 3D map descriptors.
- the r-th level retrieval is performed in the reconstructed data of the Q 3D map descriptors to obtain N 3D map descriptors.
- the decompression process of the compressed data of P 3D map descriptors includes m-level decompression (stage decompression) and k-th stage decompression (stage decompression), r and k are Positive integer, m ⁇ k, j ⁇ r ⁇ L.
- the i-th level search is performed in the binarized data of the S 3D map descriptors, and P 3D map descriptors are obtained by screening.
- P 3D map descriptors The compressed data is decompressed at the mth level (staged decompression) to obtain the reconstructed data of P 3D map descriptors.
- An r-th level search is performed on the reconstructed data of the map descriptor to obtain N 3D map descriptors, which can be used for positioning.
- At least one of the three levels of retrieval uses binarized data in compressed data, which can improve retrieval speed, and the other two levels of retrieval use reconstructed data obtained by different decompression, which can improve retrieval speed while ensuring retrieval accuracy .
- the degree of decompression or the degree of distortion of the reconstructed data obtained by the m-level decompression and the k-th level of decompression are different.
- the compressed data of the P 3D map descriptors includes the binarized data and quantized data of the P 3D map descriptors, and the m-th decompression is performed on the compressed data of the P 3D map descriptors to obtain
- the reconstructed data of the P 3D map descriptors includes: dequantizing the respective quantized data of the P 3D map descriptors to obtain P dequantized data, and the P dequantized data are used as the P 3D map descriptors Restructure data.
- Decompressing the compressed data of the Q 3D map descriptors at the kth level to obtain the reconstructed data of the Q 3D map descriptors including: dequantizing the respective quantized data of the Q 3D map descriptors to obtain Q
- the reconstructed data of the Q 3D map descriptors are obtained according to the Q dequantized data and the binarized data of the Q 3D map descriptors.
- the respective dequantized data of the P 3D map descriptors are obtained, based on the respective dequantized data of the P 3D map descriptors (used in this Reconstructed data retrieved at the first level) to perform j-level retrieval, and obtain Q 3D map descriptors by screening, and dequantize the respective quantized data of the Q 3D map descriptors to obtain Q dequantized data, according to the Q
- the dequantization data and the binarized data of each of the Q 3D map descriptors are obtained to obtain the respective reconstruction data of the Q 3D map descriptors, and the r-level retrieval is performed based on the respective reconstruction data of the Q 3D map descriptors to obtain N 3D map descriptors.
- the distortion degree of the reconstructed data retrieved at the j-level is greater than that of the reconstructed data retrieved at the r-level, that is, the rough retrieval is performed first to improve the retrieval speed, and then the fine retrieval is performed to improve the retrieval accuracy.
- the compressed data of the P 3D map descriptors includes the respective binarized data and quantized data of the P 3D map descriptors, and the compression of the P 3D map descriptors
- the data is decompressed at the mth level to obtain the reconstructed data of P 3D map descriptors, including: performing dequantization processing on the respective quantized data of the P 3D map descriptors to obtain P dequantized data, and according to the P dequantized data data and the respective binarized data of the P 3D map descriptors to obtain the respective reconstruction data of the P 3D map descriptors.
- the jth level retrieval may be the last level of the multi-level retrieval, and the last level is retrieved based on the reconstructed data obtained through complete decompression, which can improve the accuracy of the retrieval.
- the S 3D map descriptors are S representative 3D map descriptors
- the S representative 3D map descriptors respectively correspond to at least one data set
- each data set of at least one data set includes at least one 3D map Descriptor, according to the binary data of the retrieval descriptor, perform the i-th level retrieval in the binary data of the S 3D map descriptors to obtain P 3D map descriptors, including: according to the binary data of the retrieval descriptor
- the i-th level search is performed in the binarized data representing S 3D map descriptors to obtain at least one representative 3D map descriptor, and at least one representative 3D map descriptor is used to represent the 3D map in the data set corresponding to each Descriptor, as P 3D map descriptors.
- the 3D map descriptors in each data set have certain correlation or similarity.
- the representative 3D map descriptor is used to characterize common or similar features of the 3D map descriptors in the corresponding data set.
- P may be greater than S. P ⁇ T, S ⁇ T.
- one-level retrieval is performed on a small number of representative 3D map descriptors to improve the retrieval speed, and then one-level or multi-level retrieval can be performed in the data set corresponding to the representative 3D map descriptor to improve retrieval accuracy Rate.
- the retrieval speed can be improved and the retrieval accuracy can be guaranteed.
- Q ⁇ P ⁇ S ⁇ T That is, the number of 3D map descriptors in multi-level retrieval decreases step by step.
- the method may further include: according to each component of the retrieval descriptor and the corresponding component of the preset threshold vector, determining the size relationship between each component of the retrieval descriptor and the corresponding component of the preset threshold vector. Binarize the size relationship to obtain the binarized data of the retrieval descriptor.
- each component of the retrieval descriptor and the corresponding component of the preset threshold vector determine the size relationship between each component of the retrieval descriptor and the corresponding component of the preset threshold vector, including: Each component of the sub is subtracted from the corresponding component of the preset threshold vector to obtain the difference of each component; according to the difference of each component, the sign of each component is determined.
- Performing binary processing on the size relationship to obtain binary data of the retrieval descriptor includes: performing binary processing on each component symbol to obtain binary data of the retrieval descriptor.
- the embodiment of the present application provides a method for retrieving a 3D map.
- the method may include: extracting the first binarized data of the S 3D map descriptors from the compressed data of the S 3D map descriptors, and the S The 3D map descriptor corresponds to multiple 3D map points in the 3D map.
- the i-th level search is performed on the first binarized data of the S 3D map descriptors to obtain P 3D map descriptors.
- the retrieval descriptor is a feature corresponding to the real environment extracted from the visual information collected by the sensor of the electronic device.
- the j-th level retrieval is performed in the second binarization data of P 3D map descriptors to obtain Q 3D map descriptors
- S, P, Q, i and j is a positive integer
- T represents the total number of 3D map descriptors in the 3D map
- j i+1, 1 ⁇ i ⁇ L, 1 ⁇ j ⁇ L
- L represents the total number of retrieval levels of the 3D map or a threshold value of the retrieval level
- L is a positive integer greater than 1.
- the first binarized data of S 3D map descriptors are extracted from the compressed data of S 3D map descriptors, and according to the first binarized data of the retrieved descriptors, the S 3D map
- the i-level search is performed on the first binarized data of the descriptor, and P 3D map descriptors are obtained by screening. Extract the second binarized data of P 3D map descriptors from the compressed data of P 3D map descriptors, according to the second binarized data of the retrieved descriptors, in the second binary data of P 3D map descriptors
- the j-level retrieval is performed in the 2D data to filter out a smaller number of 3D map descriptors.
- the retrieval speed can be improved and the retrieval accuracy can be guaranteed.
- the first binarized data and the second binarized data may be binarized data used in any two levels in the multi-level search, and their binarized processing methods may be different, or their lengths may be different.
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors match the 3D map points corresponding to the retrieval descriptors, and N is positive Integer, 0 ⁇ N ⁇ Q, the above method further includes: performing positioning according to the 3D map points corresponding to the N 3D map descriptors, so as to obtain the pose information of the electronic device.
- the i-th level retrieval and the j-th level retrieval adopt the retrieval method based on the first distance
- the P 3D map descriptors belong to the S 3D map descriptors
- the P 3D map The position of the first binarization data of each 3D map descriptor of the descriptor in the compressed data of the 3D map descriptor, which is different from the position of the second binarization data of the 3D map descriptor in the description of the 3D map position in the compressed data of the child, where P ⁇ S.
- this position may be a start position, or an end position, etc.
- the length of the first binarized data of each 3D map descriptor of the P 3D map descriptors is smaller than the length of the second binary data of each 3D map descriptor of the P 3D map descriptors.
- the length of the valued data is smaller than the length of the first binarized data of each 3D map descriptor of the P 3D map descriptors.
- the length of the first binarized data of each 3D map descriptor in the i-level retrieval is relatively short, which can improve the retrieval speed
- the length of the second binarized data of each 3D map descriptor in the j-level retrieval Longer can improve retrieval accuracy.
- the above method further includes: receiving the retrieval descriptor, and performing binarization processing on the retrieval descriptor to obtain the first binarized data and the second binary data of the retrieval descriptor. Value data.
- the execution subject (which may be an electronic device or a server (such as a cloud server)) for executing the retrieval method described in the second aspect may receive retrieval descriptors sent by other devices.
- the execution subject for executing the retrieval method described in the second aspect receives retrieval descriptors collected and extracted by other electronic devices. Then perform binarization processing on the retrieval descriptor to obtain first binarized data and second binarized data of the retrieval descriptor.
- the above method further includes: receiving the visual information, extracting the retrieval descriptor from the visual information, and performing binarization processing on the retrieval descriptor to obtain the retrieval description The first binarized data and the second binarized data of the child.
- the execution subject (which may be an electronic device or a server) for executing the retrieval method described in the second aspect may receive visual information sent by other devices, and extract retrieval descriptors therefrom. Then perform binarization processing on the retrieval descriptor to obtain first binarized data and second binarized data of the retrieval descriptor.
- the above method further includes: responding to the operation of collecting visual information input by the user, triggering the sensor to collect visual information from the real environment, obtaining the visual information, and extracting from the visual information
- the search descriptor and performing binarization processing on the search descriptor to obtain first binarized data and second binarized data of the search descriptor.
- the execution subject (which may be an electronic device or a server) for executing the retrieval method described in the second aspect may collect the visual information and extract retrieval descriptors therefrom. Then perform binarization processing on the retrieval descriptor to obtain first binarized data and second binarized data of the retrieval descriptor.
- the length of the first binarized data of the retrieval descriptor is equal to the length of the first binarized data of each of the S 3D map descriptors, and/or, The length of the second binarized data of the retrieval descriptor is equal to the length of the second binarized data of each 3D map descriptor of the S 3D map descriptors.
- the search speed can be improved.
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data of the at least one data set
- the set includes at least one 3D map descriptor, and according to the first binarized data of the search descriptor, the ith-level search is performed on the first binarized data of the S 3D map descriptors to obtain P
- the 3D map descriptor including: according to the first binarized data of the search descriptor, perform i-level retrieval in the first binarized data of the S representative 3D map descriptors, so as to obtain at least one representative 3D map descriptors: the at least one representative 3D map descriptor in the data set corresponding to each of the 3D map descriptors is used as the P 3D map descriptors.
- Q ⁇ P ⁇ S ⁇ T That is, the number of 3D map descriptors in multi-level retrieval decreases step by step.
- the embodiment of the present application provides a method for retrieving a 3D map.
- the method may include: decompressing the compressed data of the S 3D map descriptors at the mth level to obtain the first Reconstructing data, the S 3D map descriptors correspond to multiple 3D map points in the 3D map.
- an i-th level retrieval is performed on the first reconstructed data of the S 3D map descriptors to obtain P 3D map descriptors.
- the retrieval descriptor is a feature corresponding to the real environment extracted from the visual information collected by the sensor of the electronic device.
- the decompression process of the compressed data of the S 3D map descriptors includes the k-level decompression, and the k-level decompression Includes m-th stage decompression.
- the compressed data of the S 3D map descriptors are decompressed at the mth stage to obtain the first reconstructed data of the S 3D map descriptors.
- An i-level search is performed on the first reconstructed data of the S 3D map descriptors, and P 3D map descriptors are screened out.
- Decompress the compressed data of the P 3D map descriptors at the kth level to obtain the second reconstructed data of the P 3D map descriptors.
- the j-th level retrieval is performed in the second reconstructed data to filter out a smaller number of 3D map descriptors.
- one level of retrieval is staged decompression, which can improve the retrieval speed, and the decompression differential settings of at least two levels of retrieval can ensure retrieval accuracy.
- the degree of decompression or the degree of distortion of the reconstructed data of the 3D map descriptor used by any two levels of retrieval are different. Compared with retrieving in the reconstructed data of the fully decompressed 3D map, the retrieval speed can be improved and the retrieval accuracy can be guaranteed.
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors match the 3D map points corresponding to the retrieval descriptors , N is a positive integer, 0 ⁇ N ⁇ Q, the above method further includes: performing positioning according to the 3D map points corresponding to the N 3D map descriptors, so as to obtain the pose information of the electronic device.
- positioning can be performed according to the 3D map points corresponding to the N 3D map descriptors, which is beneficial to improve positioning performance.
- the i-th level retrieval and the j-th level retrieval adopt a retrieval method based on the second distance
- the P 3D map descriptors belong to the S 3D map Descriptor
- the distortion degree of the first reconstructed data of each 3D map descriptor of the P 3D map descriptors, and the distortion of the second reconstructed data of each 3D map descriptor of the P 3D map descriptors degree is different
- the distortion degree of the first reconstructed data of each 3D map descriptor of the P 3D map descriptors is used to represent the first reconstructed data of each 3D map descriptor and the corresponding original 3D map descriptor
- the degree of difference between the second reconstruction data of each 3D map descriptor of the P 3D map descriptors is used to represent the second reconstruction data of each 3D map descriptor and the corresponding original 3D
- the distortion degree of the first reconstruction data of each 3D map descriptor of the P 3D map descriptors is greater than the second reconstruction data of each 3D map descriptor of the P 3D map descriptors Distortion of structured data.
- the degree of distortion of the first reconstructed data of each 3D map descriptor in the i-th level retrieval may be greater than the degree of distortion of the second reconstructed data of each 3D map descriptor among the P 3D map descriptors.
- the distortion degree of the previous retrieval is greater than that of the subsequent retrieval, so as to perform fast retrieval in the previous retrieval and improve the retrieval speed, and perform fine retrieval in the latter retrieval to improve retrieval accuracy Rate.
- the above method further includes: receiving the retrieval descriptor, and acquiring part or all of the data of the retrieval descriptor.
- the execution subject (which may be an electronic device or a server (such as a cloud server)) for executing the retrieval method described in the third aspect may receive retrieval descriptors sent by other devices.
- the execution subject for executing the retrieval method described in the third aspect receives retrieval descriptors collected and extracted by other electronic devices. Afterwards, part or all of the data of the retrieval descriptor is obtained according to retrieval requirements.
- the above method further includes: receiving the visual information, extracting the retrieval descriptor from the visual information, and acquiring part or all of the data of the retrieval descriptor.
- the execution subject (which may be an electronic device or a server) for executing the retrieval method described in the third aspect may receive visual information sent by other devices, and extract retrieval descriptors therefrom. Afterwards, part or all of the data of the retrieval descriptor is obtained according to retrieval requirements.
- the above method further includes: responding to the operation of collecting visual information input by the user, triggering the sensor to collect visual information from the real environment, obtaining the visual information, and extracting from the visual information
- the search descriptor and obtain part or all of the data of the search descriptor.
- the execution subject (which may be an electronic device or a server) for executing the retrieval method described in the third aspect may collect the visual information and extract the retrieval descriptor from it. Afterwards, part or all of the data of the retrieval descriptor is obtained according to retrieval requirements.
- the 3D map retrieval method in the embodiment of the present application can retrieve retrieval descriptors in different application scenarios in the 3D map, and has wide applicability.
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data of the at least one data set
- the collection includes at least one 3D map descriptor, and according to the part or all data of the retrieval descriptor, the i-th level retrieval is performed in the first reconstructed data of the S 3D map descriptors to obtain P 3D maps Descriptors, including: according to part or all of the data of the retrieval descriptors, perform the i-th level retrieval in the first reconstructed data representing the S representative 3D map descriptors, so as to obtain at least one representative 3D map descriptors ; Using the at least one representative 3D map descriptor in the data set corresponding to each of the 3D map descriptors as the P 3D map descriptors.
- the 3D map descriptors in each data set have certain correlation or similarity.
- the representative 3D map descriptor is used to characterize common or similar features of the 3D map descriptors in the corresponding data set.
- P may be greater than S. P ⁇ T, S ⁇ T.
- one-level retrieval is performed on a small number of representative 3D map descriptors to improve the retrieval speed, and then one-level or multi-level retrieval can be performed in the data set corresponding to the representative 3D map descriptor to improve retrieval accuracy Rate.
- the retrieval speed can be improved and the retrieval accuracy can be guaranteed.
- Q ⁇ P ⁇ S ⁇ T That is, the number of 3D map descriptors in multi-level retrieval decreases step by step.
- the embodiment of the present application provides a method for retrieving a 3D map
- the method may include: decompressing the compressed data of S 3D map descriptors at the mth level to obtain reconstructed data of S 3D map descriptors , the S 3D map descriptors correspond to multiple 3D map points in the 3D map.
- the i-th level retrieval is performed in the reconstructed data of the S 3D map descriptors to obtain P 3D map descriptors; the retrieval descriptor is obtained from the electronic device
- the features corresponding to the real environment are extracted from the visual information collected by the sensor.
- At least one level of retrieval is decompression in at least two levels of retrieval, which can improve retrieval accuracy
- at least one level of retrieval in at least two levels of retrieval is extraction from compressed data, which can improve retrieval speed.
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are the 3D map points corresponding to the retrieval descriptors.
- Point matching N is a positive integer, 0 ⁇ N ⁇ Q
- the method further includes: performing positioning according to the 3D map points corresponding to the N 3D map descriptors, so as to obtain the pose information of the electronic device.
- the retrieval method adopted by the i-th level retrieval is a retrieval method based on the second distance
- the retrieval method adopted by the j-th level retrieval is a retrieval method based on the first distance
- the retrieval method based on the second distance can improve the retrieval accuracy
- the retrieval method based on the first distance can increase the retrieval speed, thereby achieving a comprehensive improvement of retrieval accuracy and retrieval speed.
- the above method further includes: receiving the retrieval descriptor, and performing binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor;
- the execution subject of the retrieval method described in the aspect may receive retrieval descriptors sent by other devices.
- the execution subject for executing the retrieval method described in the first aspect receives retrieval descriptors collected and extracted by other electronic devices. Afterwards, the retrieval descriptor is binarized to obtain the binarized data of the retrieval descriptor.
- the above method further includes: receiving the visual information, extracting the retrieval descriptor from the visual information, and performing binarization processing on the retrieval descriptor to obtain the retrieval description binarized data;
- the executive body (which can be an electronic device or a server) used to execute the retrieval method described in the fourth aspect can receive visual information sent by other devices, and extract retrieval descriptors therefrom. Afterwards, the retrieval descriptor is binarized to obtain the binarized data of the retrieval descriptor.
- the above method further includes: responding to the operation of collecting visual information input by the user, triggering the sensor to collect visual information from the real environment, obtaining the visual information, and extracting from the visual information
- the retrieval descriptor and performing binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor.
- the execution subject (which may be an electronic device or a server) for executing the retrieval method described in the fourth aspect may receive visual information sent by other devices, and extract retrieval descriptors therefrom. Afterwards, the retrieval descriptor is binarized to obtain the binarized data of the retrieval descriptor.
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data of the at least one data set
- the collection includes at least one 3D map descriptor, and according to the part or all data of the search descriptor, the i-th level retrieval is performed in the reconstructed data of the S 3D map descriptors to obtain P 3D map descriptors , comprising: according to part or all of the data of the search descriptor, performing an i-th level search in the reconstructed data representing the S representative 3D map descriptors to obtain at least one representative 3D map descriptor; At least one representative 3D map descriptor in the data set corresponding to each of the 3D map descriptors is used as the P 3D map descriptors.
- the 3D map descriptors in each data set have certain correlation or similarity.
- the representative 3D map descriptor is used to characterize common or similar features of the 3D map descriptors in the corresponding data set.
- P may be greater than S. P ⁇ T, S ⁇ T.
- one-level retrieval is performed on a small number of representative 3D map descriptors to improve the retrieval speed, and then one-level or multi-level retrieval can be performed in the data set corresponding to the representative 3D map descriptor to improve retrieval accuracy Rate.
- the retrieval speed can be improved and the retrieval accuracy can be guaranteed.
- Q ⁇ P ⁇ S ⁇ T That is, the number of 3D map descriptors in multi-level retrieval decreases step by step.
- the retrieval descriptor includes an area descriptor or a 3D map point descriptor.
- the S 3D map descriptors may include S area descriptors or S 3D map point descriptors.
- the embodiment of the present application provides a device for retrieving a 3D map.
- the device may be a chip or a system-on-a-chip in an electronic device or a server, and may also be used in an electronic device or a server to implement the first aspect or the first aspect.
- the retrieval device of the 3D map includes: a retrieval module, an extraction module and a decompression module.
- the extraction module is used to extract the binarized data of S 3D map descriptors from the compressed data of S 3D map descriptors, and the S 3D map descriptors correspond to multiple A 3D map point; a retrieval module, configured to perform an i-th level retrieval in the binarized data of the S 3D map descriptors according to the binarized data of the retrieval descriptors, to obtain P 3D map descriptors;
- the retrieval descriptor is a feature corresponding to the real environment extracted from the visual information collected by the sensor of the electronic device;
- the decompression module is used to perform m-level decompression on the compressed data of the P 3D map descriptors, Obtaining the reconstruction data of the P 3D map descriptors; the decompression process of the compressed data of the P 3D map descriptors at least includes the decompression at the mth level;
- the retrieval module is configured to Some or all of the data of the descriptor is retrieved at the
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are the 3D map points corresponding to the retrieval descriptors.
- Point matching N is a positive integer, 0 ⁇ N ⁇ Q
- the device also includes: a pose determination module; the pose determination module performs positioning according to the 3D map points corresponding to the N 3D map descriptors, to The pose information of the electronic device is obtained.
- the retrieval method adopted by the i-th level retrieval is a retrieval method based on the first distance
- the retrieval method adopted by the j-th level retrieval is a retrieval method based on the second distance.
- the device further includes: an acquisition module, configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain a binarization of the retrieval descriptor data; or, receiving the visual information, extracting the retrieval descriptor from the visual information, and performing binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor; or , in response to the operation of collecting visual information input by the user, triggering the sensor to collect visual information on the real environment to obtain the visual information, extracting the retrieval descriptor from the visual information, and describing the retrieval perform binarization processing to obtain the binarized data of the retrieval descriptor.
- an acquisition module configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain a binarization of the retrieval descriptor data
- receiving the visual information extracting the retrieval descriptor from the visual information, and performing binarization processing on the retrieval descriptor to obtain binarized data of the
- the decompression module is further configured to: decompress the compressed data of the Q 3D map descriptors at the kth level to obtain Q 3D map descriptions Reconstructed data of the child; according to part or all of the data of the search descriptor, perform r-th level retrieval in the reconstructed data of the Q 3D map descriptors to obtain the N 3D map descriptors;
- N ⁇ Q ⁇ P the decompression process of the compressed data of the P 3D map descriptors includes the m-level decompression and the k-level decompression, r and k are positive integers, m ⁇ k , j ⁇ r ⁇ L.
- the compressed data of the P 3D map descriptors includes the respective binarized data and quantized data of the P 3D map descriptors
- the decompression module is specifically configured to: Dequantize the respective quantized data of the descriptors to obtain P dequantized data, and the P dequantized data are used as reconstruction data of the P 3D map descriptors; each of the Q 3D map descriptors Quantized data of the quantized data is dequantized, and Q dequantized data are obtained; according to the Q dequantized data and the binarized data of the Q 3D map descriptors, the Q 3D map descriptors are respectively obtained. of reconstructed data.
- the compressed data of the P 3D map descriptors includes the respective binarized data and quantized data of the P 3D map descriptors, and the decompression module, specifically It is used for: performing dequantization processing on the respective quantized data of the P 3D map descriptors to obtain P dequantized data; according to the respective binary values of the P dequantized data and the P 3D map descriptors to obtain the respective reconstruction data of the P 3D map descriptors.
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data of the at least one data set
- the collection includes at least one 3D map descriptor
- the retrieval module is used for: performing i-level retrieval in the S binarized data representing 3D map descriptors according to the binarized data of the retrieval descriptor, to obtain at least one representative 3D map descriptor; using the 3D map descriptors in the data sets corresponding to the at least one representative 3D map descriptor as the P 3D map descriptors.
- the embodiment of the present application provides a device for retrieving a 3D map.
- the device may be a chip or a system-on-a-chip in an electronic device or a server, and may also be used in an electronic device or a server to implement the first aspect or the first aspect.
- the retrieval device of the 3D map includes: a retrieval module, an extraction module and a decompression module.
- the extraction module is configured to extract the first binarized data of the S 3D map descriptors from the compressed data of the S 3D map descriptors, and the S 3D map descriptors correspond to 3D A plurality of 3D map points in the map; a retrieval module, configured to perform i-th level retrieval in the first binarized data of the S 3D map descriptors according to the first binarized data of the retrieved descriptors, to Obtain P 3D map descriptors; the retrieval descriptor is a feature corresponding to the real environment extracted from the visual information collected by the sensor of the electronic device; the extraction module is also used to obtain the P 3D map descriptors Extract the second binarized data of the P 3D map descriptors from the compressed data; the retrieval module is used to describe the P 3D maps according to the second binarized data of the retrieved descriptors In order to obtain Q 3D map descriptors, S, P, Q, i and j are examples of the first bin
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are the 3D map points corresponding to the retrieval descriptors.
- Point matching N is a positive integer, 0 ⁇ N ⁇ Q
- the device also includes: a pose determination module; the pose determination module is used to perform positioning according to the 3D map points corresponding to the N 3D map descriptors , to obtain the pose information of the electronic device.
- the i-th level retrieval and the j-th level retrieval adopt retrieval methods based on the first distance, and the P 3D map descriptors belong to the S 3D map descriptions
- the position of the first binarized data of each 3D map descriptor of the P 3D map descriptors in the compressed data of the 3D map descriptor is different from the second binary value of the 3D map descriptor
- the length of the first binarized data of each 3D map descriptor of the P 3D map descriptors is smaller than the length of the second binary data of each 3D map descriptor of the P 3D map descriptors.
- the length of the valued data is smaller than the length of the first binarized data of each 3D map descriptor of the P 3D map descriptors.
- the device further includes: an acquisition module, configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain a first binary value of the retrieval descriptor. valued data and second binarized data; or, receive the visual information, extract the retrieval descriptor from the visual information, and perform binarization processing on the retrieval descriptor to obtain the retrieval The first binarized data and the second binarized data of the descriptor; or, in response to the operation of collecting visual information input by the user, triggering the sensor to collect visual information on the real environment to obtain the visual information, The retrieval descriptor is extracted from the visual information, and binarization is performed on the retrieval descriptor to obtain first binarized data and second binarized data of the retrieval descriptor.
- an acquisition module configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain a first binary value of the retrieval descriptor. valued data and second binarized data; or, receive the visual information,
- the length of the first binarized data of the retrieval descriptor is equal to the length of the first binarized data of each of the S 3D map descriptors, and/or, The length of the second binarized data of the retrieval descriptor is equal to the length of the second binarized data of each 3D map descriptor of the S 3D map descriptors.
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data of the at least one data set
- the set includes at least one 3D map descriptor
- the retrieval module is specifically configured to: perform a search on the S first binarized data representing 3D map descriptors according to the first binarized data of the retrieved descriptor. Searching at the i-level to obtain at least one representative 3D map descriptor; using the 3D map descriptors in the data sets corresponding to the at least one representative 3D map descriptor as the P 3D map descriptors.
- the embodiment of the present application provides a device for retrieving a 3D map.
- the device may be a chip or a system-on-a-chip in an electronic device or a server, and may also be used in an electronic device or a server to implement the first aspect or the first aspect.
- the retrieval device of the 3D map includes: a retrieval module, an extraction module and a decompression module.
- the decompression module is used to decompress the compressed data of the S 3D map descriptors at the mth level to obtain the first reconstructed data of the S 3D map descriptors, and the S 3D map descriptors
- the map descriptor corresponds to multiple 3D map points in the 3D map
- the retrieval module is used to perform the i-th level in the first reconstruction data of the S 3D map descriptors according to part or all of the data of the retrieval descriptor.
- the retrieval descriptor is extracted from the visual information collected by the sensor of the electronic device and corresponds to the feature of the real environment; the decompression module is also used for the P
- the compressed data of the 3D map descriptors is decompressed at the kth level to obtain the second reconstructed data of the P 3D map descriptors, and the decompression process of the compressed data of the S 3D map descriptors includes the mth stage decompression and the kth stage decompression, or, the decompression process of the compressed data of the S 3D map descriptors includes the kth stage decompression, and the kth stage decompression includes the mth stage decompression Compression; a retrieval module, configured to perform j-th level retrieval in the second reconstructed data of the P 3D map descriptors according to part or all of the retrieval descriptors, so as to obtain Q 3D map descriptors, S , P, Q, i
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are the 3D map points corresponding to the retrieval descriptors.
- Point matching N is a positive integer, 0 ⁇ N ⁇ Q
- the device also includes: a pose determination module; the pose determination module is used to perform positioning according to the 3D map points corresponding to the N 3D map descriptors , to obtain the pose information of the electronic device.
- the i-th level retrieval and the j-th level retrieval adopt a retrieval method based on the second distance
- the P 3D map descriptors belong to the S 3D map Descriptor
- the distortion degree of the first reconstructed data of each 3D map descriptor of the P 3D map descriptors, and the distortion of the second reconstructed data of each 3D map descriptor of the P 3D map descriptors degree is different
- the distortion degree of the first reconstructed data of each 3D map descriptor of the P 3D map descriptors is used to represent the first reconstructed data of each 3D map descriptor and the corresponding original 3D map descriptor
- the degree of difference between the second reconstruction data of each 3D map descriptor of the P 3D map descriptors is used to represent the second reconstruction data of each 3D map descriptor and the corresponding original 3D
- the distortion degree of the first reconstruction data of each 3D map descriptor of the P 3D map descriptors is greater than the second reconstruction data of each 3D map descriptor of the P 3D map descriptors Distortion of structured data.
- the device further includes: an obtaining module, configured to: receive the retrieval descriptor, and obtain part or all of the data of the retrieval descriptor; or receive the visual information, and obtain the Extract the retrieval descriptor from the visual information, and obtain part or all of the data of the retrieval descriptor; or, in response to the operation of collecting visual information input by the user, trigger the sensor to perform visual information on the real environment Collecting, obtaining the visual information, extracting the retrieval descriptor from the visual information, and obtaining part or all of the data of the retrieval descriptor.
- an obtaining module configured to: receive the retrieval descriptor, and obtain part or all of the data of the retrieval descriptor; or receive the visual information, and obtain the Extract the retrieval descriptor from the visual information, and obtain part or all of the data of the retrieval descriptor; or, in response to the operation of collecting visual information input by the user, trigger the sensor to perform visual information on the real environment Collecting, obtaining the visual information, extracting the retrieval de
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data of the at least one data set
- the collection includes at least one 3D map descriptor
- the retrieval module is specifically configured to: perform the i-th reconstruction data in the m-th reconstructed data of the S representative 3D map descriptors according to part or all of the data of the retrieval descriptor. Level retrieval to obtain at least one representative 3D map descriptor; using the 3D map descriptors in the data sets corresponding to the at least one representative 3D map descriptor as the P 3D map descriptors.
- the embodiment of the present application provides a device for retrieving a 3D map.
- the device may be a chip or a system-on-a-chip in an electronic device or a server, and may also be used in an electronic device or a server to implement the first aspect or the first aspect.
- the retrieval device of the 3D map includes: a retrieval module, an extraction module and a decompression module.
- the decompression module is used to decompress the compressed data of the S 3D map descriptors at the mth level to obtain the reconstructed data of the S 3D map descriptors, and the S 3D map descriptors Corresponding to multiple 3D map points in the 3D map;
- the retrieval module is used to perform i-level retrieval in the reconstructed data of the S 3D map descriptors according to the partial data or all data of the retrieval descriptors, so as to obtain P A 3D map descriptor;
- the retrieval descriptor is a feature corresponding to the real environment extracted from the visual information collected by the sensor of the electronic device;
- the extraction module is used to extract from the compressed data of the P 3D map descriptors The binarized data of the P descriptors;
- the retrieval module is further configured to, according to the binarized data of the retrieved descriptors, perform j-th Level retrieval to get Q 3D map descriptors, S, P, Q
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are the 3D map points corresponding to the retrieval descriptors.
- Point matching N is a positive integer, 0 ⁇ N ⁇ Q
- the device also includes: a pose determination module; the pose determination module is used to perform positioning according to the 3D map points corresponding to the N 3D map descriptors , to obtain the pose information of the electronic device.
- the retrieval method adopted by the i-th level retrieval is a retrieval method based on the second distance
- the retrieval method adopted by the j-th level retrieval is a retrieval method based on the first distance
- the device further includes: an acquisition module, configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain a binarization of the retrieval descriptor data; or, receiving the visual information, extracting the retrieval descriptor from the visual information, and performing binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor; or , in response to the operation of collecting visual information input by the user, triggering the sensor to collect visual information on the real environment, obtaining the visual information, extracting the retrieval descriptor from the visual information, and performing the retrieval
- the descriptor is subjected to binarization processing to obtain binarized data of the retrieval descriptor.
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data of the at least one data set
- the collection includes at least one 3D map descriptor
- the retrieval module is specifically configured to: perform the i-th reconstruction data in the m-th reconstructed data of the S representative 3D map descriptors according to part or all of the data of the retrieval descriptor. Level retrieval to obtain at least one representative 3D map descriptor; using the 3D map descriptors in the data sets corresponding to the at least one representative 3D map descriptor as the P 3D map descriptors.
- the embodiment of the present application provides a device for retrieving a 3D map, including: one or more processors; a memory for storing one or more programs; when the one or more programs are used by the one or more A plurality of processors are executed, so that the one or more processors implement the method described in any one of the first to fourth aspects or any possible design of any one of the first to fourth aspects.
- an embodiment of the present application provides a computer-readable storage medium, which is characterized by comprising a computer program, and when the computer program is executed on a computer, the computer executes any one of the first to fourth aspects or The method described in any possible design of any one of the first to fourth aspects.
- the embodiment of the present application provides a computer program or computer program product, when the computer program or computer program product is executed on the computer, the computer can realize any one of the first to fourth aspects and any one of them The method described in one possible implementation.
- FIG. 1 is a schematic diagram of an application architecture provided by an embodiment of the present application
- FIG. 2 is a schematic structural diagram of an electronic device 20 provided in an embodiment of the present application.
- FIG. 3 is a schematic structural diagram of a server 30 provided in an embodiment of the present application.
- FIG. 4g is a schematic diagram of a user interface (such as a user interface of a 3D map application) displayed by an electronic device provided in an embodiment of the present application;
- FIG. 5A to FIG. 5D are schematic flowcharts of fragments of a 3D map retrieval method provided by an embodiment of the present application.
- FIG. 6 is a schematic flowchart of a 3D map retrieval method provided by an embodiment of the present application.
- FIG. 7 is a schematic flowchart of a 3D map retrieval method provided by an embodiment of the present application.
- FIG. 8 is a schematic flowchart of a 3D map retrieval method provided by an embodiment of the present application.
- Fig. 9 is a schematic diagram of the processing process of a 3D map retrieval method provided by the embodiment of the present application
- FIG. 10 is a schematic structural diagram of a 3D map retrieval device provided by an embodiment of the present application.
- FIG. 11 is a schematic block diagram of a decoding device 1100 provided by an embodiment of the present application.
- At least one (item) means one or more, and “multiple” means two or more.
- “And/or” is used to describe the association relationship of associated objects, indicating that there can be three types of relationships, for example, “A and/or B” can mean: only A exists, only B exists, and A and B exist at the same time , where A and B can be singular or plural.
- the character “/” generally indicates that the contextual objects are an “or” relationship.
- At least one of the following” or similar expressions refer to any combination of these items, including any combination of single or plural items.
- At least one item (piece) of a, b or c can mean: a, b, c, "a and b", “a and c", “b and c", or "a and b and c ", where a, b, c can be single or multiple.
- words such as “exemplary” or “for example” are used as examples, illustrations or illustrations. Any embodiment or design scheme described as “exemplary” or “for example” in the embodiments of the present application shall not be interpreted as being more preferred or more advantageous than other embodiments or design schemes. Rather, the use of words such as “exemplary” or “such as” is intended to present related concepts in a concrete manner.
- Fig. 1 is a schematic diagram of an application architecture provided by an embodiment of the present application.
- the application architecture includes a plurality of electronic devices and servers, wherein the plurality of electronic devices may include a first electronic device, one or more A second electronic device (two second electronic devices are taken as an example in FIG. 1), the one or more second electronic devices are several electronic devices other than the first electronic device. Multiple electronic devices and servers, and multiple electronic devices can communicate with each other.
- wireless-fidelity Wireless-fidelity
- WiFi wireless-fidelity
- Bluetooth Bluetooth
- cellular 2/3/4/5 generation 2/3/4/5generation, 2G/3G/4G/5G
- future communication methods which are not specifically limited.
- one or more second electronic devices in the embodiment of the present application is only used to refer to other electronic devices except the first electronic device, but does not limit whether the types of multiple electronic devices are the same.
- the above-mentioned electronic equipment can be various types of equipment equipped with cameras and display components.
- the electronic equipment can be terminal equipment such as mobile phones, tablet computers, notebook computers, video recorders (in FIG. 1, the electronic equipment is a mobile phone as an example),
- Electronic devices can also be devices used for virtual scene interaction, including VR glasses, AR devices, MR interactive devices, etc.
- Electronic devices can also be wearable electronic devices such as smart watches and smart bracelets.
- Electronic devices can also be vehicles, wireless Equipment carried in vehicles such as human-driven vehicles, drones, and industrial robots. The embodiment of the present application does not specifically limit the specific form of the electronic device.
- the above-mentioned electronic devices may also be referred to as user equipment (user equipment, UE), subscriber station, mobile unit, subscriber unit, wireless unit, remote unit, mobile device, wireless device, wireless communication device, remote device, mobile subscriber station , terminal device, access terminal, mobile terminal, wireless terminal, smart terminal, remote terminal, handset, user agent, mobile client, client, or some other suitable term.
- user equipment user equipment
- UE user equipment
- subscriber station mobile unit
- subscriber unit wireless unit
- remote unit mobile device
- wireless device wireless device
- wireless communication device remote device
- mobile subscriber station terminal device, access terminal, mobile terminal, wireless terminal, smart terminal, remote terminal, handset, user agent, mobile client, client, or some other suitable term.
- the above-mentioned servers may be one or more physical servers (a physical server is taken as an example in FIG. 1 ), may also be a computer cluster, or may be a virtual machine or a cloud server in a cloud computing scenario, and the like.
- the electronic device can install virtual scene applications (applications, APPs) such as VR applications, AR applications, or MR applications, and can run VR applications based on user operations (such as clicking, touching, sliding, shaking, voice control, etc.).
- application, AR application or MR application can collect the visual information of any object in the environment through the sensor, and then display the virtual object on the display component according to the collected visual information.
- the virtual object can be a virtual object in a VR scene, an AR scene or an MR scene (that is, a objects).
- the electronic device can install navigation, detection, and control application programs, and run corresponding applications based on user manipulation or preset programs.
- Electronic devices can perform applications such as path planning, object detection, and vehicle control based on their own pose and other state information in the current environment.
- the pose is the position and orientation information of the electronic device, which can be an absolute pose in the world coordinate system or a relative pose relative to a certain point in the environment.
- the visual information involved in the embodiments of the present application includes, but is not limited to, images and videos collected by cameras (without depth information), images and videos with depth information collected by depth sensors, and laser radar (LiDAR) acquisitions.
- the virtual scene application program in the electronic device may be an application program built in the electronic device itself, or an application program provided by a third-party service provider installed by the user himself, and no specific details are given here. limited.
- the electronic device can also be equipped with a simultaneous localization and mapping (SLAM) system.
- the SLAM system can create a map in a completely unknown environment, and use the map to perform positioning, pose (position and attitude) determination, navigation, etc.
- the map created by the SLAM system is called a SLAM map.
- the SLAM map can be understood as a map drawn by the SLAM system based on the environmental information collected by the collection device.
- the collection device may include a visual information collection device in an electronic device and Inertial measurement unit (inertial measurement unit, IMU), wherein, the visual information acquisition device may include, for example, a camera, camera, depth camera, lidar, millimeter-wave radar, etc., and the IMU may include, for example, sensors such as a gyroscope and an accelerometer.
- IMU Inertial measurement unit
- the SLAM map is also referred to as a 3D map. It should be noted that the 3D map includes but is not limited to a SLAM map, and may also include a 3D map created using other technologies, which is not specifically limited in this embodiment of the present application.
- the 3D map may include multiple 3D map points, and correspondingly, data of the 3D map may include data of multiple 3D map points.
- a 3D map point is an interesting or salient feature point in the environment.
- 3D map points are to use laser radar, aerial photography from the perspective of drones (oblique photography), high-definition panoramic cameras, high-definition industrial cameras and other equipment to shoot, through ORB, scale-invariant feature transformation ( scale-invariant feature transform, SIFT), accelerated version has robust features (speeded up robust features, SURF), binary robust independent elementary features (binary robust independent elementary features, BRIEF), binary robust invariant scalable key Point (binary robust invariant scalable keypoints, BRISK), fast retina keypoint (fast retina keypoint, FREAK), D2Net or self-supervised training-based feature point detection and descriptor extraction method (SuperPoint) and other methods from the data captured by the above equipment extract from.
- scale-invariant feature transformation scale-invariant feature transform, SIFT
- accelerated version has robust features (speeded up robust features, SURF), binary robust independent elementary features (binary robust independent elementary features, BRIEF), binary robust invariant scalable key Point (binary robust invariant scalable
- the data of 3D map points can include:
- the 3D map point descriptor is a vector (vector), which is used to represent the local features of the corresponding 3D map point.
- the 3D map point descriptor is used for matching between 3D map points.
- One possible approach is to calculate the distance between two 3D map point descriptors (it can be Euclidean distance, inner product distance, Hamming distance, etc.), and when the distance is less than a threshold, it is considered that two 3D map points match .
- the spatial position of a 3D map point can be represented by X, Y, and Z on the axis of the three-dimensional space, it can also be represented by latitude and longitude, and altitude, and it can also be represented by polar coordinates.
- the embodiment of the present application expresses the spatial position of a 3D map point No specific restrictions are made.
- the spatial position of the 3D map point can be the absolute position of the 3D map point, or the relative position of the 3D map point. For example, taking the center position of the entire area as the origin, all the spatial positions of the 3D map point are relative to the spatial position of the origin offset position.
- each 3D map point can be assigned a number and written into the data of the 3D map, and the storage order of multiple 3D map points in the memory can also be used to implicitly represent the number of the 3D map point.
- the order of multiple 3D map points contained in a 3D map has no practical significance, so the aforementioned numbers can be considered as identifiers for identifying 3D map points to distinguish each 3D map point, but this Numbering is not used to limit the order of multiple 3D map points.
- a 3D map contains 3 3D map points, and their numbers are 1, 2, and 3 respectively.
- the processing of these 3 3D map points can be done according to The order of 1, 2, 3 can also be performed in the order of 3, 2, 1, and the order of 2, 1, 3 can also be performed, and so on.
- the data of the 3D map further includes a plurality of region descriptors, and any one of the region descriptors in the plurality of region descriptors is used to describe part or all of the 3D map in the plurality of 3D map points
- the feature of the point that is, for any one of the multiple area descriptors, the area descriptor can be used to describe the features of some or all of the 3D map points in the multiple 3D map points, so that the area descriptor and the 3D map A point is a one-to-many relationship.
- each 3D map point in multiple 3D map points can be described by some or all of the multiple area descriptors, so the relationship between 3D map points and area descriptors is one-to-many. It can be seen that there is a many-to-many relationship between multiple region descriptors and multiple 3D map points.
- the generation methods of regional descriptors include, but are not limited to, traditional methods such as bag of words (BOW), vector of locally aggregated descriptors (VLAD), and methods based on NetVLAD and artificial intelligence (AI). new method.
- BOW bag of words
- VLAD vector of locally aggregated descriptors
- AI artificial intelligence
- multiple region descriptors may also be identified with a number to distinguish the multiple region descriptors, but the number is also not a limitation on the order of the multiple region descriptors.
- the data of the 3D map also includes the correspondence between the 3D map points and the descriptors, which clearly describe which 3D map points any one descriptor corresponds to, and any 3D map Which descriptors dots correspond to.
- the above correspondence can be explicitly described by a correspondence table between the number of the region descriptor and the number of the 3D map point.
- the 3D map contains 3 region descriptors, the numbers are T1 ⁇ T3, and there are 6 3D map points, where the numbers of the spatial positions of the six 3D map points are P 1 to P 6 , and the numbers of the six 3D map point descriptors are F 1 to F 6 , and the corresponding table is shown in Table 1.
- Table 1 is an example of the correspondence table between the number of the region descriptor and the number of the 3D map point, and the correspondence table can also be presented in other formats or ways, which is not specifically limited in this application.
- the above corresponding relationship can also be implicitly described by the storage location of the region descriptor and the 3D map point, for example, first store T1 in the memory, then store the data of P 1 , P 2 and P 3 , and then store T2, followed by storing the data of P2 and P3 , and finally storing T3, followed by storing the data of P3 , P4 , P5 and P6 .
- FIG. 2 is a schematic structural diagram of an electronic device 20 provided by an embodiment of the present application.
- the electronic device 20 may be the first electronic device and one or more second electronic devices in the embodiment shown in FIG. 1 at least one of the .
- the structure shown in FIG. 2 does not constitute a specific limitation on the electronic device 20 .
- the electronic device 20 may include more or fewer components than the structure shown in FIG. 2 , or combine certain components, or separate certain components, or arrange different components.
- the various components shown in Figure 2 may be implemented in hardware, software, or a combination of hardware and software including one or more signal processing and/or application specific integrated circuits.
- the electronic device 20 may include: a chip 21, a memory 22 (one or more computer-readable storage media), a user interface 23, a display component 24, a camera 25, a sensor 26, a positioning module 27 for positioning the device, and a communication device. Transceiver 28. Communication between these components may be via one or more buses 29 .
- the chip 21 may include: one or more processors 211 , a clock module 212 and a power management module 213 .
- the clock module 212 integrated in the chip 21 is mainly used to provide the processor 211 with a timer required for data transmission and timing control, and the timer can realize the clock function of data transmission and timing control.
- the processor 211 can perform operations according to instruction opcodes and timing signals, generate operation control signals, and complete control of fetching and executing instructions.
- the power management module 213 integrated in the chip 21 is mainly used to provide stable and high-precision voltage for the chip 21 and other components of the electronic device 20 .
- the processor 211 may also be referred to as a central processing unit (central processing unit, CPU), and the processor 211 may specifically include one or more processing units, for example, the processor 211 may include an application processor (application processor, AP), modem Tuning processor, graphics processing unit (graphics processing unit, GPU), image signal processor (image signal processor, ISP), controller, video codec, digital signal processor (digital signal processor, DSP), baseband processor , and/or, a neural network processor (neural-network processing unit, NPU), etc. Wherein, different processing units may be independent devices, or may be integrated in one or more processors.
- application processor application processor, AP
- modem Tuning processor graphics processing unit
- image signal processor image signal processor
- ISP image signal processor
- controller video codec
- digital signal processor digital signal processor
- baseband processor baseband processor
- NPU neural network processing unit
- different processing units may be independent devices, or may be integrated in one or more processors.
- the processor 211 may include one or more interfaces.
- the interface may include an integrated circuit (inter-integrated circuit, I2C) interface, an integrated circuit built-in audio (inter-integrated circuit sound, I2S) interface, a pulse code modulation (pulse code modulation, PCM) interface, a universal asynchronous transmitter (universal asynchronous receiver/transmitter, UART) interface, mobile industry processor interface (mobile industry processor interface, MIPI), general-purpose input and output (general-purpose input/output, GPIO) interface, subscriber identity module (subscriber identity module, SIM) interface, and /or, a universal serial bus (universal serial bus, USB) interface, etc.
- I2C integrated circuit
- I2S integrated circuit built-in audio
- PCM pulse code modulation
- PCM pulse code modulation
- UART universal asynchronous transmitter
- MIPI mobile industry processor interface
- GPIO general-purpose input and output
- subscriber identity module subscriber identity module
- SIM subscriber identity module
- USB universal serial bus
- the memory 22 can be connected to the processor 211 through the bus 29, or can be coupled with the processor 311, and is used for storing various software programs and/or multiple sets of instructions.
- the memory 22 may include a high-speed random access memory (such as a cache memory), or may include a non-volatile memory, such as one or more magnetic disk storage devices, flash memory devices or other non-volatile solid-state storage devices.
- the memory 22 can store operating systems, for example, Android (Android), Apple Mobile Platform (IOS), Microsoft Window Operating System (Windows) or embedded operating systems such as Linux.
- the memory 22 can also store data, for example, image data, point cloud data, 3D map data, pose data, coordinate system conversion information, map update information, and the like.
- the memory 22 may also store computer-executable program codes, where the computer-executable program codes include instructions, for example, communication program instructions, related program instructions of the SLAM system, and the like.
- the memory 22 can also store one or more application programs, for example, virtual scene application programs such as AR/VR/MR, map application programs, image management application programs, navigation and control application programs, and the like.
- the memory 22 can also store a user interface program.
- the user interface program can display the content of the application program through a graphical operation interface, for example, virtual objects in virtual scenes such as AR/VR/MR. 24 presentation, and realize receiving the user's control operations on the application through input controls such as menus, dialog boxes, and buttons.
- the user interface 23 may be, for example, a touch panel, and the touch panel may detect a user's operation instruction thereon, and the user interface 23 may also be, for example, a keypad, a physical button, or a mouse.
- Electronic device 20 may include one or more display components 24 .
- the electronic device 20 can jointly realize the display function through the display component 24, the graphics processing unit (GPU) and the application processor (AP) in the chip 21, etc.
- the GPU is a microprocessor that implements image processing and is connected to the display assembly 24 and the application processor.
- the GPU performs mathematical and geometric calculations for graphics rendering.
- the display component 24 can display the interface content output by the electronic device 20, for example, display images, videos, etc. in virtual scenes such as AR/VR/MR, and the interface content can include the interface of the running application program and the system level menu, etc.
- interface elements such as buttons (Button), text input boxes (Text), sliders (Scroll Bar), menus (Menu), etc.; output interface elements, such as windows (Window), labels (Label), image, video, animation, etc.
- the display component 24 can be a display panel, glasses (such as VR glasses), projection screen and so on.
- the display panel may also be called a display screen, for example, may be a touch screen, a flexible screen, a curved screen, etc., or may be other optical components.
- the display screen of the electronic device in the embodiment of the present application may be a touch screen, flexible screen, curved screen or other forms of screen, that is, the display screen of the electronic device has the function of displaying images, and the specific The material and shape are not specifically limited.
- the display panel can be a liquid crystal display (liquid crystal display, LCD), an organic light-emitting diode (organic light-emitting diode, OLED), an active matrix organic light-emitting diode or an active matrix Organic Light Emitting Diode (active-matrix organic light emitting diode, AMOLED), flexible light emitting diode (flex light-emitting diode, FLED), Miniled, MicroLed, Micro-oLed, quantum dot light emitting diodes (quantum dot light emitting diodes, QLED )Wait.
- liquid crystal display liquid crystal display
- OLED organic light-emitting diode
- AMOLED active matrix organic light-emitting diode
- FLED flexible light emitting diode
- Miniled MicroLed, Micro-oLed
- quantum dot light emitting diodes quantum dot light emitting diodes (quantum dot light emitting diodes, QLED )Wait.
- the touch panel in the user interface 23 and the display panel in the display component 24 can be coupled together, for example, the touch panel can be arranged under the display panel, and the touch panel is used for Detecting the touch pressure on the display panel when the user inputs a touch operation (such as clicking, sliding, touching, etc.) through the display panel, and the display panel is used for displaying content.
- a touch operation such as clicking, sliding, touching, etc.
- the camera 25 may be a monocular camera, a binocular camera or a depth camera, and is used for shooting/recording the environment to obtain images/video images.
- the image/video image collected by the camera 25 can be used as a kind of input data of the SLAM system, or can be displayed by the display component 24 for image/video.
- the camera 25 can also be regarded as a sensor.
- the image collected by the camera 25 may be in the IMG format or in other formats, which is not specifically limited in this embodiment of the present application.
- the sensor 26 can be used to collect data related to state changes of the electronic device 20 (for example, rotation, swing, movement, shaking, etc.), and the data collected by the sensor 26 can be used as an input data of the SLAM system.
- the sensor 26 may include one or more sensors, for example, an inertial measurement unit (inertial measurement unit, IMU), a time of flight (time of flight, TOF) sensor, and the like.
- the IMU may include sensors such as a gyroscope and an accelerometer.
- the gyroscope is used to measure the angular velocity of the electronic device when it is in motion
- the accelerometer is used to measure the acceleration of the electronic device when it is in motion.
- the TOF sensor can include a light emitter and a light receiver.
- the light emitter is used to emit light, such as laser, infrared, radar waves, etc.
- the light receiver is used to detect reflected light, such as reflected laser, infrared, and radar waves. Wait.
- the sensor 26 may also include more other sensors, such as inertial sensors, barometers, magnetometers, wheel speedometers, etc., which are not specifically limited in this embodiment of the present application.
- the positioning module 27 is used to realize the physical positioning of the electronic device 20 , for example, to acquire the initial position of the electronic device 20 .
- the positioning module 27 may include one or more of a WiFi positioning module, a Bluetooth positioning module, a base station positioning module, and a satellite positioning module.
- the satellite positioning module can be provided with a global navigation satellite system (global navigation satellite system, GNSS) to assist positioning, and GNSS is not limited to the Beidou system, the global positioning system (global positioning system, GPS) system, the GLONASS (global navigation satellite system) , GLONASS) system, Galileo satellite navigation system (Galileo) system.
- the transceiver 28 is used to enable communication between the electronic device 20 and other devices (eg, servers, other electronic devices, etc.).
- the transceiver 28 integrates a transmitter and a receiver for sending and receiving radio frequency signals, respectively.
- the transceiver 28 includes, but is not limited to: an antenna system, a radio frequency (radio frequency, RF) transceiver, one or more amplifiers, a tuner, one or more oscillators, a digital signal processor, a codec ( CODEC) chip, subscriber identification module (SIM) card and storage media, etc.
- the transceiver 28 can also be implemented on a separate chip.
- the transceiver 28 supports at least one data network communication in 2G/3G/4G/5G, etc., and/or supports at least one of the following short-range wireless communication methods: Bluetooth (bluetooth, BT) communication, wireless fidelity (wireless fidelity, WiFi) communication, near field communication (near field communication, NFC), infrared (infrared, IR) wireless communication, ultra wide band (ultra wide band, UWB) communication, ZigBee protocol (ZigBee) communication.
- Bluetooth bluetooth, BT
- wireless fidelity wireless fidelity, WiFi
- NFC near field communication
- infrared infrared
- UWB ultra wide band
- ZigBee protocol ZigBee protocol
- the processor 211 executes various functional applications and data processing of the electronic device 20 by running the program code stored in the memory 22 .
- FIG. 3 is a schematic structural diagram of a server 30 provided in an embodiment of the present application.
- the server 30 may be the server in the embodiment shown in FIG. 1 .
- Server 30 includes processor 301 , memory 302 (one or more computer-readable storage media), and transceiver 303 . Communication between these components may be via one or more buses 304 .
- the processor 301 may be one or more CPUs. In the case where the processor 301 is one CPU, the CPU may be a single-core CPU or a multi-core CPU.
- the memory 302 can be connected to the processor 301 through the bus 304, or can be coupled with the processor 301, and is used to store various program codes and/or multiple sets of instructions, and data (eg, map data, pose data, etc.).
- the memory 302 includes, but is not limited to, random access memory (Random Access Memory, RAM), read-only memory (Read-Only Memory, ROM), erasable programmable read-only memory (Erasable Programmable Read Only Memory, EPROM), or portable read-only memory (Compact Disc Read-Only Memory, CD-ROM), etc.
- the transceiver 303 mainly integrates a receiver and a transmitter, where the receiver is used to receive data sent by the electronic device (for example, requests, images, etc.), and the transmitter is used to send data to the electronic device (for example, map data, pose data, etc.) .
- server 30 shown in FIG. 3 is only an example provided by the embodiment of the present application, and the server 30 may have more components than shown in the figure, which is not specifically limited in the embodiment of the present application.
- the processor 301 executes various functional applications and data processing of the server 30 by running the program code stored in the memory 302 .
- Coupled used in the embodiments of the present application means to be directly connected or connected through one or more intermediate components or circuits.
- Figure 4a is a schematic diagram of an application scenario provided by the embodiment of the present application.
- the application scenario is that the electronic device collects visual information through a sensor, and determines the current pose of the electronic device in combination with the visual information and the 3D map from the server.
- the 3D map is provided by the server, that is, the server creates the 3D map, then compresses the 3D map, and transmits the compressed data of the 3D map to the electronic device; the electronic device performs decompression processing after receiving the compressed data of the 3D map, and obtains
- the reconstruction data of the 3D map is combined with the collected visual information and the reconstruction data of the 3D map to determine the current pose of the electronic device.
- the pose is the position information of the electronic device, which can be the absolute pose in the world coordinate system, or the relative pose relative to a certain point in the environment.
- the server may pre-create the 3D map, compress the 3D map and store it locally, which can save storage space.
- the server can also transmit the compressed data of the 3D map to other devices, such as cloud storage.
- the server creates a 3D map and compresses the compressed data of the 3D map to be stored locally.
- Compressing the 3D map by the server can save local storage space.
- the electronic device sends a map download request to the server. There are two ways to trigger the map download request:
- the user opens the map application program installed on the electronic device, and the application program uploads the location information obtained based on GPS positioning or WiFi positioning to its corresponding server, and the uploading operation can trigger a map download request. Since the uploaded content includes location information, the server can make a preliminary estimate based on the location information, and transmit the compressed data of the 3D map of the area to which the positioning point indicated by the location information belongs to the electronic device.
- the scope of the area to which the anchor point indicated by the location information can be preset.
- the area to which the anchor point belongs can be the administrative region of each level (including counties, cities, countries, or administrative districts, etc.) where the anchor point is located, or it can be centered on the anchor point , Set the distance as the circle area of the radius.
- the user opens the map application program installed on the electronic device, and actively inputs or selects an area in the application program, for example, the user actively enters "xx business center", or selects from a list "A street, B street, "A Street” was selected in "C Street”.
- the foregoing operations of the user may trigger a map download request.
- Either the user input or selection includes a geographic location, so the server transmits the compressed data of the 3D map of the geographic location to the electronic device.
- the electronic device automatically detects whether it meets the conditions for downloading a 3D map or starts downloading a 3D map, or the electronic device passes the detection A change in ambient light or a change in the environment is detected to start downloading the 3D map, so as to request the server to download a 3D map within an area, and the size of the area is not specifically limited.
- the server sends the compressed data of the 3D map to the electronic device.
- Electronic equipment collects visual information through sensors.
- step 3 and step 4 are independent of each other, and the sequence is not limited.
- the electronic device decompresses the compressed data of the 3D map to obtain reconstructed data of the 3D map.
- the electronic device performs positioning in the 3D map according to the visual information to obtain a pose corresponding to the visual information.
- the electronic device After receiving the compressed data of the 3D map, the electronic device does not need to decompress it immediately, but only needs to decompress to obtain the reconstruction data of the 3D map before positioning based on the visual information. For example, the user can pre-download the compressed data of an area-wide 3D map by downloading an "offline map", and decompress the compressed data of the 3D map when positioning is required.
- Figure 4b is a schematic diagram of an application scenario provided by the embodiment of the present application.
- the application scenario is that the electronic device collects visual information through a sensor, and the server determines the current position of the electronic device in combination with the visual information from the electronic device and a 3D map. posture.
- the 3D map is provided by the server, that is, the server creates the 3D map, then compresses the 3D map, and stores the compressed data of the 3D map locally.
- the server performs decompression processing to obtain the reconstructed data of the 3D map, and combines the visual information and the reconstructed data of the 3D map to determine the current pose of the electronic device.
- the server creates a 3D map and compresses the compressed data of the 3D map to be stored locally.
- Electronic devices collect visual information through sensors.
- the electronic device sends the visual information to the server.
- the server decompresses the compressed data of the 3D map to obtain reconstructed data of the 3D map.
- the server compresses the 3D map to save storage space.
- the server performs positioning in the 3D map according to the visual information to obtain a pose corresponding to the visual information.
- the server sends the pose to the electronic device.
- Fig. 4c is a schematic diagram of an application scenario provided by the embodiment of the present application.
- the application scenario is that the electronic device collects visual information through a sensor, and determines the current pose of the electronic device in combination with the collected visual information and a 3D map.
- the 3D map is provided by the electronic device, that is, the electronic device creates the 3D map, then compresses the 3D map, and stores the compressed data of the 3D map locally.
- the electronic device performs decompression processing to obtain the reconstructed data of the 3D map, and combines the collected visual information and the reconstructed data of the 3D map to determine the current pose of the electronic device.
- the electronic device creates a 3D map and compresses the compressed data of the 3D map to be stored locally.
- Electronic equipment collects visual information through sensors.
- the electronic device decompresses the compressed data of the 3D map to obtain reconstructed data of the 3D map.
- the electronic device performs positioning in the 3D map according to the visual information to obtain a pose corresponding to the visual information.
- Fig. 4d is a schematic diagram of an application scenario provided by an embodiment of the present application.
- the application scenario is that the second electronic device collects visual information through a sensor, and determines the location of the second electronic device in combination with the visual information and the 3D map from the server. current pose.
- the 3D map is created by the first electronic device, that is, the first electronic device creates the 3D map, compresses the 3D map, and then sends the compressed data of the 3D map to the server, and the server sends the compressed data of the 3D map to the second
- the second electronic device performs decompression processing to obtain the reconstruction data of the 3D map, and determines the current pose of the second electronic device in combination with the collected visual information and the reconstruction data of the 3D map.
- the first electronic device may pre-create a 3D map, compress the 3D map and then transmit it to the server, which can reduce transmission bandwidth.
- the first electronic device creates a 3D map and compresses it to obtain compressed data of the 3D map.
- the first electronic device sends the compressed data of the 3D map to the server.
- the first electronic device compresses the 3D map before transmitting it, which can reduce the transmission bandwidth and improve the transmission efficiency.
- the second electronic device sends a map download request to the server.
- the map download request sent by the second electronic device may also be based on the trigger manner shown in FIG. 4a.
- the server sends the compressed data of the 3D map to the second electronic device.
- the second electronic device decompresses the compressed data of the 3D map to obtain reconstructed data of the 3D map.
- the second electronic device collects the visual information through the sensor.
- the second electronic device performs positioning in the 3D map according to the visual information to obtain a pose corresponding to the visual information.
- Fig. 4e is a schematic diagram of an application scenario provided by the embodiment of the present application.
- the application scenario is that the second electronic device collects visual information through a sensor, and the server combines the visual information from the second electronic device with the visual information from the first electronic device.
- the 3D map of the device determines the current pose of the second electronic device.
- the 3D map is created by the first electronic device, that is, the first electronic device creates the 3D map, compresses the 3D map, and then sends the compressed data of the 3D map to the server.
- the server performs decompression processing to obtain the reconstructed data of the 3D map, and combines the visual information from the second electronic device and the reconstructed data of the 3D map to determine the current pose of the second electronic device.
- the first electronic device creates a 3D map and compresses it to obtain compressed data of the 3D map.
- the first electronic device sends the compressed data of the 3D map to the server.
- the second electronic device acquires visual information through sensor collection.
- the second electronic device sends a positioning request to the server, and the positioning request carries visual information.
- the server decompresses the compressed data of the 3D map to obtain reconstructed data of the 3D map.
- the server performs positioning in the 3D map according to the visual information to obtain a pose corresponding to the visual information.
- the server sends the pose obtained by positioning to the second electronic device.
- Fig. 4f is a schematic diagram of an application scenario provided by the embodiment of the present application.
- the application scenario is that the second electronic device collects visual information through a sensor, and combines the visual information and the 3D map from the first electronic device to determine the second The current pose of the electronic device.
- the 3D map is created by the first electronic device, that is, the first electronic device creates a 3D map, performs compression processing on the 3D map, and then sends the compressed data of the 3D map to the second electronic device, and the second electronic device performs decompression processing , obtain the reconstructed data of the 3D map, and determine the current pose of the second electronic device in combination with the collected visual information and the 3D map from the first electronic device.
- the first electronic device creates a 3D map, and compresses the compressed data of the 3D map to be stored locally.
- the second electronic device sends a map download request to the first electronic device.
- the first electronic device sends the compressed data of the 3D map to the second electronic device.
- the second electronic device decompresses the compressed data of the 3D map to obtain reconstructed data of the 3D map.
- the second electronic device acquires visual information through sensor collection.
- the second electronic device performs positioning in the 3D map according to the visual information to obtain a pose corresponding to the visual information.
- the positioning algorithm used may include:
- Extract the region descriptor to be retrieved from the visual information, and the algorithm used to extract the region descriptor to be retrieved is the same as the algorithm for extracting the region descriptor from the 3D map.
- the distance between the region descriptor to be retrieved and each region descriptor in multiple region descriptors can be calculated, the distance can include Hamming distance, Manhattan distance or Euclidean distance, etc., and then select the matching condition (For example, at least one region descriptor whose distance is smaller than the threshold value) is a candidate region descriptor.
- the pose solving algorithm calculates the pose of the electronic device.
- FIG. 4g is a schematic diagram of a user interface displayed by the electronic device provided in the embodiment of the present application.
- the electronic device can display the user interface shown in FIG. 4g based on the pose.
- the user interface may include a navigation arrow indication for navigating to the conference room 2, and the navigation arrow indication for navigating to the conference room 2 may be a virtual object obtained from the server or locally based on the pose.
- the user interface may also include visual information collected by sensors, for example, buildings as shown in FIG. 4g. The user goes to the conference room 2 referring to the user interface shown in FIG. 4g of the electronic device.
- the reconstructed data of the 3D map obtained through decompression involved in the embodiment of the present application may also be referred to as the reconstructed data of the 3D map.
- FIGS. 4a to 4f all involve compressing and decompressing the 3D map, and performing positioning based on the reconstructed data of the 3D map obtained through decompression, so as to obtain the current pose of the electronic device.
- a feasible way is to perform positioning based on the reconstructed data of the 3D map obtained through complete decompression.
- the 3D map retrieval method provided in the following embodiments of the present application may be used to filter out descriptors and use the descriptors for positioning.
- the following embodiments of the present application provide a retrieval method for a 3D map, by alternately performing retrieval and decompression or extracting from compressed data, so as to perform multi-level retrieval in the 3D map and obtain descriptors.
- a retrieval method for a 3D map by alternately performing retrieval and decompression or extracting from compressed data, so as to perform multi-level retrieval in the 3D map and obtain descriptors.
- at least the other level of retrieval in the two-level retrieval can guarantee the retrieval accuracy.
- Incomplete decompression refers to incomplete decompression to obtain all reconstructed data of the 3D map. A method for searching the 3D map will be described below.
- the retrieval of the 3D map involved in the embodiment of the present application refers to finding some of the most similar or most relevant 3D map points among the multiple 3D map points of the 3D map.
- the length of the binarized data involved in the embodiment of the present application may be the number of bits (also referred to as the number of bits) of the binarized data. For example, if the binarized data is 110, the number of bits of the binarized data is 3.
- retrieval method provided in the embodiment of the present application can also be applied to other technical fields, for example, retrieval of image or video data, retrieval of audio data, retrieval of point cloud data, and the like.
- FIG. 5A is a schematic flowchart of a fragment of a method for retrieving a 3D map provided by an embodiment of the present application.
- the method can be applied to any of the electronic devices illustrated in FIG. 1 to FIG. 4f, or also It can be applied to the server shown in any one of Fig. 1 to Fig. 4f.
- the method includes but is not limited to the following steps:
- S101A Extract binarized data of S 3D map descriptors from the compressed data of S 3D map descriptors, where the S 3D map descriptors correspond to multiple 3D map points in the 3D map.
- the binarized data may be first binarized data, second binarized data, ... nth binarized data, etc., n is a positive integer. It should be understood that “first”, “second” and “third” in the first binarized data, the second binarized data and the third binarized data, etc. differentiate each other.
- the n-th binarized data is taken as an example for explanation below.
- the nth binarized data of the S 3D map descriptors is extracted from the compressed data of the S 3D map descriptors.
- S is a positive integer.
- S represents the number of 3D map descriptors that need to be retrieved in the i-th level retrieval.
- S ⁇ T, T represents the total number of 3D map descriptors in the 3D map.
- the number of multiple 3D map points corresponding to the S 3D map descriptors may be greater than S, or equal to S.
- the S 3D map descriptors may be S area descriptors, or S 3D map point descriptors.
- one region descriptor corresponds to multiple 3D map points
- one 3D map point descriptor corresponds to one 3D map point.
- the compressed data of the S 3D map descriptors may be obtained by compressing the S region descriptors or the S 3D map point descriptors.
- the compression processing here may include: compaction processing.
- the compaction processing may include binarization processing and quantization processing.
- the nth binarized data of the S 3D map descriptors may be partial data of the respective compressed data of the S 3D map descriptors, for example, the nth binarized data of the S 3D map descriptors may be S 3D maps Descriptors are part or all of the binarized data of the respective compressed data.
- the binarization process includes but is not limited to: iterative quantization (iterative quantization, ITQ) hashing, locality-sensitive hashing (locality-sensitive hashing, LSH), or spectral hashing (spectral hashing).
- ITQ iterative quantization
- LSH locality-sensitive hashing
- spectral hashing spectral hashing
- the descriptor can be mapped to Hamming space (also called binary space) through binarization processing to obtain binarized data.
- the quantization process includes, but is not limited to: scalar quantization, vector quantization, or product quantization.
- the 3D map descriptor may be compressed into one or more quantization indices through quantization processing, and the one or more quantization indices are quantized data of the 3D map descriptor. Wherein, each quantization index in the one or more quantization indices corresponds to a quantization center.
- the execution subject of the embodiment of the present application may perform entropy decoding on the compressed data of the S 3D map descriptors, and extract n-th binarized data of the S 3D map descriptors therefrom.
- n-th binarized data is taken as an example for explanation.
- the i-th level search is performed on the n-th binarized data of the S 3D map descriptors to obtain P 3D map descriptors.
- the retrieval descriptor is a feature corresponding to the real environment extracted from the visual information collected by the sensor of the electronic device.
- the retrieval descriptor here may be the region descriptor to be retrieved in the above embodiment, or the 3D map point descriptor to be retrieved. That is, the retrieval descriptor may be a region to be retrieved descriptor extracted from visual information, and the region to be retrieved descriptor is used to describe features of some or all of the 3D map points among the plurality of 3D map points.
- the region descriptor to be retrieved may be a vector, for example, a 128-dimensional vector, or a 512-dimensional vector, and so on.
- the dimensions of the region descriptor to be retrieved are not limited by the above examples, and may also be other values, which are not illustrated one by one in the embodiment of the present application.
- the retrieval descriptor may be a 3D map point descriptor to be retrieved extracted from visual information, and the 3D map point descriptor to be retrieved is used to represent a local feature of the corresponding 3D map point.
- the 3D map point descriptor to be retrieved may be a vector, and the dimension of the vector may be any positive integer.
- the ways of obtaining the retrieval descriptor may include but not limited to the following ways.
- the execution subject of this embodiment may receive the retrieval descriptor sent by other devices.
- the execution subject of this embodiment may be the first electronic device as shown in FIG. 1
- the first electronic device may receive the retrieval descriptor sent by the second electronic device.
- the execution subject of this embodiment may be a server as shown in FIG. 1 , and the server may receive the retrieval descriptor sent by the first electronic device or the second electronic device.
- the execution subject of this embodiment may receive visual information sent by other devices, and extract a retrieval descriptor from the visual information.
- the execution subject of this embodiment may be the first electronic device as shown in FIG.
- the first electronic device may receive the visual information sent by the second electronic device, and the first electronic device extracts the retrieval descriptor from the visual information.
- the execution subject of this embodiment may be a server as shown in FIG. 1 , the server may receive the visual information sent by the first electronic device or the second electronic device, and the server extracts the retrieval descriptor from the visual information.
- the execution subject of this embodiment can detect the operation of collecting visual information input by the user, and in response to the operation, trigger its own sensor to collect visual information from the real environment, obtain visual information, and extract and retrieve the visual information from the visual information. descriptor.
- the execution subject of this embodiment may be the first electronic device as shown in FIG. 1.
- the first electronic device When the first electronic device detects the operation of collecting visual information input by the user, in response to the operation, it triggers its own sensor to monitor the real environment.
- the visual information is collected, the visual information is obtained, and the retrieval descriptor is extracted from the visual information.
- the operation of collecting visual information may be to start the sensor to collect visual information by clicking, touching, sliding, or shaking, or to start the sensor to collect visual information by voice control or other means, which is not limited in this embodiment of the present application.
- the first electronic device detects the user's touch operation on the shooting button on the user interface, in response to the touch operation, the camera of the first electronic device is activated and so on.
- the nth binarized data of the search descriptor may be part or all of the binarized data of the search descriptor.
- the execution subject of this embodiment can perform binarization processing on the retrieval descriptor to obtain the binarized data of the retrieval descriptor, and select (also called extraction) part or All as the nth binarized data of the retrieval descriptor.
- the execution subject of this embodiment can use the nth binarization data of the retrieval descriptor to search in the nth binarization data of S 3D map descriptors, so as to obtain the nth binarization data of the retrieval descriptor The most similar or most related P 3D map descriptors.
- P is a positive integer, and P represents the number of 3D map descriptors filtered out by the i-th level retrieval. P ⁇ T.
- P P ⁇ S.
- a 3D map descriptor set is obtained through the i-level retrieval, and the number of 3D map descriptors in the 3D map descriptor set is less than the 3D map descriptions in the 3D map description subset before the i-level retrieval the number of children.
- P may be greater than S.
- the similarity or correlation between the nth binarized data of each of the P 3D map descriptors and the nth binarized data of the retrieval descriptor degree which is higher than other 3D map descriptors in the S 3D map descriptors except the P 3D map descriptors.
- there are many ways to calculate the similarity or correlation For example, by calculating the first distance of the respective nth binarized data of two 3D map descriptors (3D map descriptor and retrieval descriptor), determine Correlation of the nth binarized data of two 3D map descriptors. Wherein, the smaller the first distance is, the higher the correlation is.
- the first distance may include a distance obtained by using binarized data, such as a Hamming distance.
- the first distance may be the Hamming distance.
- the first distance may be a distance obtained by taking the absolute value of the difference between corresponding bits of the two binarized data and adding all the absolute values of the difference.
- the retrieval based on the first distance refers to calculating the first distance to determine the degree of correlation or similarity to filter the 3D map descriptors.
- the reconstructed data may be the first reconstructed data, the second reconstructed data, . . . the mth reconstructed data, etc., where m is a positive integer. It should be understood that the "first”, “second” and “third” in the first reconstructed data, the second reconstructed data and the third reconstructed data, etc., have no order, but are only for distinguishing each other .
- the m-th reconstructed data is taken as an example for explanation below.
- P 3D map descriptors can be screened and obtained, and through S103A, the compressed data of the P 3D map descriptors can be decompressed at the mth level to obtain the mth reconstructed data of the P 3D map descriptors.
- m is a positive integer.
- the mth stage of decompression may include inverse quantization processing.
- the m-th stage of decompression may be staged decompression (that is, a part of complete decompression), or it may be complete decompression.
- the mth reconstructed data of the P 3D map descriptors may be obtained by decompressing part or all of the compressed data of the P 3D map descriptors at the mth level.
- the mth reconstructed data of the P 3D map descriptors may be dequantized data obtained by performing an inverse quantization process on part or all of the compressed data of the P 3D map descriptors.
- the dequantized data may also be referred to as reconstructed data, and the reconstructed data may be obtained through dequantization methods in the prior art.
- the mth reconstructed data of the P 3D map descriptors can be the quantized data (part of the compressed data) of the P 3D map descriptors, which are dequantized to obtain the P dequantized data, and then according to P pieces of dequantized data and P pieces of 3D map descriptors are obtained from binarized data (a part of compressed data).
- the m-th stage decompression may also include other decompression processing such as entropy decoding and prediction, and this embodiment of the present application does not illustrate one by one.
- Q is a positive integer
- T represents the total number of 3D map descriptors in the 3D map
- j i+1, 1 ⁇ i ⁇ L , 1 ⁇ j ⁇ L
- L represents the total number of retrieval levels of the 3D map or a threshold value of the retrieval number
- L is a positive integer greater than 1.
- the reconstructed data may be the first reconstructed data, the second reconstructed data, . . . the mth reconstructed data, etc., where m is a positive integer. It should be understood that the "first”, “second” and “third” in the first reconstructed data, the second reconstructed data and the third reconstructed data, etc., have no order, but are only for distinguishing each other .
- the m-th reconstructed data is taken as an example for explanation below.
- Retrieving part or all of the data of the descriptor refers to retrieving part or all of the components of the descriptor.
- the component of the retrieval descriptor refers to the value of a dimension in the vector of the retrieval descriptor.
- the retrieval descriptor may be a vector, for example, an M-dimensional vector, each of which is a component of the vector.
- the execution subject of this embodiment can use part or all of the data of the retrieval descriptor to search in the mth reconstruction data of the P 3D map descriptors to obtain the mth reconstruction data of the retrieval descriptor
- the Q 3D map descriptors with the most similar or relevant data. Q can be smaller than P. It can also be understood that another 3D map descriptor set is obtained through the j-th level retrieval, and the number of 3D map descriptors in the 3D map descriptor set is less than the number of 3D map descriptors obtained by the i-th level retrieval.
- the similarity or correlation between the mth reconstructed data of each of the Q 3D map descriptors and some or all of the data of the retrieval descriptor is higher than that of the P 3D map descriptors except for the Q 3D map descriptors Other 3D map descriptors.
- there are many ways to calculate the similarity or correlation for example, by calculating the second of two 3D map descriptors (the mth reconstructed data of the 3D map descriptor and part or all of the data of the retrieval descriptor).
- Distance which determines the degree of correlation between two 3D map descriptors.
- the second distance may include, but is not limited to, Euclidean distance, inner product distance, cosine distance, Manhattan distance, and the like. For example, the smaller the Euclidean distance, the higher the correlation, and the larger the inner product distance, the higher the correlation.
- condition judgment step can be performed after S104A.
- the search of this embodiment can be ended, and if the condition is not met, one-level or multi-level search can be continued.
- This condition can be any condition, which can be reasonably set according to requirements.
- the condition can be that the similarity or correlation is higher than or equal to the preset threshold, or that the number of retrieved 3D map descriptors is less than or equal to the preset threshold. Set the number and so on.
- the condition may be that the following N 3D map descriptors used for positioning are retrieved.
- the retrieval method of the 3D map in the embodiment of the present application is a fragment process, which may also include other levels of retrieval before and/or after the fragment process.
- the retrieval method of the 3D map may include more stages than those shown in FIG. 5A
- the multi-level search for example, 3, 4 or 5 levels and so on. For example, there may be more levels of retrieval before S101A in the embodiment shown in FIG. 5A , or there may be more levels of retrieval after S104A in the embodiment shown in FIG. 5A , which will not be illustrated here one by one.
- Q 3D map descriptors obtained through multi-level retrieval can be used for positioning. If the 3D map points corresponding to the N 3D map descriptors in the Q 3D map descriptors match the 3D map points corresponding to the retrieval descriptors, then positioning can be performed according to the 3D map points corresponding to the N 3D map descriptors to obtain Pose information of electronic devices.
- the electronic device may be an electronic device that collects the aforementioned visual information. 0 ⁇ N ⁇ Q.
- the binarized data of S 3D map descriptors are extracted from the compressed data of S 3D map descriptors, and the binarized data of S 3D map descriptors is In the i-level retrieval, P 3D map descriptors are screened. Perform m-level decompression on the compressed data of P 3D map descriptors to obtain the reconstructed data of P 3D map descriptors. The j-level retrieval is performed in the data to filter out a smaller number of 3D map descriptors.
- the first-level retrieval of at least two-level retrieval uses binary data in the compressed data, which can increase the retrieval speed, and the other-level retrieval uses the reconstructed data obtained by decompression, which can ensure retrieval accuracy.
- the retrieval method for the 3D map provided in the embodiment of the present application can improve the retrieval performance.
- FIG. 5B is a schematic flow chart of a fragment of a method for retrieving a 3D map provided by an embodiment of the present application.
- the method can be applied to any of the electronic devices illustrated in FIG. 1 to FIG. 4f, or also It can be applied to the server shown in any one of Fig. 1 to Fig. 4f.
- the method includes but is not limited to the following steps:
- S101B Extract first binarized data of S 3D map descriptors from the compressed data of S 3D map descriptors, where the S 3D map descriptors correspond to multiple 3D map points in the 3D map.
- the first binarized data may refer to any one of the first binarized data, the second binarized data, . . . the nth binarized data, and n is a positive integer. It should be understood that “first”, “second” and “third” in the first binarized data, the second binarized data and the third binarized data, etc. differentiate each other.
- the n-th binarized data is taken as an example for explanation below.
- the nth binarized data of the S 3D map descriptors are extracted from the compressed data of the S 3D map descriptors, and the S 3D map descriptors correspond to multiple 3D map points in the 3D map.
- the n-th binarized data is taken as an example for explanation below.
- the i-th level search can be performed on the nth binarized data of the S 3D map descriptors to obtain P 3D map descriptors.
- S101B and S102B reference may be made to S101A and S102A of the embodiment shown in FIG. 5A .
- P 3D map descriptors can be screened from the S 3D map descriptors, and then the qth binarization data of the P 3D map descriptors can be extracted from the compressed data of the P 3D map descriptors, To further perform retrieval based on the qth binarized data.
- the qth binarized data of the P 3D map descriptors may be part or all of the compressed data of the P 3D map descriptors, for example, the qth binarized data of the P 3D map descriptors may be P Part or all of the binarized data for each of the 3D map descriptors.
- the nth binarization data of each 3D map descriptor of the S 3D map descriptors is different from the qth binarization data of each 3D map descriptor of the P 3D map descriptors.
- the difference may be different in length, that is, the length of the nth binarized data of each 3D map descriptor of the S 3D map descriptors is different from the length of each 3D map descriptor of the P 3D map descriptors. Different lengths.
- the difference may be that the binarization processing methods corresponding to the nth binarized data and the qth binarized data are different, that is, the nth of each 3D map descriptor of the S 3D map descriptors
- the binarization processing manner corresponding to the binarization data is different from the binarization processing manner corresponding to each 3D map descriptor of the P 3D map descriptors.
- the nth binarized data of each 3D map descriptor of S 3D map descriptors is obtained through a binarization processing method (for example, locality-sensitive hashing), and each 3D data of P 3D map descriptors
- the map descriptor is obtained through another binarization method (for example, spectral hashing).
- the length of the nth binarized data can be measured or represented by the number of bits of the nth binarized data.
- the length of the qth binarized data can be measured or represented by the number of bits of the qth binarized data.
- the nth binarized data of each 3D map descriptor of P 3D map descriptors in the compressed data of 3D map descriptors The position is different from the position of the qth binarized data of the 3D map descriptor in the compressed data of the 3D map descriptor.
- the location can be a start location or an end location, etc.
- the component of the original 3D map descriptor corresponding to the nth binarization data of each 3D map descriptor of the P 3D map descriptors corresponds to the qth binarization data of the corresponding 3D map descriptor
- the components of the original 3D map descriptor of are different.
- the nth binarized data of a 3D map descriptor may be the binarized data of the first part of the 3D map descriptor (for example, the first 128-dimensional components of the 3D map descriptor), and the nth binarized data of the 3D map descriptor
- the q binarization data may be the binarization data of the second part of the 3D map descriptor (for example, the last 128-dimensional components of the 3D map descriptor).
- the length of the nth binarized data of each 3D map descriptor of P 3D map descriptors may be less than P 3D map descriptors
- the length of the q-th binarized data of each 3D map descriptor of the child is relatively short, which can improve the retrieval speed, and the length of the qth binarized data of the 3D map descriptor in the j-level retrieval is longer. It can improve retrieval accuracy. Through at least two levels of differentiated retrieval, the comprehensive improvement of retrieval speed and retrieval accuracy can be realized.
- the length of the nth binarized data of each 3D map descriptor of the P 3D map descriptors may be greater than the length of the qth binarized data of each 3D map descriptor of the P 3D map descriptors.
- the length of the nth binarized data of each 3D map descriptor of the P 3D map descriptors may be equal to the length of the qth binarized data of each 3D map descriptor of the P 3D map descriptors.
- the jth level retrieval is performed on the qth binarized data of the P 3D map descriptors to obtain Q 3D map descriptors.
- the qth binarized data of the search descriptor may be part or all of the binarized data of the search descriptor.
- the execution subject of this embodiment can perform binarization processing on the retrieval descriptor to obtain the binarized data of the retrieval descriptor, and select (also called extraction) part or All are represented as the n+1th of the retrieval descriptor. It should be noted that the nth binarized data of the retrieval descriptor and the qth binarized data of the retrieval descriptor may be the same or different.
- the execution subject of this embodiment can use the qth binarized data of the search descriptor to search in the qth binarized data of the P 3D map descriptors, so as to obtain the qth binarized data of the search descriptor
- the Q most similar or most relevant 3D map descriptors for the binarized data. Q can be smaller than P. It can also be understood that another 3D map descriptor set is obtained through the j-th level retrieval, and the number of 3D map descriptors in the 3D map descriptor set is less than the number of 3D map descriptors obtained by the i-th level retrieval.
- the similarity or correlation between the qth binarized data of each of the Q 3D map descriptors and the qth binarized data of the retrieval descriptor is higher than that of the P 3D map descriptors except the Q 3D map descriptors other 3D map descriptors.
- the explanation of the calculation method of the similarity degree or the correlation degree can refer to the relevant explanation explanation of S102B, which will not be repeated here.
- the i-th level retrieval in the above S102B and the j-th level retrieval in the above S104B all use the retrieval method based on the first distance, and its specific explanation can refer to the S102A of the embodiment shown in FIG. 5A Explanation, not repeated here.
- condition judgment step can be performed after S104B. If the condition is met, the search in this embodiment can be ended, and if the condition is not met, one or more levels of search can be continued.
- This condition can be any condition, which can be reasonably set according to requirements.
- the condition can be that the similarity or correlation is higher than or equal to the preset threshold, or that the number of retrieved 3D map descriptors is less than or equal to the preset threshold. Set the number, etc., the embodiment of the present application does not give examples one by one.
- the retrieval method of the 3D map in the embodiment of the present application is a fragment process, which may also include other levels of retrieval before and/or after the fragment process.
- the retrieval method of the 3D map may include more stages than those shown in FIG. 5B
- the multi-level search for example, 3, 4 or 5 levels and so on. For example, there may be more levels of retrieval before S101B in the embodiment shown in FIG. 5B , or there may be more levels of retrieval after S104B in the embodiment shown in FIG. 5B , which will not be illustrated here one by one.
- Q 3D map descriptors obtained through multi-level retrieval can be used for positioning. If the 3D map points corresponding to the N 3D map descriptors in the Q 3D map descriptors match the 3D map points corresponding to the retrieval descriptors, then positioning can be performed according to the 3D map points corresponding to the N 3D map descriptors to obtain Pose information of electronic devices.
- the electronic device may be an electronic device that collects the aforementioned visual information. 0 ⁇ N ⁇ Q.
- the first binarized data of S 3D map descriptors are extracted from the compressed data of S 3D map descriptors, and according to the first binarized data of the retrieved descriptors, the S 3D map descriptors are The i-level search is performed on the first binarized data, and P 3D map descriptors are obtained by screening. Extract the second binarized data of P 3D map descriptors from the compressed data of P 3D map descriptors, according to the second binarized data of the retrieved descriptors, in the second binary data of P 3D map descriptors The j-level retrieval is performed in the 2D data to filter out a smaller number of 3D map descriptors.
- the retrieval speed can be improved and the retrieval accuracy can be guaranteed.
- FIG. 5C is a schematic flowchart of a fragment of a method for retrieving a 3D map provided by an embodiment of the present application.
- the method can be applied to any of the electronic devices illustrated in FIG. 1 to FIG. 4f, or also It can be applied to the server shown in any one of Fig. 1 to Fig. 4f.
- the method includes but is not limited to the following steps:
- S101C Decompress the compressed data of the S 3D map descriptors at the mth stage to obtain the first reconstructed data of the S 3D map descriptors, and the S 3D map descriptors correspond to multiple 3D map points in the 3D map.
- the first reconstructed data may refer to any one of the first reconstructed data, the second reconstructed data, ... the mth reconstructed data, etc., m is a positive integer. It should be understood that the "first”, “second” and “third” in the first reconstructed data, the second reconstructed data and the third reconstructed data, etc., have no order, but are only for distinguishing each other . The following takes the first reconstructed data as the mth reconstructed data as an example for explanation.
- S and m are positive integers.
- S represents the number of 3D map descriptors that need to be retrieved in the i-th level retrieval.
- S ⁇ T, T represents the total number of 3D map descriptors in the 3D map.
- the number of multiple 3D map points corresponding to the S 3D map descriptors may be greater than S, or equal to S.
- the compressed data of the S 3D map descriptors may be obtained by compressing the S region descriptors or the S 3D map point descriptors.
- the compression processing here may include: quantization processing.
- the compression processing may further include at least one of the following: binarization processing, prediction processing, or entropy coding processing.
- the mth reconstructed data of the S 3D map descriptors is obtained by decompressing the partial data of the respective compressed data of the S 3D map descriptors in the mth stage.
- the mth stage of decompression may include inverse quantization processing.
- the mth stage of decompression may be a staged decompression (ie part of a full decompression).
- the m-th reconstructed data of the S 3D map descriptors may be obtained by decompressing partial data of the respective compressed data of the S 3D map descriptors through m-th stage decompression.
- the mth reconstructed data of the S 3D map descriptors may be dequantized data obtained by performing an inverse quantization process on parts of the compressed data of the S 3D map descriptors.
- the dequantized data may also be referred to as reconstructed data, and the reconstructed data may be obtained through dequantization methods in the prior art.
- the mth reconstructed data of the S 3D map descriptors may be a part of the quantized data (a part of the compressed data) of the S 3D map descriptors, which is dequantized to obtain the S dequantized data, and then It is obtained according to a part of binarized data (a part of compressed data) of S dequantized data and S 3D map descriptors.
- the m-th stage decompression may also include other decompression processing such as entropy decoding and prediction, and this embodiment of the present application does not illustrate one by one.
- the following takes the first reconstructed data as the mth reconstructed data as an example for explanation.
- the i-th level retrieval is performed in the mth reconstructed data of the S 3D map descriptors to obtain P 3D map descriptors.
- Retrieving part or all of the data of the descriptor refers to retrieving part or all of the components of the descriptor.
- the component of the retrieval descriptor refers to the value of a dimension in the vector of the retrieval descriptor.
- the retrieval descriptor may be a vector, for example, an M-dimensional vector, each of which is a component of the vector.
- the execution subject of this embodiment can use part or all of the data of the search descriptor to search the mth reconstructed data of the S 3D map descriptors to obtain the most similar or most similar to the mth reconstructed data of the search descriptor The most relevant P 3D map descriptors.
- P P ⁇ S.
- a 3D map descriptor set is obtained through the i-level retrieval, and the number of 3D map descriptors in the 3D map descriptor set is less than the 3D map descriptions in the 3D map description subset before the i-level retrieval the number of children.
- P may be greater than S.
- the second distance may include, but is not limited to, Euclidean distance, inner product distance, cosine distance, Manhattan distance, and the like. For example, the smaller the Euclidean distance, the higher the correlation, and the larger the inner product distance, the higher the correlation.
- the decompression process of the compressed data of the S 3D map descriptors includes the mth level Decompression and decompression at the kth stage, or, the decompression process of the compressed data of the S 3D map descriptors includes the kth stage of decompression, and the kth stage of decompression includes the mth stage of decompression.
- the second reconstructed data may refer to any one of the second reconstructed data, the third reconstructed data, . . . the kth reconstructed data, k is a positive integer, and m ⁇ k. It should be understood that the "second" and "third" in the second reconstructed data, the third reconstructed data, etc. have no order, but are only for distinguishing each other. The following takes the second reconstructed data as the kth reconstructed data as an example for explanation.
- P 3D map descriptors can be screened from S 3D map descriptors, and then the compressed data of the P 3D map descriptors can be decompressed at the kth level to obtain the kth of the P 3D map descriptors Restructure the data to further perform retrieval based on the kth reconstructed data.
- k m+1.
- the k-th reconstructed data of the P 3D map descriptors may be obtained by decompressing part or all of the compressed data of the P 3D map descriptors at the k-th stage.
- both the i-level retrieval and the j-level retrieval adopt the retrieval method based on the second distance
- the P 3D map descriptors belong to the S 3D map descriptors
- the P 3D map descriptors The distortion degree of the mth reconstructed data of each 3D map descriptor of P is different from the distortion degree of the kth reconstructed data of each 3D map descriptor of P 3D map descriptors, each 3D map of P 3D map descriptors
- the distortion degree of the mth reconstruction data of the descriptor is used to represent the degree of difference between the mth reconstruction data of each 3D map descriptor and the corresponding original 3D map descriptor, each 3D map description of P 3D map descriptors
- the distortion degree of the kth reconstructed data of each sub is used to represent the degree of difference between the kth reconstructed data of each 3D map descriptor and the corresponding original 3D map descriptor, wherein, P ⁇ S
- the distortion degree of the mth reconstruction data of each 3D map descriptor of the P 3D map descriptors is greater than the distortion degree of the kth reconstruction data of each 3D map descriptor of the P 3D map descriptors .
- the distortion degree of the latter level of retrieval results in at least two levels of retrieval is smaller and smaller, thereby improving the accuracy of retrieval results.
- P 3D map descriptors belong to S 3D map descriptors.
- the 3D map descriptor The compressed data includes first quantized data and second quantized data, the first reconstructed data of the 3D map descriptor is obtained by decompressing the first quantized data at the first stage, and the first reconstructed data of the 3D map descriptor can be Including reconstructed data of partial components of the original vector of the 3D map descriptor, the second reconstructed data of the 3D map descriptor is obtained by decompressing the second quantized data at the second stage, the second reconstructed data of the 3D map descriptor
- the compressed data includes first quantized data and second quantized data. Unlike the previous example, the second quantized data here can be obtained after quantizing the residual of the first quantized data.
- the first quantized data of the 3D map descriptor The reconstructed data is obtained by decompressing the first quantized data at the first level, and the first reconstructed data of the 3D map descriptor may include the first reconstructed data of the original vector of the 3D map descriptor (reconstructed data with lower precision structural data), the second reconstructed data of the 3D map descriptor is obtained by decompressing the second quantized data at the second stage, and the second reconstructed data of the 3D map descriptor may include the above-mentioned residual of the 3D map descriptor
- the following takes the second reconstructed data as the kth reconstructed data as an example for explanation.
- the j-th level retrieval is performed in the k-th reconstructed data of the P 3D map descriptors to obtain Q 3D map descriptors.
- S, P, Q, i, j, k, and m are positive integers, and 0 ⁇ Q ⁇ P, 0 ⁇ P ⁇ T, 0 ⁇ S ⁇ T, T represents the total number of 3D map descriptors in the 3D map
- condition judgment step can be performed after S104C. If the condition is met, the search in this embodiment can be ended, and if the condition is not met, one or more levels of search can be continued.
- This condition can be any condition, which can be reasonably set according to requirements.
- the condition can be that the similarity or correlation is higher than or equal to the preset threshold, or that the number of retrieved 3D map descriptors is less than or equal to the preset threshold. Set the number, etc., the embodiment of the present application does not give examples one by one.
- the retrieval method of the 3D map in the embodiment of the present application is a fragment process, which may also include other levels of retrieval before and/or after the fragment process.
- the retrieval method of the 3D map may include more stages than those shown in FIG. 5C
- the multi-level search for example, 3, 4 or 5 levels and so on. For example, there may be more levels of retrieval before S101C in the embodiment shown in FIG. 5C , or there may be more levels of retrieval after S104C in the embodiment shown in FIG. 5C , which are not illustrated here.
- Q 3D map descriptors obtained through multi-level retrieval can be used for positioning. If the 3D map points corresponding to the N 3D map descriptors in the Q 3D map descriptors match the 3D map points corresponding to the retrieval descriptors, then positioning can be performed according to the 3D map points corresponding to the N 3D map descriptors to obtain Pose information of electronic devices.
- the electronic device may be an electronic device that collects the aforementioned visual information. 0 ⁇ N ⁇ Q.
- the compressed data of S 3D map descriptors is decompressed at the mth stage to obtain the first reconstructed data of S 3D map descriptors.
- the i-th level search is performed on the first reconstructed data of the map descriptor, and P 3D map descriptors are screened out.
- Decompress the compressed data of the P 3D map descriptors at the kth level to obtain the second reconstructed data of the P 3D map descriptors.
- the j-th level retrieval is performed in the second reconstructed data to filter out a smaller number of 3D map descriptors.
- a kind of reconstructed data of the 3D map descriptor obtained by decompression which is alternately performed with the first-level retrieval.
- one level of retrieval is staged decompression, which can improve the retrieval speed, and the decompression differential settings of at least two levels of retrieval can ensure retrieval accuracy.
- the degree of decompression or the degree of distortion of the reconstructed data of the 3D map descriptor used by any two levels of retrieval are different. Compared with retrieving in the reconstructed data of the fully decompressed 3D map, the retrieval speed can be improved and the retrieval accuracy can be guaranteed.
- FIG. 5D is a schematic flowchart of a fragment of a method for retrieving a 3D map provided by an embodiment of the present application.
- the method can be applied to any of the electronic devices illustrated in FIG. 1 to FIG. 4f, or also It can be applied to the server shown in any one of Fig. 1 to Fig. 4f.
- the method includes but is not limited to the following steps:
- S101D Decompress the compressed data of the S 3D map descriptors at the mth level to obtain the reconstructed data of the S 3D map descriptors, and the S 3D map descriptors correspond to multiple 3D map points in the 3D map.
- the reconstructed data may refer to any one of the first reconstructed data, the second reconstructed data, ... the mth reconstructed data, etc., m is a positive integer. It should be understood that the "first”, “second” and “third” in the first reconstructed data, the second reconstructed data and the third reconstructed data, etc., have no order, but are only for distinguishing each other . The following takes the reconstructed data as the mth reconstructed data as an example for explanation.
- the S 3D map descriptors correspond to multiple 3D map points in the 3D map.
- S103D Extract binarized data of P 3D map descriptors from the compressed data of P 3D map descriptors.
- the compressed data of S 3D map descriptors is decompressed at the mth level to obtain the reconstructed data of S 3D map descriptors, and according to the partial data or all data of the retrieval descriptors, the S 3D map descriptions
- the i-th level retrieval is performed in the reconstructed data of the child, and P 3D map descriptors are screened out. Extract the binarized data of the P descriptors from the compressed data of the P 3D map descriptors, and perform jth in the binarized data of the P 3D map descriptors according to the binarized data of the retrieved descriptors.
- Extraction or decompression from compressed data, and retrieval are performed alternately to perform multi-level retrieval in the 3D map to obtain a 3D map descriptor that can be used for positioning.
- one level of retrieval is decompression, which can improve retrieval accuracy
- one level of retrieval is extraction from compressed data, which can improve retrieval speed.
- the retrieval method of the 3D map provided by the embodiment of the present application can improve the retrieval speed and ensure the retrieval accuracy.
- retrieval method flows shown in FIG. 5A to FIG. 5C above can be flexibly combined with each other to form various multi-level retrieval methods.
- FIG. 6 is a schematic flowchart of a method for retrieving a 3D map provided by an embodiment of the present application.
- the method can be applied to any of the electronic devices illustrated in FIG. 1 to FIG. 4f, or can also be Applied to the server shown in any one of Fig. 1 to Fig. 4f.
- This embodiment takes multi-level retrieval as an example of 4-level retrieval for illustration, wherein, the first-level retrieval and the second-level retrieval adopt the retrieval method based on the first distance, and the third-level retrieval and the fourth-level retrieval adopt the retrieval method based on the second distance. How to retrieve the distance.
- the method includes but is not limited to the following steps:
- the retrieval descriptor is a feature corresponding to the real environment extracted from the visual information collected by the sensor of the electronic device.
- the S 1 3D map descriptors may be S 1 region descriptors, or S 1 3D map point descriptors. Wherein, one area descriptor corresponds to multiple 3D map points, and its specific explanation can refer to the explanation of the foregoing embodiments, and will not be repeated here.
- the first binarized data of the retrieval descriptor may be partial binarized data of the retrieval descriptor.
- the execution subject of this embodiment can perform binarization processing on the retrieval descriptor to obtain the binary data of the retrieval descriptor, and select a part from the binary data of the retrieval descriptor as the first part of the retrieval descriptor. Binarize data.
- P 1 is smaller than S 1 . It can also be understood that a subset is obtained through retrieval, and the number of 3D map descriptors in this subset is less than the number of 3D map descriptors in the 3D map descriptor set before retrieval.
- the bit overhead of the second binarized data in this embodiment may be greater than the bit overhead of the first binarized data, so that the retrieval speed of the previous stage of retrieval can be improved, and the accuracy of the subsequent stage of retrieval can be guaranteed.
- the execution subject of this embodiment can use the second binarized data of the search descriptor to perform a search based on the first distance among the second binarized data of P 1 3D map descriptors, so as to obtain the The Q 1 D map descriptors most similar or most relevant to the second binarized data.
- Q 1 is less than or equal to P 1 . It can also be understood that another subset is obtained through retrieval, and the number of 3D map descriptors in this subset is less than the number of 3D map descriptors in the 3D map descriptor subset retrieved in step 203 .
- Q 2 is less than or equal to Q 1 .
- another subset is obtained through retrieval, and the number of 3D map descriptors in this subset is less than the number of 3D map descriptors in the 3D map descriptor subset retrieved in step 205 .
- the retrieval based on the second distance in the third stage can improve the retrieval accuracy.
- Q 3 is less than or equal to Q 2 . It can also be understood that another subset is obtained through retrieval, and the number of 3D map descriptors in this subset is less than the number of 3D map descriptors in the 3D map descriptor subset retrieved in step 207 .
- the binarized data of the 3D map descriptor is extracted from the compressed data or the reconstruction data of the 3D map descriptor is obtained by decompression, which is alternately performed with the first-level search, so as to perform multi-level search in the 3D map, and obtain A 3D map descriptor that can be used for localization.
- the retrieval speed can be improved and the retrieval accuracy can be guaranteed.
- FIG. 7 is a schematic flowchart of a method for retrieving a 3D map provided by an embodiment of the present application.
- the method can be applied to any of the electronic devices illustrated in FIG. 1 to FIG. 4f, or can also be Applied to the server shown in any one of Fig. 1 to Fig. 4f.
- This embodiment is illustrated by taking the S 3D map descriptors in the embodiment shown in FIG. 5A as S representative 3D map descriptors.
- the S representative 3D map descriptors each correspond to at least one data set.
- the specific possible ways of 102A in the above-mentioned FIG. 5A may include but not limited to the following steps:
- At least one representative 3D map descriptor corresponds to at least one data set, and each of the at least one data set includes compressed data of a plurality of 3D map descriptors.
- the 3D map descriptors in each data set have certain correlation or similarity.
- S represent the binarized data of the 3D map descriptor
- the binarized data of the retrieval descriptor may be part or all of the binarized data of the retrieval descriptor.
- the execution subject of this embodiment can use the binarized data of the search descriptor to perform a search based on the first distance among S binarized data representing 3D map descriptors, so as to obtain the binarized data corresponding to the search descriptor At least one of the most similar or most related represents a 3D map descriptor.
- the number of at least one representative 3D map descriptor is less than or equal to S.
- At least one data set is obtained by retrieving at least one representative 3D map descriptor, so as to obtain compressed data of P 3D map descriptors in the at least one data set.
- the S representative 3D map descriptors are T 1 , T 2 and T 3 , T 1 corresponds to data set 1 , T 2 corresponds to data set 2 , and T 3 corresponds to data set 3 .
- At least one representative 3D map descriptor obtained by searching is T 3 , thereby obtaining a data set 3 , which may include compressed data of P 3D map descriptors.
- one-level retrieval is performed on a small number of representative 3D map descriptors to improve the retrieval speed, and then one-level or multi-level retrieval can be performed in the data set corresponding to the representative 3D map descriptor to improve retrieval accuracy Rate.
- the retrieval speed can be improved and the retrieval accuracy can be guaranteed.
- S 3D map descriptors in the embodiment shown in FIG. 5C to FIG. 5D are S representative 3D map descriptors, which is similar to the above-mentioned embodiment shown in FIG. 7 .
- S representative 3D map descriptors which is similar to the above-mentioned embodiment shown in FIG. 7 .
- FIG. 8 is a schematic flowchart of a method for retrieving a 3D map provided by an embodiment of the present application.
- FIG. 9 is a schematic diagram of a processing procedure of a method for retrieving a 3D map provided by an embodiment of the present application. This embodiment includes a server and electronic equipment. The method includes but is not limited to the following steps:
- the server determines the size relationship between each component of the S 3D map descriptors and the corresponding component of the preset threshold vector according to each component of the S 3D map descriptors and the corresponding component of the preset threshold vector.
- the server performs binarization processing on the above size relationship to obtain binarized data of each of the S 3D map descriptors.
- the server quantifies the absolute value of the difference between each component of the S 3D map descriptors and the corresponding component of the preset threshold vector, to obtain quantized data of each of the S 3D map descriptors.
- each component included in the preset threshold vector is an arbitrary value.
- the server encapsulates the binarized data and quantized data of each of the S 3D map descriptors to obtain a code stream of the 3D map.
- the server sends the code stream of the 3D map to the electronic device.
- the electronic device acquires the retrieval descriptor.
- the electronic device may extract the retrieval descriptor from the visual information collected by the sensor.
- the electronic device decapsulates the code stream of the 3D map to obtain the respective binarized data and quantized data of the S 3D map descriptors, and extract the respective binarized data of the S 3D map descriptors.
- the electronic device can decode the code stream of the 3D map Encapsulate to obtain the binarized data and quantized data of each of the five 3D map descriptors as shown in FIG. 9 .
- the electronic device performs a first-level search in the respective binary data of the S 3D map descriptors according to the binarized data of the search descriptors, so as to obtain P 3D map descriptors.
- the binarized data of the retrieval descriptor is obtained in the following manner: the electronic device can determine the relationship between each component of the retrieval descriptor and the corresponding component of the preset threshold vector according to each component of the retrieval descriptor and the corresponding component of the preset threshold vector. size relationship. The size relationship is binarized to obtain the binarized data of the retrieval descriptor.
- a first-distance-based retrieval method may be used to perform the first-level retrieval in the binarized data of each of the S 3D map descriptors.
- the first-level retrieval is performed in five 3D map descriptors (T 1 , T 2 , T 3 , T 4 and T 5 ).
- the Hamming distance between the binarized data of five 3D map descriptors (T 1 , T 2 , T 3 , T 4 and T 5 ) and the binarized data of the retrieval descriptor can be calculated respectively, according to Hamming distance filters out 3D map descriptors.
- it is taken as an example to filter out two 3D map descriptors, that is, to filter out T 1 and T 2 shown in the dotted line box in FIG. 9 . That is, P 2.
- the electronic device decompresses the respective quantized data of the P 3D map descriptors to obtain the respective reconstruction data of the P 3D map descriptors.
- the electronic device may perform dequantization processing on respective quantized data of the P 3D map descriptors to obtain respective dequantized data of the P 3D map descriptors. According to the respective inverse quantization data and binarized data of the P 3D map descriptors, respective reconstruction data of the P 3D map descriptors are obtained.
- the electronic device performs a second-level search in the reconstructed data of the P 3D map descriptors, so as to obtain the Q 3D map descriptors.
- a second distance-based retrieval method can be collected for the second-level retrieval in the respective reconstructed data of the P 3D map descriptors.
- the second-level retrieval is performed in T1 and T2 according to all components of the retrieval descriptor.
- the Euclidean distance between the reconstructed data of T 1 and T 2 and the retrieval descriptor can be calculated respectively, and the 3D map descriptor can be screened out according to the Euclidean distance.
- it is taken as an example to filter out one 3D map descriptor, that is, to filter out T 1 as shown in FIG. 9 . That is, Q 1.
- the server compresses the S 3D map descriptors of the 3D map to obtain the binarized data and quantized data of each of the S 3D map descriptors, so as to reduce the resource overhead required for transmitting the 3D map.
- the method for retrieving a 3D map according to the embodiment of the present application is described in detail above with reference to the accompanying drawings, and the apparatus for retrieving the 3D map according to the embodiment of the present application will be introduced below with reference to FIG. 10 and FIG. 11 . It should be understood that the apparatus for retrieving a 3D map can execute the method for retrieving a 3D map in the embodiment of the present application. In order to avoid unnecessary repetition, repeated descriptions are appropriately omitted when introducing the 3D map retrieval apparatus according to the embodiment of the present application below.
- FIG. 10 is a schematic structural diagram of a 3D map retrieval device provided by an embodiment of the present application.
- the 3D map retrieval apparatus 1000 may include: a retrieval module 1001 , an extraction module 1002 and a decompression module 1003 .
- the extraction module 1002 is configured to extract binarized data of S 3D map descriptors from compressed data of S 3D map descriptors, and the S 3D map descriptors correspond to 3D A plurality of 3D map points in the map; a retrieval module 1001, configured to perform i-level retrieval in the binarized data of the S 3D map descriptors according to the binarized data of the retrieved descriptors, so as to obtain P 3D map descriptor; the retrieval descriptor is a feature corresponding to the real environment extracted from the visual information collected by the sensor of the electronic device; the decompression module 1003 is used to perform compression on the compressed data of the P 3D map descriptors decompressing at the mth level to obtain the reconstructed data of the P 3D map descriptors; the decompression process of the compressed data of the P 3D map descriptors at least includes the mth level decompression; the retrieval module 1001 , for performing
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are the 3D map points corresponding to the retrieval descriptors Matching, N is a positive integer, 0 ⁇ N ⁇ Q, the device also includes: a pose determination module; the pose determination module performs positioning according to the 3D map points corresponding to the N 3D map descriptors to obtain The pose information of the electronic device.
- the retrieval method adopted by the i-th level retrieval is a retrieval method based on the first distance
- the retrieval method adopted by the j-th level retrieval is a retrieval method based on the second distance.
- the device further includes: an acquisition module, configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor or, receiving the visual information, extracting the retrieval descriptor from the visual information, and performing binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor; or, In response to the operation of collecting visual information input by the user, the sensor is triggered to collect visual information from the real environment to obtain the visual information, and the retrieval descriptor is extracted from the visual information, and the retrieval descriptor is Perform binarization processing to obtain binarized data of the retrieval descriptor.
- the decompression module 1003 is further configured to: perform k-level decompression on the compressed data of the Q 3D map descriptors to obtain Q 3D map descriptions Reconstructed data of the child; according to part or all of the data of the search descriptor, perform r-th level retrieval in the reconstructed data of the Q 3D map descriptors to obtain the N 3D map descriptors;
- N ⁇ Q ⁇ P the decompression process of the compressed data of the P 3D map descriptors includes the m-level decompression and the k-level decompression, r and k are positive integers, m ⁇ k , j ⁇ r ⁇ L.
- the compressed data of the P 3D map descriptors includes respective binarization data and quantized data of the P 3D map descriptors
- the decompression module 1003 is specifically configured to: process the P 3D maps Dequantize the respective quantized data of the descriptors to obtain P dequantized data, and the P dequantized data are used as reconstruction data of the P 3D map descriptors; each of the Q 3D map descriptors Quantized data of the quantized data is dequantized, and Q dequantized data are obtained; according to the Q dequantized data and the binarized data of the Q 3D map descriptors, the Q 3D map descriptors are respectively obtained. of reconstructed data.
- the compressed data of the P 3D map descriptors includes the respective binarized data and quantized data of the P 3D map descriptors, and the decompression module 1003, specifically It is used for: performing dequantization processing on the respective quantized data of the P 3D map descriptors to obtain P dequantized data; according to the respective binary values of the P dequantized data and the P 3D map descriptors to obtain the respective reconstruction data of the P 3D map descriptors.
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data set of the at least one data set Including at least one 3D map descriptor, the retrieval module 1001 is used for: according to the binarization data of the retrieval descriptor, perform i-level retrieval in the S binarized data representing 3D map descriptors, to obtain at least one representative 3D map descriptor; using the 3D map descriptors in the data sets corresponding to the at least one representative 3D map descriptor as the P 3D map descriptors.
- the extraction module 1002 is configured to extract the first binarized data of the S 3D map descriptors from the compressed data of the S 3D map descriptors, and the S 3D map descriptors
- the descriptor corresponds to a plurality of 3D map points in the 3D map
- the retrieval module 1001 is configured to, according to the first binarized data of the retrieved descriptor, perform the second binarized data of the S 3D map descriptors.
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are the 3D map points corresponding to the retrieval descriptors Matching, N is a positive integer, 0 ⁇ N ⁇ Q, the device further includes: a pose determination module; the pose determination module is used to perform positioning according to the 3D map points corresponding to the N 3D map descriptors, to obtain the pose information of the electronic device.
- the i-th level retrieval and the j-th level retrieval adopt a retrieval method based on the first distance
- the P 3D map descriptors belong to the S 3D map descriptors
- the position of the first binarization data of each 3D map descriptor of the P 3D map descriptors in the compressed data of the 3D map descriptor is different from the second binarization of the 3D map descriptor The position of the data in the compressed data of the 3D map descriptor, where P ⁇ S.
- the length of the first binarized data of each 3D map descriptor of the P 3D map descriptors is smaller than the second binary value of each 3D map descriptor of the P 3D map descriptors the length of the optimized data.
- the device further includes: an acquisition module, configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain a first binary value of the retrieval descriptor Or, receive the visual information, extract the retrieval descriptor from the visual information, and perform binarization processing on the retrieval descriptor to obtain the retrieval description The first binarized data and the second binarized data; or, in response to the operation of collecting visual information input by the user, triggering the sensor to collect visual information on the real environment to obtain the visual information, from The retrieval descriptor is extracted from the visual information, and the retrieval descriptor is binarized to obtain first binarized data and second binarized data of the retrieval descriptor.
- an acquisition module configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain a first binary value of the retrieval descriptor
- receive the visual information extract the retrieval descriptor from the visual information, and perform binarization processing on the retrieval descriptor to
- the length of the first binarized data of the retrieval descriptor is equal to the length of the first binarized data of each of the S 3D map descriptors, and/or, The length of the second binarized data of the retrieval descriptor is equal to the length of the second binarized data of each 3D map descriptor of the S 3D map descriptors.
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data set of the at least one data set It includes at least one 3D map descriptor, and the retrieval module 1001 is specifically configured to: according to the first binarized data of the retrieved descriptor, perform Searching at the i-level to obtain at least one representative 3D map descriptor; using the 3D map descriptors in the data sets corresponding to the at least one representative 3D map descriptor as the P 3D map descriptors.
- the decompression module 1003 is configured to decompress the compressed data of the S 3D map descriptors at the mth level to obtain the first reconstructed data of the S 3D map descriptors, so The S 3D map descriptors correspond to a plurality of 3D map points in the 3D map; the retrieval module 1001 is configured to reconstruct data in the first reconstructed data of the S 3D map descriptors according to part or all of the data of the retrieval descriptors
- the i-th level search is performed; the search descriptor is extracted from the visual information collected by the sensor of the electronic device and corresponds to the feature of the real environment; the decompression module 1003 also uses performing k-level decompression on the compressed data of the P 3D map descriptors to obtain the second reconstructed data of the P 3D map descriptors; decompressing the compressed data of the S 3D map descriptors
- the process includes the
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are the 3D map points corresponding to the retrieval descriptors Matching, N is a positive integer, 0 ⁇ N ⁇ Q, the device further includes: a pose determination module; the pose determination module is used to perform positioning according to the 3D map points corresponding to the N 3D map descriptors, to obtain the pose information of the electronic device.
- the i-th level retrieval and the j-th level retrieval adopt a retrieval method based on the second distance
- the P 3D map descriptors belong to the S 3D map descriptions
- the distortion degree of the first reconstruction data of each 3D map descriptor of the P 3D map descriptors, and the distortion degree of the second reconstruction data of each 3D map descriptor of the P 3D map descriptors Differently, the distortion degree of the first reconstructed data of each 3D map descriptor of the P 3D map descriptors is used to represent the difference between the first reconstructed data of each 3D map descriptor and the corresponding original 3D map descriptor
- the degree of difference between the P 3D map descriptors, the distortion degree of the second reconstruction data of each 3D map descriptor is used to represent the second reconstruction data of each 3D map descriptor and the corresponding original 3D map
- the degree of distortion of the first reconstructed data of each 3D map descriptor of the P 3D map descriptors is greater than the second reconstruction of each 3D map descriptor of the P 3D map descriptors Data distortion.
- the device further includes: an obtaining module, configured to: receive the retrieval descriptor, and obtain part or all of the data of the retrieval descriptor; or, receive the visual information, and obtain from the extracting the retrieval descriptor from the visual information, and obtaining part or all of the data of the retrieval descriptor; or, in response to the operation of visual information collection input by the user, triggering the sensor to collect visual information from the real environment , to obtain the visual information, extract the retrieval descriptor from the visual information, and acquire part or all of the data of the retrieval descriptor.
- an obtaining module configured to: receive the retrieval descriptor, and obtain part or all of the data of the retrieval descriptor; or, receive the visual information, and obtain from the extracting the retrieval descriptor from the visual information, and obtaining part or all of the data of the retrieval descriptor; or, in response to the operation of visual information collection input by the user, triggering the sensor to collect visual information from the real environment , to obtain the visual information, extract the
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data set of the at least one data set Including at least one 3D map descriptor, the retrieval module 1001 is specifically configured to: perform the i-th reconstruction data in the m-th reconstructed data of the S representative 3D map descriptors according to part or all of the data of the retrieval descriptor. Level retrieval to obtain at least one representative 3D map descriptor; using the 3D map descriptors in the data sets corresponding to the at least one representative 3D map descriptor as the P 3D map descriptors.
- the decompression module 1003 is configured to decompress the compressed data of S 3D map descriptors at the mth level to obtain reconstructed data of S 3D map descriptors, and the S The 3D map descriptor corresponds to multiple 3D map points in the 3D map; the retrieval module 1001 is configured to perform the i-th level in the reconstructed data of the S 3D map descriptors according to part or all of the data of the retrieval descriptor.
- N 3D map descriptors among the Q 3D map descriptors are used for positioning, and the 3D map points corresponding to the N 3D map descriptors are the 3D map points corresponding to the retrieval descriptors Matching, N is a positive integer, 0 ⁇ N ⁇ Q, the device further includes: a pose determination module; the pose determination module is used to perform positioning according to the 3D map points corresponding to the N 3D map descriptors, to obtain the pose information of the electronic device.
- the retrieval method adopted by the i-th level retrieval is a retrieval method based on the second distance
- the retrieval method adopted by the j-th level retrieval is a retrieval method based on the first distance
- the device further includes: an acquisition module, configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor or, receiving the visual information, extracting the retrieval descriptor from the visual information, and performing binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor; or, Responding to the operation of collecting visual information input by the user, triggering the sensor to collect visual information on the real environment, obtaining the visual information, extracting the retrieval descriptor from the visual information, and describing the retrieval perform binarization processing to obtain the binarized data of the retrieval descriptor.
- an acquisition module configured to: receive the retrieval descriptor, and perform binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descriptor or, receiving the visual information, extracting the retrieval descriptor from the visual information, and performing binarization processing on the retrieval descriptor to obtain binarized data of the retrieval descript
- the S 3D map descriptors are S representative 3D map descriptors, and each of the S representative 3D map descriptors corresponds to at least one data set, and each data set of the at least one data set Including at least one 3D map descriptor, the retrieval module 1001 is specifically configured to: perform the i-th reconstruction data in the m-th reconstructed data of the S representative 3D map descriptors according to part or all of the data of the retrieval descriptor. Level retrieval to obtain at least one representative 3D map descriptor; using the 3D map descriptors in the data sets corresponding to the at least one representative 3D map descriptor as the P 3D map descriptors.
- the 3D map retrieval device 1000 can execute any one of FIG. 5A to FIG. 5D or the 3D map retrieval device method of the embodiment shown in FIG. 6 or FIG. 7 , or execute S406- Relevant content of S410, or execute relevant content of the embodiment shown in FIG. 9 .
- FIG. 5A to FIG. 5D the 3D map retrieval device method of the embodiment shown in FIG. 6 or FIG. 7
- S406- Relevant content of S410 or execute relevant content of the embodiment shown in FIG. 9 .
- Fig. 11 is a schematic block diagram of an implementation manner of a decoding device 1100 used in the embodiment of the present application.
- the decoding device 1100 may include a processor 1101 , a memory 1102 and a bus system 1103 .
- the processor 1101 and the memory 1102 are connected through the bus system 1103, the memory 1102 is used to store instructions, and the processor 1101 is used to execute the instructions stored in the memory 1102 to execute various 3D map retrieval methods described in this application. To avoid repetition, no detailed description is given here.
- the processor 1101 may be a central processing unit (central processing unit, CPU), and the processor 1101 may also be other general-purpose processors, DSP, ASIC, FPGA or other programmable logic devices, discrete gates or Transistor logic devices, discrete hardware components, and more.
- a general-purpose processor may be a microprocessor, or the processor may be any conventional processor, or the like.
- the memory 1102 may include a ROM device or a RAM device. Any other suitable type of storage device may also be used as memory 1102 .
- Memory 1102 may include code and data 11021 accessed by processor 1101 using bus 1103 .
- the memory 1102 may further include an operating system 11023 and an application program 11022, where the application program 11022 includes at least one program that allows the processor 1101 to execute the 3D map retrieval method described in this application.
- the application program 11022 may include applications 1 to N, which further include a 3D map application that executes the 3D map retrieval method described in this application.
- bus system 1103 may also include a power bus, a control bus, and a status signal bus.
- bus system 1103 may also include a power bus, a control bus, and a status signal bus.
- bus system 1103 may also include a power bus, a control bus, and a status signal bus.
- bus system 1103 may also include a power bus, a control bus, and a status signal bus.
- bus system 1103 may also include a power bus, a control bus, and a status signal bus.
- the decoding apparatus 1100 may further include one or more output devices, such as a display 1104 .
- display 1104 may be a touch-sensitive display that incorporates a display with a haptic unit operable to sense touch input.
- the display 1104 may be connected to the processor 1101 via the bus 1103 .
- the decoding device 1100 can implement the 3D map retrieval method in this application.
- Computer-readable media may include computer-readable storage media, which correspond to tangible media, such as data storage media, or communication media including any medium that facilitates transfer of a computer program from one place to another (eg, according to a communication protocol) .
- a computer-readable medium may generally correspond to (1) a non-transitory tangible computer-readable storage medium, or (2) a communication medium, such as a signal or carrier wave.
- Data storage media may be any available media that can be accessed by one or more computers or one or more processors to retrieve instructions, code and/or data structures for implementation of the techniques described in this application.
- a computer program product may include a computer readable medium.
- such computer-readable storage media may include RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage, flash memory, or any other medium that can contain the desired program code in the form of a computer and can be accessed by a computer.
- any connection is properly termed a computer-readable medium.
- coaxial cable, fiber optic cable, twisted pair, digital subscriber line (DSL), or wireless technologies such as infrared, radio, and microwave
- coaxial cable Wire, fiber optic cable, twisted pair, DSL, or wireless technologies such as infrared, radio, and microwave are included in the definition of media.
- Disk and disc includes compact disc (CD), laser disc, optical disc, digital versatile disc (DVD) and blu-ray disc where disks usually reproduce data magnetically, while discs reproduce optically with lasers data. Combinations of the above should also be included within the scope of computer-readable media.
- DSPs digital signal processors
- ASICs application specific integrated circuits
- FPGAs field programmable logic arrays
- DSPs digital signal processors
- ASICs application specific integrated circuits
- FPGAs field programmable logic arrays
- the techniques of the present application may be implemented in a wide variety of devices or devices, including wireless handsets, an integrated circuit (IC), or a group of ICs (eg, a chipset).
- IC integrated circuit
- a group of ICs eg, a chipset
- Various components, modules, or units are described in this application to emphasize functional aspects of means for performing the disclosed techniques, but do not necessarily require realization by different hardware units. Indeed, as described above, the various units may be combined in a codec hardware unit in conjunction with suitable software and/or firmware, or by interoperating hardware units (comprising one or more processors as described above) to supply.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Remote Sensing (AREA)
- Multimedia (AREA)
- Radar, Positioning & Navigation (AREA)
- Software Systems (AREA)
- Geometry (AREA)
- Databases & Information Systems (AREA)
- Signal Processing (AREA)
- Computer Graphics (AREA)
- Library & Information Science (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Networks & Wireless Communication (AREA)
- Processing Or Creating Images (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Image Processing (AREA)
Abstract
Description
Claims (18)
- 一种3D地图的检索方法,其特征在于,所述方法包括:从S个3D地图描述子的压缩数据中提取S个3D地图描述子的二值化数据,所述S个3D地图描述子对应3D地图中的多个3D地图点;根据检索描述子的二值化数据,在所述S个3D地图描述子的二值化数据中进行第i级检索,以得到P个3D地图描述子;所述检索描述子是从电子设备的传感器采集的视觉信息所提取出对应于真实环境的特征;对所述P个3D地图描述子的压缩数据进行第m级解压缩,得到所述P个3D地图描述子的重构数据;所述P个3D地图描述子的压缩数据的解压缩过程至少包括所述第m级解压缩;根据所述检索描述子的部分数据或全部数据,在所述P个3D地图描述子的重构数据中进行第j级检索,以得到Q个3D地图描述子,S、P、Q、i、j和m为正整数,且0<Q<P,0<P<T,0<S≤T,T表示所述3D地图中的3D地图描述子的总个数,j=i+1,1≤i<L,1<j≤L,L表示所述3D地图的检索的总级数或者检索级数阈值,L为大于1的正整数。
- 根据权利要求1所述的方法,其特征在于,所述Q个3D地图描述子中的N个3D地图描述子用于定位,所述N个3D地图描述子对应的3D地图点与所述检索描述子对应的3D地图点匹配,N为正整数,0<N≤Q,所述方法还包括:根据所述N个3D地图描述子对应的3D地图点进行定位,以得到所述电子设备的位姿信息。
- 根据权利要求1或2所述的方法,其特征在于,所述第i级检索所采用的检索方式为基于第一距离的检索方式,所述第j级检索所采用的检索方式为基于第二距离的检索方式。
- 根据权利要求1至3任一项所述的方法,其特征在于,所述方法还包括:接收所述检索描述子,并对所述检索描述子进行二值化处理,以得到所述检索描述子的二值化数据;或者,接收所述视觉信息,从所述视觉信息中提取所述检索描述子,并对所述检索描述子进行二值化处理,以得到所述检索描述子的二值化数据;或者,响应于用户输入的视觉信息采集的操作,触发所述传感器对所述真实环境进行视觉信息采集,得到所述视觉信息,从所述视觉信息中提取所述检索描述子,对所述检索描述子进行二值化处理,以得到所述检索描述子的二值化数据。
- 根据权利要求2所述的方法,其特征在于,在N<Q的情况下,所述方法还包括:对所述Q个3D地图描述子的压缩数据进行第k级解压缩,以得到Q个3D地图描述子的重构数据;根据所述检索描述子的部分数据或全部数据,在所述Q个3D地图描述子的重构数据中进行第r级检索,以得到所述N个3D地图描述子;其中,N<Q<P,所述P个3D地图描述子的压缩数据的解压缩过程包括所述第m级解压缩和所述第k级解压缩,r和k为正整数,m<k,j<r≤L。
- 根据权利要求5所述的方法,其特征在于,所述P个3D地图描述子的压缩数据包括P个3D地图描述子各自的二值化数据和量化数据,所述对所述P个3D地图描述子的压缩数据进行第m级解压缩得到所述P个3D地图描述子的重构数据,包括:对所述P个3D地图描述子各自的量化数据进行反量化处理,得到P个反量化数据,所述P个反量化数据用作所述P个3D地图描述子的重构数据;所述对所述Q个3D地图描述子的压缩数据进行第k级解压缩得到所述Q个3D地图描述子的重构数据,包括:对所述Q个3D地图描述子各自的量化数据,进行反量化处理,得到Q个反量化数据;根据所述Q个反量化数据和所述Q个3D地图描述子各自的二值化数据,得到所述Q个3D地图描述子各自的重构数据。
- 根据权利要求2所述的方法,其特征在于,在N=Q的情况下,所述P个3D地图描述子的压缩数据包括P个3D地图描述子各自的二值化数据和量化数据,所述对所述P个3D地图描述子的压缩数据进行第m级解压缩得到所述P个3D地图描述子的重构数据,包括:对所述P个3D地图描述子各自的量化数据,进行反量化处理,得到P个反量化数据;根据所述P个反量化数据和所述P个3D地图描述子各自的二值化数据,得到所述P个3D地图描述子各自的重构数据。
- 根据权利要求1至7任一项所述的方法,其特征在于,所述S个3D地图描述子为S个代表3D地图描述子,所述S个代表3D地图描述子各自分别对应至少一个数据集合,所述至少一个数据集合的各个数据集合包括至少一个3D地图描述子,所述根据检索描述子的二值化数据,在所述S个3D地图描述子的二值化数据中进行第i级检索,以得到P个3D地图描述子,包括:根据所述检索描述子的二值化数据,在所述S个代表3D地图描述子的二值化数据中进行第i级检索,以得到至少一个代表3D地图描述子;将所述至少一个代表3D地图描述子各自对应的数据集合中的3D地图描述子,作为所述P个3D地图描述子。
- 一种3D地图的检索装置,其特征在于,所述装置包括:提取模块,用于从S个3D地图描述子的压缩数据中提取S个3D地图描述子的二值化数据,所述S个3D地图描述子对应3D地图中的多个3D地图点;检索模块,用于根据检索描述子的二值化数据,在所述S个3D地图描述子的二值化数据中进行第i级检索,以得到P个3D地图描述子;所述检索描述子是从电子设备的传感器采集的视觉信息所提取出对应于真实环境的特征;解压缩模块,用于对所述P个3D地图描述子的压缩数据进行第m级解压缩,得到所述P个3D地图描述子的重构数据;所述P个3D地图描述子的压缩数据的解压缩过程至少包括所述第m级解压缩;所述检索模块,用于根据所述检索描述子的部分数据或全部数据,在所述P个3D地图描述子的重构数据中进行第j级检索,以得到Q个3D地图描述子,S、P、Q、i、j和m为正整数,且0<Q<P,0<P<T,0<S≤T,T表示所述3D地图中的3D地图描述子的总个数,j=i+1,1≤i<L,1<j≤L,L表示所述3D地图的检索的总级数或者检索级数阈值,L为大于1的正整数。
- 根据权利要求9所述的装置,其特征在于,所述Q个3D地图描述子中的N个3D地图描述子用于定位,所述N个3D地图描述子对应的3D地图点与所述检索描述子对应的3D地图 点匹配,N为正整数,0<N≤Q,所述装置还包括:位姿确定模块;所述位姿确定模块,根据所述N个3D地图描述子对应的3D地图点进行定位,以得到所述电子设备的位姿信息。
- 根据权利要求9或10所述的装置,其特征在于,所述第i级检索所采用的检索方式为基于第一距离的检索方式,所述第j级检索所采用的检索方式为基于第二距离的检索方式。
- 根据权利要求9至11任一项所述的装置,其特征在于,所述装置还包括:获取模块,用于:接收所述检索描述子,并对所述检索描述子进行二值化处理,以得到所述检索描述子的二值化数据;或者,接收所述视觉信息,从所述视觉信息中提取所述检索描述子,并对所述检索描述子进行二值化处理,以得到所述检索描述子的二值化数据;或者,响应于用户输入的视觉信息采集的操作,触发所述传感器对所述真实环境进行视觉信息采集,得到所述视觉信息,从所述视觉信息中提取所述检索描述子,对所述检索描述子进行二值化处理,以得到所述检索描述子的二值化数据。
- 根据权利要求10所述的装置,其特征在于,在N<Q的情况下,所述解压缩模块还用于:对所述Q个3D地图描述子的压缩数据进行第k级解压缩,以得到Q个3D地图描述子的重构数据;根据所述检索描述子的部分数据或全部数据,在所述Q个3D地图描述子的重构数据中进行第r级检索,以得到所述N个3D地图描述子;其中,N<Q<P,所述P个3D地图描述子的压缩数据的解压缩过程包括所述第m级解压缩和所述第k级解压缩,r和k为正整数,m<k,j<r≤L。
- 根据权利要求13所述的装置,其特征在于,所述P个3D地图描述子的压缩数据包括P个3D地图描述子各自的二值化数据和量化数据,所述解压模块具体用于:对所述P个3D地图描述子各自的量化数据进行反量化处理,得到P个反量化数据,所述P个反量化数据用作所述P个3D地图描述子的重构数据;对所述Q个3D地图描述子各自的量化数据,进行反量化处理,得到Q个反量化数据;根据所述Q个反量化数据和所述Q个3D地图描述子各自的二值化数据,得到所述Q个3D地图描述子各自的重构数据。
- 根据权利要求10所述的装置,其特征在于,在N=Q的情况下,所述P个3D地图描述子的压缩数据包括P个3D地图描述子各自的二值化数据和量化数据,所述解压缩模块具体用于:对所述P个3D地图描述子各自的量化数据,进行反量化处理,得到P个反量化数据;根据所述P个反量化数据和所述P个3D地图描述子各自的二值化数据,得到所述P个3D地图描述子各自的重构数据。
- 根据权利要求9至15任一项所述的装置,其特征在于,所述S个3D地图描述子为S个代表3D地图描述子,所述S个代表3D地图描述子各自分别对应至少一个数据集合,所述至少一个数据集合的各个数据集合包括至少一个3D地图描述子,所述检索模块用于:根据所述检索描述子的二值化数据,在所述S个代表3D地图描述子的二值化数据中进 行第i级检索,以得到至少一个代表3D地图描述子;将所述至少一个代表3D地图描述子各自对应的数据集合中的3D地图描述子,作为所述P个3D地图描述子。
- 一种3D地图的检索装置,其特征在于,包括:一个或多个处理器;存储器,用于存储一个或多个程序;当所述一个或多个程序被所述一个或多个处理器执行,使得所述一个或多个处理器实现如权利要求1-8中任一项所述的方法。
- 一种计算机可读存储介质,其特征在于,包括计算机程序,所述计算机程序在计算机上被执行时,使得所述计算机执行权利要求1-8中任一项所述的方法。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202180099018.XA CN117460972A (zh) | 2021-06-04 | 2021-07-09 | 3d地图的检索方法和装置 |
EP21943703.5A EP4336223A1 (en) | 2021-06-04 | 2021-07-09 | 3d map retrieval method and apparatus |
KR1020247000191A KR20240017906A (ko) | 2021-06-04 | 2021-07-09 | 3d 지도 검색 방법 및 장치 |
US18/526,846 US20240111793A1 (en) | 2021-06-04 | 2023-12-01 | Method and apparatus for retrieving 3d map |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CNPCT/CN2021/098483 | 2021-06-04 | ||
PCT/CN2021/098482 WO2022252234A1 (zh) | 2021-06-04 | 2021-06-04 | 3d地图的编码装置和方法 |
CNPCT/CN2021/098482 | 2021-06-04 | ||
PCT/CN2021/098483 WO2022252235A1 (zh) | 2021-06-04 | 2021-06-04 | 3d地图的解码装置和方法、及3d地图的编码码流 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US18/526,846 Continuation US20240111793A1 (en) | 2021-06-04 | 2023-12-01 | Method and apparatus for retrieving 3d map |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022252346A1 true WO2022252346A1 (zh) | 2022-12-08 |
Family
ID=84322741
Family Applications (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/104601 WO2022252337A1 (zh) | 2021-06-04 | 2021-07-05 | 3d地图的编解码方法及装置 |
PCT/CN2021/105624 WO2022252346A1 (zh) | 2021-06-04 | 2021-07-09 | 3d地图的检索方法和装置 |
PCT/CN2021/105623 WO2022252345A1 (zh) | 2021-06-04 | 2021-07-09 | 3d地图的压缩、解压缩方法和装置 |
PCT/CN2021/105625 WO2022252347A1 (zh) | 2021-06-04 | 2021-07-09 | 3d地图的检索方法和装置 |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/104601 WO2022252337A1 (zh) | 2021-06-04 | 2021-07-05 | 3d地图的编解码方法及装置 |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/105623 WO2022252345A1 (zh) | 2021-06-04 | 2021-07-09 | 3d地图的压缩、解压缩方法和装置 |
PCT/CN2021/105625 WO2022252347A1 (zh) | 2021-06-04 | 2021-07-09 | 3d地图的检索方法和装置 |
Country Status (5)
Country | Link |
---|---|
US (3) | US20240112372A1 (zh) |
EP (3) | EP4336223A1 (zh) |
KR (3) | KR20240013238A (zh) |
CN (4) | CN117413523A (zh) |
WO (4) | WO2022252337A1 (zh) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000285121A (ja) * | 1999-03-30 | 2000-10-13 | Sony Corp | 地図検索装置、地図検索方法、地図検索制御プログラムを記録した記録媒体、ナビゲーション装置、ナビゲート方法及びナビゲート制御プログラムを記録した記録媒体 |
CN102168983A (zh) * | 2010-02-12 | 2011-08-31 | 三星电子株式会社 | 用于发送和接收地图数据的方法和装置 |
CN103745498A (zh) * | 2014-01-16 | 2014-04-23 | 中国科学院自动化研究所 | 一种基于图像的快速定位方法 |
US20150043828A1 (en) * | 2013-08-08 | 2015-02-12 | Stmicroelectronics Sa | Method for searching for a similar image in an image database based on a reference image |
CN111551186A (zh) * | 2019-11-29 | 2020-08-18 | 福瑞泰克智能系统有限公司 | 一种车辆实时定位方法、系统及车辆 |
Family Cites Families (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR100238066B1 (ko) * | 1996-06-26 | 2000-01-15 | 윤종용 | 양자화활동도를 이용한 비트 발생량 제어방법 및 그 영상 부호화장치 |
KR100668714B1 (ko) * | 2005-01-14 | 2007-01-16 | 한국전자통신연구원 | 효과적인 텍스처 매핑을 위한 3차원 메쉬 정보의 텍스처좌표 부호화 및 복호화 방법 |
CN102611892B (zh) * | 2006-03-16 | 2014-10-08 | 华为技术有限公司 | 在编码过程中实现自适应量化的方法及装置 |
US20120306850A1 (en) * | 2011-06-02 | 2012-12-06 | Microsoft Corporation | Distributed asynchronous localization and mapping for augmented reality |
CN104794219A (zh) * | 2015-04-28 | 2015-07-22 | 杭州电子科技大学 | 一种基于地理位置信息的场景检索方法 |
CN116883522A (zh) * | 2015-12-14 | 2023-10-13 | 松下电器(美国)知识产权公司 | 三维数据编码方法、解码方法、编码装置、解码装置 |
KR20170082797A (ko) * | 2016-01-07 | 2017-07-17 | 주식회사 케이티 | 컨텐츠 기반 영상 검색을 위한 특징 서술자 부호화 방법 및 장치 |
EP3407607A1 (en) * | 2017-05-24 | 2018-11-28 | Thomson Licensing | Method and device for encoding and reconstructing a point cloud |
CN107590814A (zh) * | 2017-09-01 | 2018-01-16 | 上海斐讯数据通信技术有限公司 | 一种三维图像的二值化处理方法及系统 |
CN108182205A (zh) * | 2017-12-13 | 2018-06-19 | 南京信息工程大学 | 一种基于sift的hash算法的图像快速检索方法 |
JP7273029B2 (ja) * | 2018-04-10 | 2023-05-12 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 三次元データ符号化方法、三次元データ復号方法、三次元データ符号化装置、及び三次元データ復号装置 |
JP7323545B2 (ja) * | 2018-10-09 | 2023-08-08 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 三次元データ符号化方法、三次元データ復号方法、三次元データ符号化装置、及び三次元データ復号装置 |
WO2020143005A1 (zh) * | 2019-01-10 | 2020-07-16 | 深圳市大疆创新科技有限公司 | 对三维数据点集进行编码或解码的方法和设备 |
CN112384950A (zh) * | 2019-06-12 | 2021-02-19 | 浙江大学 | 点云的编解码方法及装置 |
US20210042958A1 (en) * | 2019-08-09 | 2021-02-11 | Facebook Technologies, Llc | Localization and mapping utilizing visual odometry |
CN112348885A (zh) * | 2019-08-09 | 2021-02-09 | 华为技术有限公司 | 视觉特征库的构建方法、视觉定位方法、装置和存储介质 |
CN110706248B (zh) * | 2019-08-20 | 2024-03-12 | 广东工业大学 | 一种基于slam的视觉感知建图方法及移动机器人 |
CN111639147B (zh) * | 2020-04-24 | 2021-02-05 | 视辰信息科技(上海)有限公司 | 一种地图压缩方法、系统及计算机可读存储介质 |
CN111882590A (zh) * | 2020-06-24 | 2020-11-03 | 广州万维创新科技有限公司 | 一种基于单张图片定位的ar场景应用方法 |
CN111982152B (zh) * | 2020-07-31 | 2022-04-05 | 广州文远知行科技有限公司 | 点云地图量化方法、装置、计算机设备和存储介质 |
-
2021
- 2021-07-05 CN CN202180098947.9A patent/CN117413523A/zh active Pending
- 2021-07-05 WO PCT/CN2021/104601 patent/WO2022252337A1/zh unknown
- 2021-07-09 KR KR1020237045074A patent/KR20240013238A/ko active Search and Examination
- 2021-07-09 WO PCT/CN2021/105624 patent/WO2022252346A1/zh active Application Filing
- 2021-07-09 EP EP21943703.5A patent/EP4336223A1/en active Pending
- 2021-07-09 EP EP21943704.3A patent/EP4336454A1/en active Pending
- 2021-07-09 KR KR1020247000191A patent/KR20240017906A/ko unknown
- 2021-07-09 CN CN202180098601.9A patent/CN117397242A/zh active Pending
- 2021-07-09 WO PCT/CN2021/105623 patent/WO2022252345A1/zh active Application Filing
- 2021-07-09 EP EP21943702.7A patent/EP4339881A1/en active Pending
- 2021-07-09 CN CN202180099018.XA patent/CN117460972A/zh active Pending
- 2021-07-09 CN CN202180099015.6A patent/CN117461056A/zh active Pending
- 2021-07-09 KR KR1020247000354A patent/KR20240017918A/ko unknown
- 2021-07-09 WO PCT/CN2021/105625 patent/WO2022252347A1/zh active Application Filing
-
2023
- 2023-12-01 US US18/526,734 patent/US20240112372A1/en active Pending
- 2023-12-01 US US18/526,801 patent/US20240095265A1/en active Pending
- 2023-12-01 US US18/526,846 patent/US20240111793A1/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2000285121A (ja) * | 1999-03-30 | 2000-10-13 | Sony Corp | 地図検索装置、地図検索方法、地図検索制御プログラムを記録した記録媒体、ナビゲーション装置、ナビゲート方法及びナビゲート制御プログラムを記録した記録媒体 |
CN102168983A (zh) * | 2010-02-12 | 2011-08-31 | 三星电子株式会社 | 用于发送和接收地图数据的方法和装置 |
US20150043828A1 (en) * | 2013-08-08 | 2015-02-12 | Stmicroelectronics Sa | Method for searching for a similar image in an image database based on a reference image |
CN103745498A (zh) * | 2014-01-16 | 2014-04-23 | 中国科学院自动化研究所 | 一种基于图像的快速定位方法 |
CN111551186A (zh) * | 2019-11-29 | 2020-08-18 | 福瑞泰克智能系统有限公司 | 一种车辆实时定位方法、系统及车辆 |
Also Published As
Publication number | Publication date |
---|---|
CN117461056A (zh) | 2024-01-26 |
WO2022252347A1 (zh) | 2022-12-08 |
US20240112372A1 (en) | 2024-04-04 |
EP4339881A1 (en) | 2024-03-20 |
KR20240017918A (ko) | 2024-02-08 |
KR20240013238A (ko) | 2024-01-30 |
CN117413523A (zh) | 2024-01-16 |
US20240111793A1 (en) | 2024-04-04 |
EP4336223A1 (en) | 2024-03-13 |
CN117397242A (zh) | 2024-01-12 |
WO2022252337A1 (zh) | 2022-12-08 |
US20240095265A1 (en) | 2024-03-21 |
EP4336454A1 (en) | 2024-03-13 |
WO2022252345A1 (zh) | 2022-12-08 |
CN117460972A (zh) | 2024-01-26 |
KR20240017906A (ko) | 2024-02-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20200175054A1 (en) | System and method for determining a location on multimedia content | |
WO2019000880A1 (zh) | 实景导航方法、装置、设备及计算机可读存储介质 | |
JP2012160047A (ja) | 対応参照画像検索装置及び方法、コンテンツ重畳装置、システム、及び方法、並びにコンピュータプログラム | |
CN113822427A (zh) | 一种模型训练的方法、图像匹配的方法、装置及存储介质 | |
CN111832579A (zh) | 地图兴趣点数据处理方法、装置、电子设备以及可读介质 | |
Guan et al. | On-device mobile visual location recognition by using panoramic images and compressed sensing based visual descriptors | |
WO2021088497A1 (zh) | 虚拟物体显示方法、全局地图更新方法以及设备 | |
WO2022252346A1 (zh) | 3d地图的检索方法和装置 | |
WO2023051383A1 (zh) | 一种设备定位方法、设备及系统 | |
WO2022253228A1 (zh) | 3d地图的压缩、解压缩方法和装置 | |
WO2022252238A1 (zh) | 3d地图的压缩、解压缩方法和装置 | |
CN114925680A (zh) | 物流兴趣点信息生成方法、装置、设备和计算机可读介质 | |
WO2022252237A1 (zh) | 3d地图的编解码方法及装置 | |
WO2022252234A1 (zh) | 3d地图的编码装置和方法 | |
CN114610938A (zh) | 遥感图像检索方法、装置、电子设备和计算机可读介质 | |
WO2022252236A1 (zh) | 3d地图的编解码方法及装置 | |
US20240119639A1 (en) | Apparatus and method for decoding 3d map, and encoded bitstream of 3d map |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21943703 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2023574725 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2021943703 Country of ref document: EP |
|
ENP | Entry into the national phase |
Ref document number: 2021943703 Country of ref document: EP Effective date: 20231208 |
|
ENP | Entry into the national phase |
Ref document number: 20247000191 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1020247000191 Country of ref document: KR |
|
NENP | Non-entry into the national phase |
Ref country code: DE |