WO2021237443A1 - Procédé et appareil de positionnement visuel, dispositif et support d'enregistrement lisible - Google Patents

Procédé et appareil de positionnement visuel, dispositif et support d'enregistrement lisible Download PDF

Info

Publication number
WO2021237443A1
WO2021237443A1 PCT/CN2020/092284 CN2020092284W WO2021237443A1 WO 2021237443 A1 WO2021237443 A1 WO 2021237443A1 CN 2020092284 W CN2020092284 W CN 2020092284W WO 2021237443 A1 WO2021237443 A1 WO 2021237443A1
Authority
WO
WIPO (PCT)
Prior art keywords
photos
positioning
visual positioning
panoramic
neural network
Prior art date
Application number
PCT/CN2020/092284
Other languages
English (en)
Chinese (zh)
Inventor
陈尊裕
吴珏其
胡斯洋
陈欣
吴沛谦
张仲文
Original Assignee
蜂图志科技控股有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 蜂图志科技控股有限公司 filed Critical 蜂图志科技控股有限公司
Priority to CN202080001067.0A priority Critical patent/CN111758118B/zh
Priority to JP2022566049A priority patent/JP7446643B2/ja
Priority to PCT/CN2020/092284 priority patent/WO2021237443A1/fr
Publication of WO2021237443A1 publication Critical patent/WO2021237443A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/97Determining parameters from multiple pictures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Definitions

  • This application relates to the field of positioning technology, and in particular to a visual positioning method, device, equipment, and readable storage medium.
  • the principle of visual positioning based on machine learning Use a large number of real scene photos with location markers for training, and get a neural network model whose input is a photo (RGB numerical matrix) and output is a specific location. After obtaining the trained neural network model, the user only needs to take a picture of the environment to get the specific shooting location.
  • This method needs to collect a large number of photo samples of the use environment as a training data set. For example, some documents record that in order to realize the visual positioning of a 35-meter-wide street corner store, 330 photos need to be collected, and in order to realize the visual positioning of a 140-meter street (positioning only on one side), 1,500 Multiple photos; in order to realize the positioning of a certain factory, the factory needs to be divided into 18 areas, and each area needs to take 200 images. It can be seen that in order to ensure the visual positioning effect, it is necessary to collect a large number of on-site photos as training data, and these photos must be taken to every corner of the scene, which is very time-consuming and labor-intensive.
  • the purpose of this application is to provide a visual positioning method, device, equipment, and readable storage medium, which use panoramic photos in a real map to train a neural network model, which can solve the problem of difficult sample collection in visual positioning.
  • a visual positioning method including:
  • the positioning model is a neural network model trained by using panoramic photos in a real map
  • the final position is determined.
  • said determining the final position using a plurality of said candidate positions includes:
  • the geometric center of the geometric figure is taken as the final positioning.
  • it also includes:
  • the standard deviation is used as the positioning error of the final positioning.
  • the process of training the neural network model includes:
  • the geographic mark includes a geographic location and a specific orientation
  • the neural network model is trained using the training samples, and the trained neural network model is determined as the positioning model.
  • said performing the anti-distortion transformation on several of said panoramic photos to obtain several groups of plane projection photos with the same aspect ratio including:
  • each of the panoramic photos is divided according to different focal length parameters, and several groups of plane projection photos with different viewing angles are obtained.
  • dividing each of the panoramic photos according to different focal length parameters to obtain several groups of plane projection photos with different viewing angles includes:
  • Each of the panoramic photos is segmented according to the number of segments corresponding to the original image coverage greater than a specified percentage, and several groups of adjacent images have plane projection photos with overlapping viewing angles.
  • the process of training the neural network model further includes:
  • the training samples are supplemented by using scene photos obtained from the Internet or environment photos collected from the positioning environment.
  • performing random segmentation on the wide-angle photos to obtain the atlas to be tested includes:
  • random division is performed on the wide-angle photo with an original image coverage greater than a specified percentage, and a set of atlases to be tested matching the number of divisions is obtained.
  • a visual positioning device includes:
  • the atlas to be tested acquisition module is used to acquire wide-angle photos, and randomly segment the wide-angle photos to obtain the atlas to be tested;
  • the candidate positioning acquisition module is used to input the atlas to be tested into a positioning model for positioning recognition to obtain multiple candidate positionings;
  • the positioning model is a neural network model trained by using panoramic photos in a real-world map;
  • the positioning output module is used to determine the final positioning by using a plurality of the candidate positionings.
  • a visual positioning device including:
  • Memory used to store computer programs
  • the processor is used to implement the above-mentioned visual positioning method when the computer program is executed.
  • the real map is a map where you can see the real street scene, and the real map includes a 360-degree real scene.
  • the panoramic photo in the real map is the real street view map, which overlaps with the application environment of visual positioning.
  • the neural network module is trained by using the panoramic photos in the real map to obtain a positioning model for visual positioning. After obtaining the wide-angle photos, perform random segmentation on the wide-angle photos to obtain the atlas to be tested. Input the atlas to be tested into the positioning model for positioning recognition, and then multiple candidate positionings can be obtained. Based on these candidate positions, the final position can be determined.
  • a positioning model can be obtained by training the neural network model based on the panoramic photos in the real scene map, and the visual positioning can be completed based on the positioning model, which solves the problem of difficulty in the collection of visual positioning training samples.
  • the embodiments of the present application also provide devices, equipment, and readable storage media corresponding to the above-mentioned visual positioning method, which have the above-mentioned technical effects, and will not be repeated here.
  • Fig. 1 is an implementation flowchart of a visual positioning method in an embodiment of the application
  • FIG. 2 is a schematic diagram of a perspective segmentation in an embodiment of this application.
  • FIG. 3 is a schematic structural diagram of a visual positioning device in an embodiment of the application.
  • FIG. 4 is a schematic structural diagram of a visual positioning device in an embodiment of this application.
  • Fig. 5 is a schematic diagram of a specific structure of a visual positioning device in an embodiment of the application.
  • the visual positioning method provided in the embodiment of the present invention can be directly applied to a cloud server, or in a local device.
  • the device that needs to be positioned can be positioned through a wide-angle photo if it has the functions of taking pictures and networking.
  • FIG. 1 is a flowchart of a visual positioning method in an embodiment of the application. The method includes the following steps:
  • Wide-angle that is, pictures taken with a wide-angle lens or panoramic mode. Simply put, the smaller the focal length, the wider the field of view, and the wider the range of the scene that can be accommodated in the photo.
  • the panoramic photos in the real map are used to train the neural network model. Therefore, in order to better perform visual positioning, when using the positioning model for visual positioning, the required photos are also wide-angle photos.
  • the user can use the wide-angle mode (or ultra-wide-angle mode) or the panoramic mode to take a picture of the surrounding environment at a location that needs to be positioned.
  • the angle of view exceeds 120 degrees (of course, it can also be other degrees, such as 140 degrees, 180 degrees, etc.). Photo.
  • the wide-angle photos are obtained, they are randomly divided to obtain the atlas to be tested composed of several divided photos.
  • the specific number of photos divided into the wide-angle photo can be set according to the training effect of the positioning model of the world and the actual positioning accuracy requirements.
  • the larger the number of divisions the higher the positioning accuracy.
  • the more training iterations of the model The longer the training time.
  • the wide-angle photo when segmenting the wide-angle photo, can also be randomly divided according to the number of divisions with the original image coverage greater than a specified percentage, to obtain the atlas to be tested matching the number of divisions.
  • the wide-angle photos can be randomly divided into N pieces with an aspect ratio of 1:1 (it should be noted that the aspect ratio can also be other ratios, and the aspect ratio is the same as the aspect ratio of the training sample used for training the positioning model. That is, the image whose height is 1/3 to 1/2 of the height of the wide-angle photo is used as the atlas to be measured.
  • the number of N is set according to the training effect and positioning accuracy needs.
  • N When the training effect is slightly poor and the positioning accuracy is high, select a higher value of N.
  • the number of N can be set to 100 (of course, other values can also be selected , Such as 50, 80, etc., will not be enumerated here one by one).
  • the random segmentation result requires the coverage of the original image (that is, the wide-angle photo) to be >95% (of course, it can also be set to other percentages, which will not be enumerated here).
  • S102 Input the atlas to be tested into the positioning model for positioning recognition, and obtain multiple candidate positioning.
  • the positioning model is a neural network model trained by using panoramic photos in the real map.
  • each segmented photo in the atlas to be measured is input into the positioning model for positioning recognition, and an output about the positioning result is obtained for each photo.
  • the positioning result corresponding to each divided photo is used as a candidate positioning.
  • the process of training a neural network model includes:
  • Step 1 Obtain a number of panoramic photos from the real-world map, and determine the geographic location of each real-world photo;
  • Step 2 Perform anti-distortion transformation on several panoramic photos to obtain several groups of plane projection photos with the same aspect ratio;
  • Step 3 Mark a geographic mark for each group of plane projection photos according to the correspondence with the panoramic photos; the geographic mark includes the geographic location and the specific orientation;
  • Step 4 Use the geo-tagged plane projection photos as training samples
  • Step 5 Use the training samples to train the neural network model, and determine the trained neural network model as the positioning model.
  • the panoramic photo can be subjected to anti-distortion transformation, and then several groups of plane projection photos with the same length ratio can be obtained. Since there is a corresponding relationship between the panoramic photos and the geographic locations in the real scene map, in this embodiment, the geographic locations of a group of planar projection photos divided from the same panoramic photo correspond to the geographic locations of the panoramic photos. In addition, when dividing a panoramic photo, the segmentation is performed based on the angle of view. Therefore, the orientation of the divided photo is also clear. In this embodiment, the geographic location and the specific orientation are used as geographic markers and added. In other words, every flat projection photo has a corresponding geographic location and specific orientation.
  • the trained neural network model is the positioning model. Specifically, a collection of photos with specific locations and specific orientations can be used as the data pool. Randomly select 80% of the data pool as the training set, and the remaining 20% as the test set. The ratio can also be adjusted according to the actual training situation. Input the training set into the initialized or pre-trained neural network model of the large-scale image set for training, and use the test set to verify the training results.
  • CNN Convolutional Neural Network
  • convolutional neural network which is a feedforward neural network, including convolutional layer (alternating convolutional layer) and pooling layer) and its derivative structure
  • LSTM Long Short-Term Memory, long and short-term memory network
  • RNN time recurrent neural network
  • hybrid structures etc.
  • the specific neural network used is not limited.
  • the panoramic photos can be segmented according to different focal length parameters, so as to obtain plane projection photos with different viewing angles as training samples.
  • each panoramic photo can be divided according to different focal length parameters in the anti-distortion transformation to obtain several groups of plane projection photos with different viewing angles. That is, the number of divisions n is determined according to the focal length parameter F.
  • Fig. 2 is a schematic diagram of a viewing angle segmentation in an embodiment of this application.
  • the focal length parameter F can also be changed to other values, such as 1.0 and 1.3, to obtain plane projection photos with other perspectives.
  • the panoramic photo when segmenting the panoramic photo, can also be segmented according to the number of segments corresponding to the original image coverage greater than a specified percentage. That is, under the same viewing angle, the adjacent pictures have a flat projection photo with a covering angle. Specifically, each panoramic photo is segmented according to the number of segments corresponding to the original image coverage greater than a specified percentage, and several groups of adjacent images have overlapping perspective plane projection photos. That is, in order to enrich the shooting angle of the photo, it is recommended that the number of divisions be greater than the number of equal divisions when the focal length is fixed.
  • the axis perpendicular to the ground of the panoramic photo projection spherical surface is the rotation axis, and the center of the line of sight (the arrow in the figure 2) is rotated every 45 degrees to split a plane projection photo with a viewing angle of 90 degrees.
  • the adjacent pictures There will be a 45-degree overlapping viewing angle.
  • the orientation data is then marked for the resulting flat projection photo. Because the value of F can also be 1.0 and 1.3, the viewing angle is about 60 degrees and 30 degrees, respectively, and the value of n can also be 12 and 24. You can also set more F values and increase the number of n to further improve the coverage of the training set. Generally, the coverage rate is greater than 95%.
  • the process of training the neural network model can also be used from Use the Internet to obtain scene photos, or supplement the training samples with environmental photos collected from the positioning environment.
  • the final location can be determined based on these candidate locations. After obtaining the final positioning, it can be output for the user to view.
  • one location can be randomly selected from the candidate locations as the final location, or several candidate locations can be randomly selected from the candidate locations, and the geometric centers of geometric figures corresponding to these candidate locations can be taken as the final location.
  • candidate locations with a high degree of overlap can also be used as the final location.
  • the candidate positions can be clustered and filtered, and the candidate positions that are free from most positioning positions can be removed, and then based on The remaining candidate positions determine the final position.
  • the implementation process includes:
  • Step 1 Perform clustering processing on multiple candidate locations, and use the clustering results to screen multiple candidate locations;
  • Step 2 Use the selected candidate locations to construct geometric figures
  • Step three take the geometric center of the geometric figure as the final positioning.
  • a clustering algorithm such as DBSCAN (Density-Based Spatial Clustering of Applications with Noise) can be used to classify candidate locations, and adjacent location data can be classified into one category.
  • the positioning error can also be determined.
  • the final location is used to calculate the standard deviation of multiple candidate locations; the standard deviation is used as the location error of the final location. That is, the variance between each candidate location and the final location is calculated and accumulated to obtain the final location error.
  • the real map is a map where you can see the real street scene, and the real map includes a 360-degree real scene.
  • the panoramic photo in the real map is the real street view map, which overlaps with the application environment of visual positioning.
  • the neural network module is trained by using the panoramic photos in the real map to obtain a positioning model for visual positioning. After obtaining the wide-angle photos, perform random segmentation on the wide-angle photos to obtain the atlas to be tested. Input the atlas to be tested into the positioning model for positioning recognition, and then multiple candidate positionings can be obtained. Based on these candidate positions, the final position can be determined.
  • a positioning model can be obtained by training the neural network model based on the panoramic photos in the real scene map, and the visual positioning can be completed based on the positioning model, which solves the problem of difficulty in the collection of visual positioning training samples.
  • the embodiments of the present application also provide corresponding improvement solutions.
  • the same steps as in the above-mentioned embodiments or the corresponding steps can be referred to each other, and the corresponding beneficial effects can also be referred to each other, which will not be repeated in the preferred/improved embodiments herein.
  • the embodiment of the present application also provides a visual positioning device.
  • the visual positioning device described below and the visual positioning method described above can be referred to each other.
  • the visual positioning device includes:
  • the atlas to be tested acquisition module 101 is used to acquire wide-angle photos, and randomly segment the wide-angle photos to obtain the atlas to be tested;
  • the candidate location acquisition module 102 is used to input the atlas to be tested into a location model for location recognition to obtain multiple candidate locations;
  • the location model is a neural network model trained by using panoramic photos in the real map;
  • the positioning output module 103 is used for determining the final positioning using multiple candidate positionings.
  • the wide-angle photos are obtained, and the wide-angle photos are randomly divided to obtain the atlas to be tested; the atlas to be tested is input to the positioning model for positioning recognition, and multiple candidate positions are obtained; the positioning model is Use the neural network model trained on the panoramic photos in the real map; use multiple candidate locations to determine the final location.
  • the real map is a map where you can see the real street scene, and the real map includes a 360-degree real scene.
  • the panoramic photo in the real map is the real street view map, which overlaps with the application environment of visual positioning.
  • the neural network module is trained by using the panoramic photos in the real map to obtain a positioning model for visual positioning. After obtaining the wide-angle photos, perform random segmentation on the wide-angle photos to obtain the atlas to be tested. Input the atlas to be tested into the positioning model for positioning recognition, and multiple candidate positionings can be obtained. Based on these candidate positions, the final position can be determined.
  • a positioning model can be obtained by training the neural network model based on the panoramic photos in the real map, and based on the positioning model, the visual positioning can be completed, which solves the problem of difficulty in the collection of visual positioning training samples.
  • the positioning output module 103 specifically includes:
  • the positioning screening unit is used to perform clustering processing on multiple candidate locations, and use the clustering results to screen multiple candidate locations;
  • the geometric figure construction unit is used to construct a geometric figure by using several candidate positions obtained by screening;
  • the final positioning determining unit is used to take the geometric center of the geometric figure as the final positioning.
  • the positioning output module 103 further includes:
  • the positioning error determining unit is used to calculate the standard deviation of multiple candidate positioning by using the final positioning; the standard deviation is used as the positioning error of the final positioning.
  • the model training module includes:
  • the panoramic photo obtaining unit is used to obtain several panoramic photos from the real-world map and determine the geographic location of each real-world photo;
  • the anti-distortion transformation unit is used to perform anti-distortion transformation on several panoramic photos to obtain several groups of plane projection photos with the same aspect ratio;
  • the geotagging unit is used to tag each group of plane projection photos with geotags according to the corresponding relationship with the panoramic photos; the geotags include geographic location and specific orientation;
  • the training sample determination unit is used to use the geographic-tagged plane projection photos as the training sample
  • the model training unit is used to train the neural network model using training samples, and determine the trained neural network model as a positioning model.
  • the anti-warping transformation unit is specifically used to segment each panoramic photo according to different focal length parameters in the anti-warping transformation to obtain several groups of plane projection photos with different viewing angles.
  • the anti-distortion transformation unit is specifically used to divide each panoramic photo according to the number of divisions whose coverage ratio of the corresponding original image is greater than a specified percentage, to obtain planes with overlapping viewing angles in several groups of adjacent pictures Project photos.
  • model training module further includes:
  • the sample supplement unit is used to supplement the training samples by using the scene photos obtained from the Internet or the environment photos collected from the positioning environment.
  • the atlas acquisition module 101 to be tested is specifically configured to perform random segmentation of the wide-angle photo with the original image coverage greater than a specified percentage according to the number of segments, to obtain the image to be tested matching the number of segments set.
  • the embodiment of the present application also provides a visual positioning device.
  • the visual positioning device described below and the visual positioning method described above can be referenced correspondingly.
  • the visual positioning device includes:
  • the memory 410 is used to store computer programs
  • the processor 420 is configured to implement the steps of the visual positioning method provided in the foregoing method embodiment when executing a computer program.
  • FIG. 5 is a schematic diagram of a specific structure of a visual positioning device provided by this embodiment.
  • the visual positioning device may have relatively large differences due to different configurations or performance, and may include one or more processors ( Central processing units, CPU) 420 (for example, one or more processors) and memory 410, one or more of which store computer application programs 413 or data 412.
  • the memory 410 may be short-term storage or persistent storage.
  • the computer application program may include one or more modules (not shown in the figure), and each module may include a series of instruction operations on the data processing device.
  • the central processing unit 420 may be configured to communicate with the memory 410 and execute a series of instruction operations in the memory 410 on the visual positioning device 301.
  • the visual positioning device 400 may also include one or more power supplies 430, one or more wired or wireless network interfaces 440, one or more input and output interfaces 450, and/or one or more operating systems 411.
  • the steps in the visual positioning method described above can be implemented by the structure of the visual positioning device.
  • the embodiment of the present application also provides a readable storage medium, and a readable storage medium described below and a visual positioning method described above can be referenced correspondingly.
  • a readable storage medium in which a computer program is stored, and when the computer program is executed by a processor, the steps of the visual positioning method provided by the foregoing method embodiment are implemented.
  • the readable storage medium can specifically be a U disk, a mobile hard disk, a read-only memory (Read-Only Memory, ROM), a random access memory (Random Access Memory, RAM), a magnetic disk, or an optical disk that can store program codes. Readable storage medium.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Software Systems (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

Procédé et appareil de positionnement visuel, dispositif, et support d'enregistrement lisible, le procédé comprenant : l'acquisition d'une photo à grand angle, et la segmentation aléatoire de la photo à grand angle pour obtenir un atlas à mesurer (S101) ; et l'entrée de l'atlas dans un modèle de positionnement pour une identification de position pour obtenir une pluralité de positions candidates, le modèle de positionnement étant un modèle de réseau de neurones artificiels formé à l'aide des photos panoramiques dans une carte réelle (S102) ; et la détermination de la position finale à l'aide de la pluralité de positions candidates (S103). Le modèle de positionnement peut être obtenu par formation du modèle de réseau de neurones artificiels sur la base des photos panoramiques dans la carte réelle, et le positionnement visuel peut être achevé sur la base du modèle de positionnement, ce qui résout le problème de la difficulté de collecte d'échantillons d'apprentissage pour un positionnement visuel.
PCT/CN2020/092284 2020-05-26 2020-05-26 Procédé et appareil de positionnement visuel, dispositif et support d'enregistrement lisible WO2021237443A1 (fr)

Priority Applications (3)

Application Number Priority Date Filing Date Title
CN202080001067.0A CN111758118B (zh) 2020-05-26 2020-05-26 一种视觉定位方法、装置、设备及可读存储介质
JP2022566049A JP7446643B2 (ja) 2020-05-26 2020-05-26 視覚測位方法、装置、機器及び可読記憶媒体
PCT/CN2020/092284 WO2021237443A1 (fr) 2020-05-26 2020-05-26 Procédé et appareil de positionnement visuel, dispositif et support d'enregistrement lisible

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2020/092284 WO2021237443A1 (fr) 2020-05-26 2020-05-26 Procédé et appareil de positionnement visuel, dispositif et support d'enregistrement lisible

Publications (1)

Publication Number Publication Date
WO2021237443A1 true WO2021237443A1 (fr) 2021-12-02

Family

ID=72713357

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/092284 WO2021237443A1 (fr) 2020-05-26 2020-05-26 Procédé et appareil de positionnement visuel, dispositif et support d'enregistrement lisible

Country Status (3)

Country Link
JP (1) JP7446643B2 (fr)
CN (1) CN111758118B (fr)
WO (1) WO2021237443A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117289626A (zh) * 2023-11-27 2023-12-26 杭州维讯机器人科技有限公司 一种用于工业化的虚拟仿真方法及系统

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113724284A (zh) * 2021-09-03 2021-11-30 四川智胜慧旅科技有限公司 一种位置锁定装置、山岳型景区的搜救系统及搜救方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109308678A (zh) * 2017-07-28 2019-02-05 株式会社理光 利用全景图像进行重定位的方法、装置及设备
CN109829406A (zh) * 2019-01-22 2019-05-31 上海城诗信息科技有限公司 一种室内空间识别方法
CN110298370A (zh) * 2018-03-21 2019-10-01 北京猎户星空科技有限公司 网络模型训练方法、装置及物体位姿确定方法、装置
CN110298320A (zh) * 2019-07-01 2019-10-01 北京百度网讯科技有限公司 一种视觉定位方法、装置及存储介质
CN110636274A (zh) * 2019-11-11 2019-12-31 成都极米科技股份有限公司 超短焦画幕对齐方法、装置和超短焦投影机及存储介质

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3650578B2 (ja) * 2000-09-28 2005-05-18 株式会社立山アールアンドディ 画像の歪みを補正するためのニューラル・ネットワークを用いたパノラマ画像ナビゲーションシステム
JP4264380B2 (ja) * 2004-04-28 2009-05-13 三菱重工業株式会社 自己位置同定方法及び該装置
CN202818503U (zh) * 2012-09-24 2013-03-20 天津市亚安科技股份有限公司 多方向监控区域预警定位自动跟踪监控装置
CN104200188B (zh) * 2014-08-25 2017-02-15 北京慧眼智行科技有限公司 一种快速定位qr码位置探测图形的方法和系统
CN108009588A (zh) * 2017-12-01 2018-05-08 深圳市智能现实科技有限公司 定位方法及装置、移动终端
JP6676082B2 (ja) * 2018-01-18 2020-04-08 光禾感知科技股▲ふん▼有限公司 屋内測位方法及びシステム、ならびにその屋内マップを作成するデバイス
US11195010B2 (en) * 2018-05-23 2021-12-07 Smoked Sp. Z O. O. Smoke detection system and method
KR102227583B1 (ko) * 2018-08-03 2021-03-15 한국과학기술원 딥 러닝 기반의 카메라 캘리브레이션 방법 및 장치
CN109285178A (zh) * 2018-10-25 2019-01-29 北京达佳互联信息技术有限公司 图像分割方法、装置及存储介质
CN110136136B (zh) * 2019-05-27 2022-02-08 北京达佳互联信息技术有限公司 场景分割方法、装置、计算机设备及存储介质
CN110503037A (zh) * 2019-08-22 2019-11-26 三星电子(中国)研发中心 一种在区域内定位物品的方法及系统

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109308678A (zh) * 2017-07-28 2019-02-05 株式会社理光 利用全景图像进行重定位的方法、装置及设备
CN110298370A (zh) * 2018-03-21 2019-10-01 北京猎户星空科技有限公司 网络模型训练方法、装置及物体位姿确定方法、装置
CN109829406A (zh) * 2019-01-22 2019-05-31 上海城诗信息科技有限公司 一种室内空间识别方法
CN110298320A (zh) * 2019-07-01 2019-10-01 北京百度网讯科技有限公司 一种视觉定位方法、装置及存储介质
CN110636274A (zh) * 2019-11-11 2019-12-31 成都极米科技股份有限公司 超短焦画幕对齐方法、装置和超短焦投影机及存储介质

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117289626A (zh) * 2023-11-27 2023-12-26 杭州维讯机器人科技有限公司 一种用于工业化的虚拟仿真方法及系统
CN117289626B (zh) * 2023-11-27 2024-02-02 杭州维讯机器人科技有限公司 一种用于工业化的虚拟仿真方法及系统

Also Published As

Publication number Publication date
CN111758118B (zh) 2024-04-16
CN111758118A (zh) 2020-10-09
JP2023523364A (ja) 2023-06-02
JP7446643B2 (ja) 2024-03-11

Similar Documents

Publication Publication Date Title
US10922844B2 (en) Image positioning method and system thereof
US9542773B2 (en) Systems and methods for generating three-dimensional models using sensed position data
CN111046125A (zh) 一种视觉定位方法、系统及计算机可读存储介质
WO2021237443A1 (fr) Procédé et appareil de positionnement visuel, dispositif et support d'enregistrement lisible
US9551579B1 (en) Automatic connection of images using visual features
WO2021027692A1 (fr) Procédé et appareil de construction de bibliothèque de caractéristiques visuelles, procédé et appareil de positionnement visuel et support d'informations
CN111626295B (zh) 车牌检测模型的训练方法和装置
US20090232415A1 (en) Platform for the production of seamless orthographic imagery
CN111383204A (zh) 视频图像融合方法、融合装置、全景监控系统及存储介质
US20170039450A1 (en) Identifying Entities to be Investigated Using Storefront Recognition
WO2021004416A1 (fr) Procédé et appareil permettant d'établir une carte de balises sur la base de balises visuelles
CN115049878B (zh) 基于人工智能的目标检测优化方法、装置、设备及介质
CN113808267A (zh) 一种基于gis地图的社区三维展示方法及系统
Zhao et al. Spherical superpixel segmentation
CN112637519A (zh) 多路4k准实时拼接视频的全景拼接算法
CN114820924A (zh) 一种基于bim和视频监控的博物馆参观分析的方法及系统
CN112613107A (zh) 杆塔工程确定施工进度的方法、装置、存储介质及设备
WO2022247126A1 (fr) Procédé et appareil de localisation visuelle, dispositif, support et programme
CN113379748A (zh) 一种点云全景分割方法和装置
San Blas et al. A Platform for Swimming Pool Detection and Legal Verification Using a Multi-Agent System and Remote Image Sensing.
Abrams et al. Webcams in context: Web interfaces to create live 3D environments
CN117253022A (zh) 一种对象识别方法、装置及查验设备
Budianti et al. Background blurring and removal for 3d modelling of cultural heritage objects
CN111107307A (zh) 基于单应性变换的视频融合方法、系统、终端及介质
Wang et al. Identifying people wearing masks in a 3D-scene

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20937750

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2022566049

Country of ref document: JP

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20937750

Country of ref document: EP

Kind code of ref document: A1