WO2024096717A1 - Procédé et système d'acquisition automatique d'une paire de correspondance de points caractéristiques entre des images de vue de rue à l'aide d'un modèle tridimensionnel - Google Patents

Procédé et système d'acquisition automatique d'une paire de correspondance de points caractéristiques entre des images de vue de rue à l'aide d'un modèle tridimensionnel Download PDF

Info

Publication number
WO2024096717A1
WO2024096717A1 PCT/KR2023/017670 KR2023017670W WO2024096717A1 WO 2024096717 A1 WO2024096717 A1 WO 2024096717A1 KR 2023017670 W KR2023017670 W KR 2023017670W WO 2024096717 A1 WO2024096717 A1 WO 2024096717A1
Authority
WO
WIPO (PCT)
Prior art keywords
street view
building
model
view image
matching
Prior art date
Application number
PCT/KR2023/017670
Other languages
English (en)
Korean (ko)
Inventor
전준호
Original Assignee
네이버랩스 주식회사
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 네이버랩스 주식회사 filed Critical 네이버랩스 주식회사
Publication of WO2024096717A1 publication Critical patent/WO2024096717A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • G06T7/344Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods involving models
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/10Constructive solid geometry [CSG] using solid primitives, e.g. cylinders, cubes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/14Transformations for image registration, e.g. adjusting or mapping for alignment of images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/18Image warping, e.g. rearranging pixels individually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects
    • G06V20/653Three-dimensional objects by matching three-dimensional models, e.g. conformal mapping of Riemann surfaces
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/698Control of cameras or camera modules for achieving an enlarged field of view, e.g. panoramic image capture

Definitions

  • the present disclosure relates to a method and system for automatically acquiring feature point matching pairs between street view images using a 3D model. Specifically, information included in the 3D model when performing feature matching between street view images captured on the ground. It relates to a method and system for more accurately acquiring feature point matching pairs by using .
  • 3D content can realistically express the object of expression, so it is widely used not only in entertainment fields such as movies and games, but also in map information services such as navigation and autonomous driving services.
  • map information services such as navigation and autonomous driving services.
  • research on technology for converting already existing 2D content into 3D content is being actively conducted.
  • street view service As an area of map information service, street view service is provided.
  • the street view image in the street view service is an omnidirectional panoramic image captured at a specific location and can contain more information compared to a regular camera image with a narrow angle of view. Therefore, when producing 3D content based on street view images, it is more effective because it is possible to produce more 3D content with fewer images than using images captured using a general camera.
  • panoramic street view images which are generated using equirectangular projection, geometric distortion may occur depending on the location of the object area included in the street view image, making it difficult to produce 3D content based on this. .
  • the present disclosure provides a method for solving the above problems, a computer-readable non-transitory recording medium on which instructions are recorded, and a device (system).
  • the present disclosure may be implemented in various ways, including a method, a device (system), or a computer-readable non-transitory recording medium recording instructions.
  • a method of acquiring feature point matching pairs between street view images using a 3D model is performed for a specific area including 3D geometric information expressed as an absolute position.
  • Receiving a 3D model receiving a first street view image captured at a first node within a specific area - the first street view image includes a first building -, receiving a first street view image captured at a second node within a specific area 2.
  • Receiving a street view image - the second street view image includes the first building - and acquiring a plurality of feature point matching pairs based on the first street view image, the second street view image, and the 3D model.
  • a computer-readable non-transitory recording medium recording instructions for executing a method according to an embodiment of the present disclosure on a computer is provided.
  • an information processing system includes a communication module, a memory, and at least one processor connected to the memory and configured to execute at least one computer-readable program included in the memory.
  • At least one program receives a 3D model for a specific area including 3D geometric information expressed as an absolute position, receives a first street view image taken at a first node within the specific area, and - a first street view image.
  • the review image includes a first building -, a second street view image taken from a second node within a specific area is received, - the second street view image includes a first building, -, the first street view image, the second street view image is received.
  • the quality of feature matching for a geometrically distorted image can be improved by automatically acquiring a feature point matching pair associated with a matching target building between street view data.
  • the same matching target building is determined among numerous buildings with high appearance similarity, and feature point matching pairs for the building are automatically generated. This can reduce the cost and effort of acquiring matching pairs of feature points for each of multiple buildings included in the street view data.
  • FIG. 1 is a diagram illustrating an example of a method for matching a 3D model and street view data according to an embodiment of the present disclosure.
  • Figure 2 is a schematic diagram showing a configuration in which an information processing system according to an embodiment of the present disclosure is connected to enable communication with a plurality of user terminals.
  • Figure 3 is a block diagram showing the internal configuration of a user terminal and an information processing system according to an embodiment of the present disclosure.
  • FIG. 4 is a diagram illustrating an example of determining a matching candidate area among a plurality of buildings included in a street view image captured from the ground according to an embodiment of the present disclosure.
  • FIG. 5 is a diagram illustrating an example of a process of performing feature matching based on a first matching candidate area in a first street view image and a second matching candidate area in a second street view image according to an embodiment of the present disclosure.
  • FIG. 6 is a diagram illustrating an example of a plurality of feature points detected from a planar image of a matching target building according to an embodiment of the present disclosure.
  • FIG. 7 is a diagram illustrating an example of a matching result between a first planar image and a second planar image according to an embodiment of the present disclosure.
  • FIG. 8 is a diagram illustrating an example of a feature point matching result between street view images acquired based on matching results for each of a plurality of matching target buildings according to an embodiment of the present disclosure.
  • Figure 9 is a diagram showing an example of selecting a pair of street view images that commonly include the same building based on a street view image map.
  • FIG. 10 is a flowchart illustrating an example of a method for acquiring matching pairs of feature points between street view images using a 3D model according to an embodiment of the present disclosure.
  • a modulee' or 'unit' refers to a software or hardware component, and the 'module' or 'unit' performs certain roles.
  • 'module' or 'unit' is not limited to software or hardware.
  • a 'module' or 'unit' may be configured to reside on an addressable storage medium and may be configured to run on one or more processors.
  • a 'module' or 'part' refers to components such as software components, object-oriented software components, class components and task components, processes, functions and properties. , procedures, subroutines, segments of program code, drivers, firmware, microcode, circuits, data, databases, data structures, tables, arrays, or variables.
  • Components and 'modules' or 'parts' may be combined into smaller components and 'modules' or 'parts' or further components and 'modules' or 'parts'.
  • a 'module' or 'unit' may be implemented with a processor and memory.
  • 'Processor' should be interpreted broadly to include general-purpose processors, central processing units (CPUs), microprocessors, digital signal processors (DSPs), controllers, microcontrollers, state machines, etc.
  • 'processor' may refer to an application-specific integrated circuit (ASIC), programmable logic device (PLD), field programmable gate array (FPGA), etc.
  • ASIC application-specific integrated circuit
  • PLD programmable logic device
  • FPGA field programmable gate array
  • 'Processor' refers to a combination of processing devices, for example, a combination of a DSP and a microprocessor, a combination of a plurality of microprocessors, a combination of one or more microprocessors in combination with a DSP core, or any other such combination of configurations. You may. Additionally, 'memory' should be interpreted broadly to include any electronic component capable of storing electronic information.
  • RAM random access memory
  • ROM read-only memory
  • NVRAM non-volatile random access memory
  • PROM programmable read-only memory
  • EPROM erasable-programmable read-only memory
  • a memory is said to be in electronic communication with a processor if the processor can read information from and/or write information to the memory.
  • the memory integrated into the processor is in electronic communication with the processor.
  • 'system' may include at least one of a server device and a cloud device, but is not limited thereto.
  • a system may consist of one or more server devices.
  • a system may consist of one or more cloud devices.
  • the system may be operated with a server device and a cloud device configured together.
  • 'display' may refer to any display device associated with a computing device, e.g., any display device capable of displaying any information/data controlled by or provided by the computing device. can refer to.
  • 'each of a plurality of A' or 'each of a plurality of A' may refer to each of all components included in a plurality of A, or may refer to each of some components included in a plurality of A. .
  • 'street view data' may refer to data including road view data including images captured on the roadway and location information, as well as walk view data including images captured on the sidewalk and location information. .
  • 'street view data' may further include images and location information taken at random points outdoors (or indoors facing the outdoors), as well as roadways and sidewalks.
  • FIG. 1 is a diagram illustrating an example of a method of matching a 3D model 110 and street view data 120 according to an embodiment of the present disclosure.
  • the information processing system may acquire/receive the 3D model 110 and street view data 120 for a specific area.
  • the 3D model 110 may include 3D geometric information expressed in absolute coordinate positions and texture information corresponding thereto.
  • the location information included in the 3D model 110 may be information of higher accuracy than the location information included in the street view data 120.
  • the texture information included in the 3D model 110 may be of lower quality (eg, lower resolution) than the texture information included in the street view data 120.
  • 3D geometric information expressed as an absolute coordinate position may be generated based on an aerial photograph taken of a specific area from above the specific area.
  • the 3D model (110) for a specific area includes a 3D building model (112), a digital elevation model (DEM) (114), a true ortho image (116) for a specific area, and a digital surface. It may include a digital surface model (DSM), road layout, road DEM, etc.
  • the 3D model 110 for a specific area includes a digital surface model (DSM) containing geometric information about the ground of the specific area and an orthoimage 116 for the specific area corresponding thereto. It may be a model created based on, but is not limited to, this.
  • a precise orthoimage 116 of a specific area may be generated based on a plurality of aerial photos and the absolute coordinate location information and direction information of each aerial photo.
  • the street view data 120 may include a plurality of street view images captured at a plurality of nodes within a specific area and absolute coordinate location information for each of the plurality of street view images.
  • the location information included in the street view data 120 may be information of lower accuracy than the location information included in the 3D model 110, and the texture information included in the street view image is included in the 3D model 110. It may be information of higher quality (e.g., higher resolution) than the included texture information.
  • the location information included in the street view data 120 may be location information obtained using a GPS device when a node captures a street view image. Location information obtained using a vehicle's GPS equipment may have an error of about 5 to 10 meters.
  • street view data may include direction information (i.e., image shooting direction information) for each of a plurality of street view images.
  • the information processing system may perform map matching 130 between the 3D model 110 and street view data 120. Specifically, the information processing system may perform feature matching between texture information included in the 3D model 110 and a plurality of street view images included in the street view data 120. To perform map matching 130, the information processing system may convert at least some of the plurality of street view images included in the street view data 120 into a top view image. As a result of map matching 130, a plurality of map matching points/map matching lines 132 can be extracted.
  • the map matching point may represent a corresponding pair of one point in the street view image and one point in the 3D model 110.
  • the type of map matching point may vary depending on the type of 3D model 110 used for map matching 130, the location of the point, etc.
  • map matching points are Ground Control Points (GCP), which are point correspondence pairs on the ground within a specific area, and Building Control Points (BCP), which are point correspondence pairs on buildings within a specific area.
  • GCP Ground Control Points
  • BCP Building Control Points
  • Map matching points can be extracted not only from the ground, buildings, and structures described above, but also from street view images and arbitrary areas of the 3D model 110.
  • the map matching line may represent a corresponding pair of one line of the street view image and one line of the 3D model 110.
  • the type of map matching line may vary depending on the type of 3D model 110 used for map matching 130, the location of the line, etc.
  • map matching lines include Ground Control Line (GCL), which is a corresponding pair of lines on the ground within a specific area, and Building Control Line (BCL), which is a corresponding pair of lines on buildings within a specific area.
  • GCL Ground Control Line
  • BCL Building Control Line
  • Map matching lines can be extracted from the ground, buildings, structures, and lanes described above, as well as street view images and arbitrary areas of the 3D model 110.
  • the information processing system may perform feature matching 150 between a plurality of street view images to extract a plurality of feature point correspondence sets 152.
  • feature matching 150 between a plurality of street view images may be performed using at least a portion of the 3D model 110.
  • feature matching 150 between street view images can be performed using the 3D building model 112 included in the 3D model 110.
  • a method of acquiring feature point matching pairs between street view images using a 3D model will be described in detail later with reference to FIGS. 4 to 10.
  • the information processing system provides absolute coordinate position information and Direction information can be estimated (160).
  • the processor may estimate absolute coordinate position information and direction information for a plurality of street view images using a bundle adjustment technique (160).
  • the estimated absolute coordinate position information and direction information 162 is information in an absolute coordinate system representing the 3D model 110, and may be a parameter of 6 degrees of freedom (DoF).
  • DoF degrees of freedom
  • the absolute coordinate location information and direction information 162 estimated through this process may be data with higher precision than the absolute coordinate location information and direction information included in the street view data 120.
  • the quality of feature matching for a geometrically distorted image of an equirectangular projection can be improved by automatically acquiring a feature point matching pair associated with a matching target building between the street view data 120.
  • the 3D model 110 including 3D geometric information expressed in absolute position it is possible to determine the same matching target building among numerous buildings with high appearance similarity and automatically obtain matching pairs of feature points for the building. Therefore, the cost and effort of obtaining matching pairs of feature points for each of a plurality of buildings included in the street view data 120 can be reduced.
  • Automatically acquired feature point matching pairs between matching target buildings are used to determine the relative posture between street view data (120) captured at different nodes, providing 3D geometric information, high-accuracy location information and direction information, and high-quality It can be used for various services that utilize all texture information.
  • Figure 2 is a schematic diagram showing a configuration in which the information processing system 230 according to an embodiment of the present disclosure is connected to communicate with a plurality of user terminals 210_1, 210_2, and 210_3.
  • a plurality of user terminals 210_1, 210_2, and 210_3 may be connected to an information processing system 230 capable of providing a map information service through a network 220.
  • the plurality of user terminals 210_1, 210_2, and 210_3 may include terminals of users receiving a map information service.
  • the plurality of user terminals 210_1, 210_2, and 210_3 may be cars that capture street view images from nodes.
  • the information processing system 230 includes one or more server devices and/or databases capable of storing, providing, and executing computer-executable programs (e.g., downloadable applications) and data related to providing map information services, etc.
  • it may include one or more distributed computing devices and/or distributed databases based on cloud computing services.
  • the map information service provided by the information processing system 230 may be provided to the user through an application or web browser installed on each of the plurality of user terminals 210_1, 210_2, and 210_3.
  • the information processing system 230 may provide information corresponding to a street view image request, an image-based location recognition request, etc. received from the user terminals 210_1, 210_2, and 210_3 through an application or perform corresponding processing. You can.
  • a plurality of user terminals 210_1, 210_2, and 210_3 may communicate with the information processing system 230 through the network 220.
  • the network 220 may be configured to enable communication between a plurality of user terminals 210_1, 210_2, and 210_3 and the information processing system 230.
  • the network 220 may be, for example, a wired network such as Ethernet, a wired home network (Power Line Communication), a telephone line communication device, and RS-serial communication, a mobile communication network, a wireless LAN (WLAN), It may consist of wireless networks such as Wi-Fi, Bluetooth, and ZigBee, or a combination thereof.
  • the communication method is not limited, and may include communication methods utilizing communication networks that the network 220 may include (e.g., mobile communication networks, wired Internet, wireless Internet, broadcasting networks, satellite networks, etc.) as well as user terminals (210_1, 210_2, 210_3). ) may also include short-range wireless communication between the network 220 may include (e.g., mobile communication networks, wired Internet, wireless Internet, broadcasting networks, satellite networks, etc.) as well as user terminals (210_1, 210_2, 210_3). ) may also include short-range wireless communication between the network 220 may include (e.g., mobile communication networks, wired Internet, wireless Internet, broadcasting networks, satellite networks, etc.) as well as user terminals (210_1, 210_2, 210_3). ) may also include short-range wireless communication between the network 220 may include (e.g., mobile communication networks, wired Internet, wireless Internet, broadcasting networks, satellite networks, etc.) as well as user terminals (210_1, 210_2,
  • the mobile phone terminal (210_1), tablet terminal (210_2), and PC terminal (210_3) are shown as examples of user terminals, but they are not limited thereto, and the user terminals (210_1, 210_2, 210_3) use wired and/or wireless communication.
  • This may be any computing device capable of installing and executing an application or a web browser.
  • user terminals include AI speakers, smartphones, mobile phones, navigation, computers, laptops, digital broadcasting terminals, PDAs (Personal Digital Assistants), PMPs (Portable Multimedia Players), tablet PCs, game consoles, It may include wearable devices, IoT (internet of things) devices, VR (virtual reality) devices, AR (augmented reality) devices, set-top boxes, etc.
  • three user terminals (210_1, 210_2, 210_3) are shown as communicating with the information processing system 230 through the network 220, but this is not limited to this, and a different number of user terminals are connected to the network ( It may be configured to communicate with the information processing system 230 through 220).
  • FIG. 3 is a block diagram showing the internal configuration of the user terminal 210 and the information processing system 230 according to an embodiment of the present disclosure.
  • the user terminal 210 may refer to any computing device capable of executing an application or a web browser and capable of wired/wireless communication, for example, the mobile phone terminal 210_1, tablet terminal 210_2 of FIG. 2, It may include a PC terminal (210_3), etc.
  • the user terminal 210 may include a memory 312, a processor 314, a communication module 316, and an input/output interface 318.
  • information processing system 230 may include memory 332, processor 334, communication module 336, and input/output interface 338. As shown in FIG.
  • the user terminal 210 and the information processing system 230 are configured to communicate information and/or data through the network 220 using respective communication modules 316 and 336. It can be. Additionally, the input/output device 320 may be configured to input information and/or data to the user terminal 210 through the input/output interface 318 or to output information and/or data generated from the user terminal 210.
  • Memories 312 and 332 may include any non-transitory computer-readable recording medium. According to one embodiment, the memories 312 and 332 are non-permanent mass storage devices such as read only memory (ROM), disk drive, solid state drive (SSD), flash memory, etc. It can be included. As another example, non-perishable mass storage devices such as ROM, SSD, flash memory, disk drive, etc. may be included in the user terminal 210 or the information processing system 230 as a separate persistent storage device that is distinct from memory. Additionally, the memories 312 and 332 may store an operating system and at least one program code (eg, code for an application installed and running on the user terminal 210).
  • ROM read only memory
  • SSD solid state drive
  • flash memory etc. It can be included.
  • non-perishable mass storage devices such as ROM, SSD, flash memory, disk drive, etc. may be included in the user terminal 210 or the information processing system 230 as a separate persistent storage device that is distinct from memory.
  • the memories 312 and 332 may store an
  • These software components may be loaded from a computer-readable recording medium separate from the memories 312 and 332.
  • This separate computer-readable recording medium may include a recording medium directly connectable to the user terminal 210 and the information processing system 230, for example, a floppy drive, disk, tape, DVD/CD- It may include computer-readable recording media such as ROM drives and memory cards.
  • software components may be loaded into the memories 312 and 332 through a communication module rather than a computer-readable recording medium. For example, at least one program is loaded into memory 312, 332 based on a computer program installed by files provided over the network 220 by developers or a file distribution system that distributes installation files for applications. It can be.
  • the processors 314 and 334 may be configured to process instructions of a computer program by performing basic arithmetic, logic, and input/output operations. Instructions may be provided to the processors 314 and 334 by memories 312 and 332 or communication modules 316 and 336. For example, processors 314 and 334 may be configured to execute received instructions according to program codes stored in recording devices such as memories 312 and 332.
  • the communication modules 316 and 336 may provide a configuration or function for the user terminal 210 and the information processing system 230 to communicate with each other through the network 220, and may provide a configuration or function for the user terminal 210 and/or information processing.
  • the system 230 may provide a configuration or function for communicating with other user terminals or other systems (for example, a separate cloud system, etc.). For example, a request or data generated by the processor 314 of the user terminal 210 according to a program code stored in a recording device such as the memory 312 (e.g., data associated with a request for a street view image taken on the ground) etc.) may be transmitted to the information processing system 230 through the network 220 under the control of the communication module 316.
  • a control signal or command provided under the control of the processor 334 of the information processing system 230 is transmitted through the communication module 316 of the user terminal 210 through the communication module 336 and the network 220. It may be received by the user terminal 210. For example, the user terminal 210 may receive data related to a street view image for a specific area from the information processing system 230.
  • the input/output interface 318 may be a means for interfacing with the input/output device 320.
  • input devices may include devices such as cameras, keyboards, microphones, mice, etc., including audio sensors and/or image sensors
  • output devices may include devices such as displays, speakers, haptic feedback devices, etc. You can.
  • the input/output interface 318 may be a means for interfacing with a device that has components or functions for performing input and output, such as a touch screen, integrated into one.
  • the processor 314 of the user terminal 210 uses information and/or data provided by the information processing system 230 or another user terminal when processing instructions of a computer program loaded in the memory 312. A service screen, etc.
  • the input/output device 320 is shown not to be included in the user terminal 210, but the present invention is not limited to this and may be configured as a single device with the user terminal 210. Additionally, the input/output interface 338 of the information processing system 230 may be connected to the information processing system 230 or means for interfacing with a device (not shown) for input or output that the information processing system 230 may include. It can be. In FIG. 3 , the input/output device 320 is shown not to be included in the user terminal 210, but the present invention is not limited to this and may be configured as a single device with the user terminal 210. Additionally, the input/output interface 338 of the information processing system 230 may be connected to the information processing system 230 or means for interfacing with a device (not shown) for input or output that the information processing system 230 may include. It can be. In FIG.
  • the input/output interfaces 318 and 338 are shown as elements configured separately from the processors 314 and 334, but the present invention is not limited thereto, and the input/output interfaces 318 and 338 may be configured to be included in the processors 314 and 334. there is.
  • the user terminal 210 and information processing system 230 may include more components than those in FIG. 3 . However, there is no need to clearly show most prior art components. According to one embodiment, the user terminal 210 may be implemented to include at least some of the input/output devices 320 described above. Additionally, the user terminal 210 may further include other components such as a transceiver, a global positioning system (GPS) module, a camera, various sensors, and a database.
  • GPS global positioning system
  • the user terminal 210 may include components generally included in a smartphone, such as an acceleration sensor, a gyro sensor, an image sensor, a proximity sensor, a touch sensor, Various components such as an illuminance sensor, a camera module, various physical buttons, buttons using a touch panel, input/output ports, and a vibrator for vibration may be implemented to be further included in the user terminal 210.
  • the processor 314 of the user terminal 210 may be configured to operate an application that provides a map information service. At this time, code associated with the corresponding application and/or program may be loaded into the memory 312 of the user terminal 210.
  • the processor 314 uses input devices such as a touch screen, a keyboard, a camera including an audio sensor and/or an image sensor, and a microphone connected to the input/output interface 318. It is possible to receive text, images, videos, voices, and/or actions input or selected through, and store the received text, images, videos, voices, and/or actions in the memory 312 or use the communication module 316 and It can be provided to the information processing system 230 through the network 220. For example, the processor 314 may receive a user's input requesting a street view image for a specific area and provide the input to the information processing system 230 through the communication module 316 and the network 220.
  • input devices such as a touch screen, a keyboard, a camera including an audio sensor and/or an image sensor, and a microphone connected to the input/output interface 318. It is possible to receive text, images, videos, voices, and/or actions input or selected through, and store the received text, images, videos, voices, and/or actions in the memory 312 or use
  • the processor 314 of the user terminal 210 manages, processes, and/or stores information and/or data received from the input/output device 320, other user terminals, the information processing system 230, and/or a plurality of external systems. It can be configured to do so. Information and/or data processed by processor 314 may be provided to information processing system 230 via communication module 316 and network 220.
  • the processor 314 of the user terminal 210 may transmit information and/or data to the input/output device 320 through the input/output interface 318 and output the information. For example, the processor 314 may display the received information and/or data on the screen of the user terminal.
  • the processor 334 of the information processing system 230 may be configured to manage, process, and/or store information and/or data received from a plurality of user terminals 210 and/or a plurality of external systems. Information and/or data processed by the processor 334 may be provided to the user terminal 210 through the communication module 336 and the network 220.
  • FIG. 4 is a diagram illustrating an example of determining a matching candidate area among a plurality of buildings included in a street view image captured from the ground according to an embodiment of the present disclosure.
  • the information processing system may determine a matching candidate area for acquiring a plurality of feature point matching pairs based on a street view image captured on the ground.
  • the street view image may be a 360-degree panoramic image captured using a vehicle equipped with at least one camera. That is, the street view image may be a panoramic image created using equirectangular projection.
  • street view images may include low-precision GPS location information acquired when shooting.
  • the information processing system may determine a matching candidate area associated with a matching target building among a plurality of buildings included in the street view image, based on a street view image for a specific area and a 3D model for a specific area.
  • the 3D model may include 3D building models for a plurality of buildings.
  • a matching candidate area may be determined through the first state 410 and the second state 420.
  • the first state 410 represents an example of a result of projecting a 3D building model for a plurality of buildings onto a street view image.
  • the information processing system may project a 3D building model for a plurality of buildings onto the street view image using the location information and direction information of the street view image.
  • a plurality of 3D absolute coordinate position information associated with a 3D building model for each of a plurality of buildings can be converted into a plurality of two-dimensional points on the spherical coordinate system of the street view image and projected onto the street view image.
  • the second state 420 represents an example of a result in which an area associated with each of a plurality of buildings in the street view image is determined based on the projection result.
  • the information processing system may determine an area associated with each of a plurality of buildings in the street view image based on the projection result. For example, the information processing system may determine a minimum image area including a plurality of two-dimensional points for each of a plurality of buildings as an area associated with the building.
  • the minimum image area can be determined using the Convex Hull algorithm, etc.
  • the information processing system may determine a matching candidate area associated with the matching target building, based on the area associated with each of the plurality of buildings, by excluding buildings that are obscured by other buildings.
  • 3D absolute coordinate location information included in the 3D building model can be used. For example, determine the distance from the location where the street view image was taken to each building, and based on that distance, match the distant building if the area of the building located further away is obscured by the area of the building located closer. It can be excluded from the target building.
  • a minimum image area containing a two-dimensional point for each of a plurality of buildings may be determined as an area associated with the building. That is, the first building area 422, the second building area 424, the third building area 426, and the fourth building area 428 may be determined based on two-dimensional points associated with different buildings.
  • the information processing system then controls the first building area 422, the second building area 424, and the third building area that are not obscured by other buildings, except for the fourth building area 428 that is obscured by the other building.
  • Area 426 may be determined as the matching target building.
  • the information processing system may perform feature matching based on an area associated with a matching target building that is commonly included in a plurality of street view images. Specifically, the information processing system may use a 3D building model for the matching target building included in the first street view image to determine the area associated with the matching target building in the first street view image as the first matching candidate area. Additionally, the information processing system may use a 3D building model for a specific building included in the second street view image to determine an area associated with a specific building in the second street view image as the second matching candidate area. Then, in response to determining that the first matching candidate area and the second matching candidate area are associated with the matching target building, the information processing system performs feature matching based on the first matching candidate area and the second matching candidate area. can do.
  • Figure 4 shows that the first to fourth building areas 422 to 428 are determined among a plurality of buildings in the street view image, but this is for convenience of explanation, and the information processing system is included in the street view image.
  • Each building area can be determined for all of the plurality of buildings, and among them, the building that is not obscured by other buildings can be determined as the matching target building.
  • Figure 5 shows a process of performing feature matching based on the first matching candidate area 512 in the first street view image and the second matching candidate area 522 in the second street view image according to an embodiment of the present disclosure.
  • the information processing system uses the first matching candidate area 512 and the second matching candidate area to perform feature matching based on the first matching candidate area 512 and the second matching candidate area 522.
  • Each area 522 can be converted into a flat image.
  • the information processing system may convert the first matching candidate area 512 into a first planar image 514 using a perspective projection method.
  • the information processing system may determine the direction and size of a specific matching target building in the spherical coordinate system of the first street view image.
  • the information processing system may convert the first matching candidate area 512 into a first planar image 514 based on the determined direction and size of the first building. Similarly, the information processing system may convert the second matching candidate area 522 into a second planar image 524 using a perspective projection method. The information processing system may then perform feature matching between the first planar image 514 and the second planar image 524.
  • a geometric distortion image of an equirectangular projection including a matching target building is converted into a flat image, and then , the quality of feature matching for street view images can be improved by performing feature matching.
  • FIG. 6 is a diagram illustrating an example of a plurality of feature points detected from a planar image 600 of a matching target building according to an embodiment of the present disclosure.
  • the planar image 600 for the matching target building may represent an undistorted planar image generated based on a matching candidate area associated with the matching target building in the street view image.
  • the information processing system may detect a plurality of local feature points for generating a feature point matching pair associated with the matching target building from the planar image 600 of the matching target building. For example, the information processing system may detect a plurality of local feature points from the planar image 600 and then determine the feature point associated with the matching target building. Additionally, the information processing system can extract a visual descriptor of each feature point. Multiple feature points can be extracted using SIFT (Scale-Invariant Feature Transform), SuperPoint, and R2D2 (Repeatable and Reliable Detector and Descriptor) techniques, but are not limited to this.
  • SIFT Scale-Invariant Feature Transform
  • SuperPoint SuperPoint
  • R2D2 Repeatable and Reliable Detector and Descriptor
  • FIG. 7 is a diagram illustrating an example of a matching result between a first planar image 710 and a second planar image 720 according to an embodiment of the present disclosure.
  • the information processing system may extract a plurality of point correspondence pairs for matching target buildings between the first planar image 710 and the second planar image 720 and store them as matching results.
  • the first planar image 710 and the second planar image 720 are images of the matching target building created based on the first street view image and the second street view image captured at different nodes in a specific area. You can.
  • the information processing system includes a first set of feature points extracted from the first planar image 710 for the matching target building and a second set of feature points extracted from the second planar image 720 for the matching target building.
  • a plurality of point correspondence pairs can be extracted and stored as a feature matching result for the matching target building.
  • the first set of feature points and the second set of feature points may be extracted using the same or the same type of feature extractor.
  • the feature extractor is a visual feature extraction neural network model, and the visual feature descriptor of the corresponding pixel between the plurality of planar images (710, 720) generated by converting the street view image is similar. It may be a model that has been trained to extract the data properly.
  • the method of performing feature matching is not limited to a specific method, and any feature matching method (NN Matching, SuperPoint/Glue, deep learning-based feature matching method such as R2D2, etc.) can be used.
  • FIG. 8 is a diagram illustrating an example of a feature point matching result 820 between street view images acquired based on a matching result 810 for each of a plurality of matching target buildings according to an embodiment of the present disclosure.
  • the information processing system may obtain matching results 810 for each of a plurality of matching target buildings included in the street view image using the method described in FIGS. 4 to 7.
  • the matching result 810 for each of a plurality of matching target buildings may be a matching result between the first planar image 812_1 and the second planar image 812_2 for the first building, and the third planar image 812_2 for the second building.
  • first building, second building, and third building may represent different buildings included in the street view image associated with each planar image.
  • the information processing system may acquire a feature point matching result 820 between street view images in a spherical coordinate system based on a matching result 810 for each of a plurality of matching target buildings in a planar coordinate system.
  • the information processing system can convert two-dimensional coordinate position information in a planar image of a plurality of feature point matching pairs into two-dimensional position information in a street view image using inversion of the perspective projection method.
  • the information processing system uses two-dimensional coordinate position information in the first planar image 812_1 and two-dimensional coordinate position information in the second planar image 812_2 of the plurality of feature point matching pairs for the first building, respectively.
  • the information processing system includes two-dimensional coordinate location information in the planar images 814_1 and 814_2 of the plurality of feature point matching pairs for the second building and the planar images 816_1 and the plurality of feature point matching pairs for the third building.
  • the 2D coordinate location information in 816_2) can be converted into 2D location information in the first street view image 822 and 2D location information in the second street view image 824.
  • mapping pairs of feature points between street view images are quickly and accurately acquired automatically for multiple buildings (e.g., first building, second building, third building, etc.) included in different street view images. can do.
  • FIG. 9 is a diagram illustrating an example of selecting a pair of street view images that commonly include the same building based on the street view image map 900.
  • the street view image map 900 may include relationship information between a plurality of street view images in a specific area. For example, referring to FIG. 9 , circles or dots may represent nodes where each of a plurality of street view images is captured. Additionally, edges or lines connecting each node may indicate relationships between street view images that commonly include buildings included in the street view images.
  • the information processing system may select a pair of street view images that commonly include the same building in a plurality of street view images. For example, the information processing system may receive a plurality of street view image maps 900 captured from a plurality of nodes within a specific area. Then, the information processing system uses the three-dimensional model to display a first street view image captured at the first node 910 and a second street view image captured at the second node 920 on the plurality of street view image maps 900. It may be determined that the same building is commonly included in the street view image and the third street view image captured by the third node 930.
  • the information processing system includes a first street view image captured at the first node 910, a second street view image captured at the second node 920, and a third street view image captured at the third node 930.
  • a combination of two street view images among the images can be selected as a pair of street view images to perform feature matching.
  • Figure 9 shows an example of selecting a street view image pair based on three street view images taken at three nodes, but the present invention is not limited to this, and the street view image pair is two street view images among a plurality of street view images. It can be made up of a combination of .
  • FIG. 10 is a flowchart illustrating an example of a method 1000 for acquiring feature point matching pairs between street view images using a 3D model according to an embodiment of the present disclosure.
  • Method 1000 may be initiated by a processor (e.g., at least one processor of an information processing system) receiving a three-dimensional model for a specific region containing three-dimensional geometric information expressed in absolute positions (S1010).
  • the 3D model may include multiple 3D building models for multiple buildings.
  • the processor may receive a plurality of street view images taken from a plurality of nodes within a specific area (S1020).
  • the street view image may be a panoramic image created using equirectangular projection.
  • the plurality of nodes includes a first node and a second node, and the processor generates a first street view image captured at a first node within a specific area and a second street view image captured at a second node within a specific area. can receive.
  • the first street view image and the second street view image may include the first building.
  • the 3D model may include a 3D building model for the first building.
  • the processor may determine that the first street view image and the second street view image include the first building using a 3D model (S1030). In this case, the processor may select the first street view image and the second street view image among the plurality of street view images as a pair of street view images to perform feature matching (S1040). Afterwards, the processor may acquire a plurality of feature point matching pairs based on the first street view image, the second street view image, and the 3D model (S1050). For example, multiple feature point matching pairs can be acquired using NN Matching or SuperGlue techniques.
  • the processor determines an area associated with the first building in the first street view image as the first matching candidate area using a 3D building model for the first building to obtain a plurality of feature point matching pairs. You can.
  • the processor may use the 3D building model for the first building to determine an area associated with the first building in the second street view image as the second matching candidate area.
  • the processor may then perform feature matching based on the first matching candidate area and the second matching candidate area in response to determining that the first matching candidate area and the second matching candidate area are associated with the first building. .
  • the step of determining an area associated with the first building in the first street view image as the first matching candidate area includes using location information and direction information associated with the first street view image, It may include projecting the 3D building model onto the first street view image and determining an area associated with the first building in the first street view image based on the projection result.
  • the step of projecting the 3D building model for the first building onto the first street view image involves projecting the 3D building model for the first building using location information and direction information associated with the first street view image. It may include converting a plurality of 3D absolute coordinate location information into a plurality of 2D points on a spherical coordinate system of the first street view image.
  • determining the area associated with the first building in the first street view image based on the projection result may include determining a minimum image area including a plurality of two-dimensional points as the area associated with the first building. You can.
  • the minimum image area can be determined using the Convex Hull algorithm.
  • the 3D model may further include a 3D building model for a second building.
  • the step of acquiring a plurality of feature point matching pairs includes determining an area associated with the second building in the first street view image using a 3D building model for the second building and an area associated with the first building. It may include excluding an area obscured by an area associated with the second building from the first matching candidate area. In this case, the distance from the first node to the first building may be greater than the distance from the first node to the second building.
  • performing feature matching based on the first matching candidate area and the second matching candidate area includes converting the first matching candidate area into a first planar image using a perspective projection method, the step of converting the first matching candidate area into a first planar image using a perspective projection method. It may include converting the second matching candidate area into a second planar image using a method and performing feature matching between the first planar image and the second planar image.
  • the step of converting the first matching candidate area into a first plane image includes determining the direction and size of the first building in the spherical coordinate system of the first street view image, and based on the determined direction and size of the first building , It may include converting the first matching candidate area into a first planar image.
  • performing feature matching between the first planar image and the second planar image includes extracting a first set of feature points based on the first planar image, and extracting a first set of feature points based on the second planar image. It may include extracting feature points of the set, performing feature matching between feature points of the first set and feature points of the second set, and acquiring a plurality of feature point matching pairs based on the feature matching results. You can.
  • the first set of feature points and the second set of feature points may be extracted using the same type of feature extractor.
  • the first set of feature points and the second set of feature points may be extracted using SIFT (Scale-Invariant Feature Transform), SuperPoint, or R2D2 (Repeatable and Reliable Detector and Descriptor) techniques.
  • the step of acquiring a plurality of feature point matching pairs based on the feature matching result includes two-dimensional coordinate position information in the first plane image of the plurality of feature point matching pairs using inversion of a perspective projection method, and a second pair of feature point matching pairs. It may include converting 2D coordinate location information in the two-plane image into 2D location information in the first street view image and 2D location information in the second street view image, respectively.
  • the above-described method may be provided as a computer program stored in a computer-readable recording medium for execution on a computer.
  • the medium may continuously store a computer-executable program, or may temporarily store it for execution or download.
  • the medium may be a variety of recording or storage means in the form of a single or several pieces of hardware combined. It is not limited to a medium directly connected to a computer system and may be distributed over a network. Examples of media include magnetic media such as hard disks, floppy disks and magnetic tapes, optical recording media such as CD-ROMs and DVDs, magneto-optical media such as floptical disks, And there may be something configured to store program instructions, including ROM, RAM, flash memory, etc. Additionally, examples of other media include recording or storage media managed by app stores that distribute applications, sites or servers that supply or distribute various other software, etc.
  • the processing units used to perform the techniques may include one or more ASICs, DSPs, digital signal processing devices (DSPDs), programmable logic devices (PLDs). ), field programmable gate arrays (FPGAs), processors, controllers, microcontrollers, microprocessors, electronic devices, and other electronic units designed to perform the functions described in this disclosure. , a computer, or a combination thereof.
  • the various illustrative logical blocks, modules, and circuits described in connection with this disclosure may be general-purpose processors, DSPs, ASICs, FPGAs or other programmable logic devices, discrete gate or transistor logic, discrete hardware components, or It may be implemented or performed as any combination of those designed to perform the functions described in.
  • a general-purpose processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine.
  • a processor may also be implemented as a combination of computing devices, such as a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other configuration.
  • RAM random access memory
  • ROM read-only memory
  • NVRAM non-volatile random access memory
  • PROM on computer-readable media such as programmable read-only memory (EPROM), electrically erasable PROM (EEPROM), flash memory, compact disc (CD), magnetic or optical data storage devices, etc. It may also be implemented as stored instructions. Instructions may be executable by one or more processors and may cause the processor(s) to perform certain aspects of the functionality described in this disclosure.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Signal Processing (AREA)
  • Computer Graphics (AREA)
  • Processing Or Creating Images (AREA)

Abstract

La présente divulgation concerne un procédé mis en œuvre par au moins un processeur pour acquérir une paire de correspondance de points caractéristiques entre des images de vue de rue à l'aide d'un modèle tridimensionnel. Le procédé d'acquisition d'une paire de correspondance de points caractéristiques entre des vues de rue à l'aide d'un modèle tridimensionnel comprend les étapes consistant à : recevoir un modèle tridimensionnel par rapport à une zone spécifique, qui comprend des informations géométriques tridimensionnelles exprimées en tant que position absolue ; recevoir une première image de vue de rue capturée par un premier nœud dans une zone spécifique, la première image de vue de rue comprenant un premier bâtiment ; recevoir une seconde image de vue de rue capturée par un second nœud dans une zone spécifique, la seconde image de vue de rue comprenant le premier bâtiment ; et acquérir de multiples paires de correspondance de points de caractéristique sur la base de la première image de vue de rue, de la seconde image de vue de rue et du modèle tridimensionnel.
PCT/KR2023/017670 2022-11-04 2023-11-06 Procédé et système d'acquisition automatique d'une paire de correspondance de points caractéristiques entre des images de vue de rue à l'aide d'un modèle tridimensionnel WO2024096717A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
KR1020220146221A KR20240064332A (ko) 2022-11-04 2022-11-04 3차원 모델을 이용한 거리뷰 영상 사이의 특징점 매칭쌍 자동 취득 방법 및 시스템
KR10-2022-0146221 2022-11-04

Publications (1)

Publication Number Publication Date
WO2024096717A1 true WO2024096717A1 (fr) 2024-05-10

Family

ID=90931108

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2023/017670 WO2024096717A1 (fr) 2022-11-04 2023-11-06 Procédé et système d'acquisition automatique d'une paire de correspondance de points caractéristiques entre des images de vue de rue à l'aide d'un modèle tridimensionnel

Country Status (2)

Country Link
KR (1) KR20240064332A (fr)
WO (1) WO2024096717A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20100054056A (ko) * 2008-11-13 2010-05-24 엔에이치엔(주) 전자 지도를 이용하여 서비스를 제공하기 위한 방법, 시스템 및 컴퓨터 판독 가능한 기록 매체
US20160104289A1 (en) * 2014-10-09 2016-04-14 Caterpillar Inc. Real-time range map generation
KR20170123328A (ko) * 2015-05-27 2017-11-07 구글 엘엘씨 파노라마 가상 현실 콘텐츠의 옴니스테레오 캡쳐 및 렌더링
KR20210147855A (ko) * 2020-05-28 2021-12-07 네이버랩스 주식회사 시각 특징 맵 생성 방법 및 시스템
KR20220022350A (ko) * 2020-08-18 2022-02-25 삼성전자주식회사 이미지를 이용한 위치 정보 획득 방법 및 이를 지원하는 전자 장치

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20100054056A (ko) * 2008-11-13 2010-05-24 엔에이치엔(주) 전자 지도를 이용하여 서비스를 제공하기 위한 방법, 시스템 및 컴퓨터 판독 가능한 기록 매체
US20160104289A1 (en) * 2014-10-09 2016-04-14 Caterpillar Inc. Real-time range map generation
KR20170123328A (ko) * 2015-05-27 2017-11-07 구글 엘엘씨 파노라마 가상 현실 콘텐츠의 옴니스테레오 캡쳐 및 렌더링
KR20210147855A (ko) * 2020-05-28 2021-12-07 네이버랩스 주식회사 시각 특징 맵 생성 방법 및 시스템
KR20220022350A (ko) * 2020-08-18 2022-02-25 삼성전자주식회사 이미지를 이용한 위치 정보 획득 방법 및 이를 지원하는 전자 장치

Also Published As

Publication number Publication date
KR20240064332A (ko) 2024-05-13

Similar Documents

Publication Publication Date Title
WO2015174729A1 (fr) Procédé et système de fourniture de réalité augmentée destinés à fournir des informations spatiales, ainsi que support d'enregistrement et système de distribution de fichier
US20140300637A1 (en) Method and apparatus for determining camera location information and/or camera pose information according to a global coordinate system
WO2010027193A2 (fr) Affichage à corrélation spatiale d'un contenu tridimensionnel sur des composants d'affichage ayant des positions arbitraires
WO2011031026A2 (fr) Système de délivrance de service d'avatar en 3 dimensions et procédé d'utilisation d'image d'arrière-plan
Huo et al. SynchronizAR: Instant Synchronization for Spontaneous and Spatial Collaborations in Augmented Reality.
WO2021241847A1 (fr) Procédé et système de génération de carte de caractéristiques visuelles
WO2016182167A1 (fr) Dispositif de mise en œuvre d'une visualisation en trois dimensions d'une conception plane
WO2021025364A1 (fr) Procédé et système utilisant un lidar et une caméra pour améliorer des informations de profondeur concernant un point caractéristique d'image
WO2021125578A1 (fr) Procédé et système de reconnaissance de position reposant sur un traitement d'informations visuelles
WO2021132947A1 (fr) Dispositif électronique de fourniture de localisation d'intérieur et procédé correspondant
WO2019221340A1 (fr) Procédé et système de calcul de coordonnées spatiales d'une région d'intérêt et support d'enregistrement non transitoire lisible par ordinateur
WO2024096717A1 (fr) Procédé et système d'acquisition automatique d'une paire de correspondance de points caractéristiques entre des images de vue de rue à l'aide d'un modèle tridimensionnel
CN116858215B (zh) 一种ar导航地图生成方法及装置
WO2024106833A1 (fr) Procédé et système permettant d'acquérir automatiquement un point de commande de bâtiment
WO2024085630A1 (fr) Procédé et système d'entraînement de modèle de réseau neuronal d'extraction de caractéristiques visuelles
WO2024085628A1 (fr) Procédé et système d'acquisition automatique de point de commande au sol
WO2024085600A1 (fr) Procédé et système de mise en correspondance de modèle tridimensionnel et de données de vue de rue
WO2024101833A1 (fr) Procédé et système de génération d'une carte de caractéristiques visuelles à l'aide d'un modèle tridimensionnel et d'une image de vue de rue
WO2024085631A1 (fr) Procédé et système d'acquisition automatique de ligne de commande au sol
WO2024101776A1 (fr) Procédé et système de génération de modèle de vue de rue tridimensionnelle utilisant un modèle de construction tridimensionnel et un modèle de route
WO2021206200A1 (fr) Dispositif et procédé permettant de traiter des informations de nuage de points
WO2024085455A1 (fr) Procédé et système de correction de pose d'objet
WO2021210725A1 (fr) Appareil et procédé permettant de traiter des informations de nuage de points
WO2021075878A1 (fr) Procédé permettant de fournir un service d'enregistrement de réalité augmentée et terminal utilisateur
KR102675282B1 (ko) 자동 생성된 컨트롤 피처를 이용한 거리뷰 영상의 절대 포즈 추정 방법 및 시스템

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23886411

Country of ref document: EP

Kind code of ref document: A1