WO2022078240A1 - Camera precise positioning method applied to electronic map, and processing terminal - Google Patents

Camera precise positioning method applied to electronic map, and processing terminal Download PDF

Info

Publication number
WO2022078240A1
WO2022078240A1 PCT/CN2021/122607 CN2021122607W WO2022078240A1 WO 2022078240 A1 WO2022078240 A1 WO 2022078240A1 CN 2021122607 W CN2021122607 W CN 2021122607W WO 2022078240 A1 WO2022078240 A1 WO 2022078240A1
Authority
WO
WIPO (PCT)
Prior art keywords
camera
feature point
electronic map
dimensional
shooting
Prior art date
Application number
PCT/CN2021/122607
Other languages
French (fr)
Chinese (zh)
Inventor
高星
徐建明
石立阳
Original Assignee
佳都科技集团股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 佳都科技集团股份有限公司 filed Critical 佳都科技集团股份有限公司
Publication of WO2022078240A1 publication Critical patent/WO2022078240A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/04Indexing scheme for image data processing or generation, in general involving 3D image data
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Definitions

  • the invention relates to the technical field of electronic maps, in particular to a camera precise positioning method and a processing terminal applied in electronic maps.
  • the GPS coordinates of the camera are usually determined by means of geographic surveying and mapping.
  • the one-machine-one-file module on the security surveillance video networking platform stores the camera's location information (latitude and longitude information).
  • the workload is huge, and professional surveying and mapping equipment such as differential GPS instrument and total station needs to be used, and each camera needs to be manually surveyed and mapped, and the workload is huge.
  • the longitude and latitude of the camera are often marked manually on the three-dimensional electronic map. What's more, the longitude and latitude of the camera is replaced by the longitude and latitude of the target object (usually a building) where the camera is located, which is essentially the approximate longitude and latitude of the camera, and accurate positioning cannot be achieved.
  • the current camera positioning method makes it impossible to know the specific height of the camera, whether it is indoors or outdoors, the specific direction the camera is facing, etc. in the electronic map, and it is impossible to automatically determine which cameras for a specific target area through algorithms.
  • the target area can be covered, and it is impossible to automatically calculate whether there are other available surveillance camera resources in a specific surveillance video screen based on algorithms, so as to realize high-low point linkage video jumping.
  • one of the objectives of the present invention is to provide a method for precise positioning of a camera applied in an electronic map, which can solve the problem of the accurate position and attitude of the camera applied to the target area based on visual coverage in the electronic map. and the problem of internal parameters;
  • the second object of the present invention is to provide a processing terminal, which can solve the problem of accurate position, attitude and internal parameters of a camera applied to a target area based on visual coverage in an electronic map.
  • a technical solution for realizing one of the objectives of the present invention is: a method for precise positioning of a camera applied in an electronic map, comprising the following steps:
  • Step 1 Generate a 3D model of the target area according to the image data of the target area;
  • Step 2 Extract the first image feature point of each shot from the image data
  • Step 3 obtaining the shooting data of the camera to be positioned, and extracting any video frame from the shooting data, denoting it as the target video frame, and extracting the second image feature point of the target video frame;
  • Step 4 Compare the target video frame with each shooting picture of the image data, and use the shooting picture corresponding to the first image feature point with the highest matching degree as the shooting picture ref,
  • the pixel coordinates of the ref feature points of the shooting picture are respectively established with the pixel coordinates of each feature point of the target video frame and the three-dimensional coordinates of the sparse three-dimensional point cloud;
  • Step 5 Select an initial value of an internal parameter of the camera to be positioned, and determine the initial external parameter of the target video frame according to the initial value of the internal parameter;
  • Step 6 According to the initial external parameters, perform beam method adjustment processing with all the shooting pictures of the image data, and calculate the reprojection error;
  • Step 7 Update the initial value of the internal parameter to obtain a new initial value of the internal parameter
  • Step 8 Repeat steps 5-7 until the current new initial value of the internal parameter exceeds the preset value, then stop, and calculate the reprojection error corresponding to the initial value of each internal parameter, so as to obtain the internal parameter and attitude corresponding to the minimum reprojection error. and 3D position.
  • step 1 before performing the step 1, it also includes step 0: photographing the target area, and obtaining image data including the target area and the photographing position, posture and internal/external parameters of the photographing camera.
  • the target area is captured and photographed by a parade
  • the parade capture and photography include aerial drone tilt photography and road surface capture vehicle photography, or the drone shoots at high altitude and low altitude respectively.
  • the first image feature point is a corner point of a road sign marking or a general computer vision feature point.
  • mapping relationship is established between the pixel coordinates of the ref feature points of the shooting screen and the pixel coordinates of each feature point of the target video frame and the three-dimensional coordinates of the sparse three-dimensional point cloud, specifically,
  • the pixel coordinates of each feature point of the target video frame correspond to the pixel coordinates of the ref feature points of the shooting picture, and the pixel coordinates of the ref feature points of the shooting picture also correspond to the three-dimensional coordinates of the sparse three-dimensional point cloud.
  • the target video frame is compared with each shooting picture of the image data, and the first image feature point with the highest matching degree with the second image feature point is found, specifically,
  • a feature point word bag is established for each shooting picture, and through the feature point word bag, the feature points and positions of the target video frame and each shooting picture in the image data are matched, and the feature points corresponding to the second image are found.
  • the preset value in step 8 is the range of the field of view of the camera to be positioned.
  • step 8 it also includes,
  • Step 9 According to the final calculated internal parameters, attitude and three-dimensional position of the camera to be positioned, map the internal parameters, attitude and three-dimensional position to the three-dimensional model in step 1, and process through the spatial analysis of the three-dimensional map and the field of view analysis , to determine the specific height of the camera to be positioned in the three-dimensional electronic map and whether it is indoors or outdoors, and whether a certain target area in the three-dimensional electronic map can be covered by the camera to be positioned.
  • a processing terminal which includes:
  • the processor is used for running the program instructions to execute the steps of the camera precise positioning method applied to the electronic map.
  • the present invention can determine information such as the specific height of the camera, whether it is indoors or outdoors, the specific direction the camera is facing, etc. in the electronic three-dimensional map, and can automatically determine which cameras can be used for a specific target area. Cover the target area, and automatically calculate whether there are other available surveillance camera resources in a specific surveillance video screen, so as to realize the subsequent application of high-low point linkage video jumping.
  • FIG. 2 is a schematic diagram of a processing terminal.
  • a method for precise positioning of a camera applied to an electronic map includes the following steps:
  • Step 1 Take a tour of the target area to collect and shoot, and obtain the image data including the target area and the internal/external parameters of the shooting camera.
  • the target area is usually selected as an urban area of a city or a specific designated area.
  • the parade collection and photography include aerial drone tilt photography and road surface collection vehicle photography, such as street view collection vehicles, so that the three-dimensional position information and road surface information of each target object in the target area can be obtained, and the road surface information includes road signs and signs. Wire.
  • the three-dimensional position information and road surface information of each target object (including the road surface) in the target area can be obtained, thereby laying a foundation for subsequent conversion into a three-dimensional model.
  • a drone or a collection vehicle can be used for shooting, and the collection vehicle can also capture the three-dimensional position information and road surface information of the target object.
  • the drone When only the drone is used for shooting, the drone can be shot at high altitude and low altitude, so as to better obtain the three-dimensional position information of the target object and clearly shoot the road surface information.
  • the high-altitude and low-altitude here only refer to the relative height of the drone's shooting position, and do not limit the specific height of the drone's shooting position.
  • Step 2 Generate a three-dimensional model of the target area according to the obtained image data.
  • the shooting position, attitude and internal/external parameters of the shooting camera at any shooting point in the target area can be obtained.
  • the shooting camera also corresponds to the camera during the actual cruise acquisition and shooting.
  • the camera For the drone, it is the camera carried on the drone itself, and for the street view acquisition vehicle, it is the camera carried on the acquisition vehicle.
  • the image data a three-dimensional model of the target area is generated, which can be processed by existing photogrammetric modeling software.
  • Step 3 Extract the first image feature point of each shooting picture from the image data, establish a feature point word bag for each shooting picture according to the first image feature point, and classify the first image according to the internal parameters and external parameters of the shooting camera.
  • the image feature points in the 3D model generate sparse 3D point cloud by triangulation method, and the 3D model here refers to the 3D electronic map.
  • the first image feature point may be a corner point of a road sign and marking line extracted based on image recognition, or a general computer vision feature point of any one of ORB, SIFT, and SURF.
  • Step 4 Obtain the shooting data and position information of the camera to be positioned.
  • the shooting data is usually the video captured by the camera, and the second image feature point of any video frame is extracted from the shooting data, and extracted from the shooting data of the camera to be positioned.
  • the video frame is recorded as the target video frame, and the position information of the camera to be positioned can be obtained through a one-machine-one-file module, and its position information is rough latitude and longitude information.
  • the camera to be positioned is fixed on a building in a certain target area, and its position is fixed rather than moving. Therefore, the shooting angle of the camera is usually unchanged. Therefore, any one can be found from the shooting data. It is not necessary to extract each video frame, and then extract the second image feature point from the video frame.
  • the second image feature point and the first image feature point are the same type of image feature point, that is, if the first image feature point is the corner point of the road sign marking, the second image feature point is also the corner point of the road marking marking , the first image feature point is ORB, the second image feature point is also ORB, and the others are the same.
  • Step 5 Through the feature point word bag, the target video frame is matched with the feature points and positions of each shooting picture in the image data, and the first image feature point with the highest matching degree with the second image feature point is found, so as to obtain the matching degree
  • the shooting picture corresponding to the highest first image feature point and the shooting picture corresponding to the first image feature point with the highest matching degree are recorded as the shooting picture ref. That is, the second image feature point is matched and compared with the first image feature point corresponding to each captured picture in the feature point word bag, and the first image feature point with the highest matching degree is found, and the first image feature point corresponds to
  • the shooting screen is the shooting screen ref that needs to be found in this step.
  • the pixel coordinates of each feature point of the target video frame correspond to the pixel coordinates of the ref feature points of the shooting screen, and the pixel coordinates of the ref feature points of the shooting screen also correspond to a sparse three-dimensional point cloud.
  • the three-dimensional coordinates are the geographic coordinates including the height.
  • Step 6 Select an initial value of the internal parameter within the range of the angle of view of the camera to be positioned, and the initial value of the initial value of the internal parameter is preferably the lower limit of the range of the angle of view of the camera to be positioned.
  • the field of view of the camera has been set at the beginning of the factory. At present, the field of view of the camera is usually in the range of 30-150°. Therefore, the initial value of the internal parameter can be selected as 30°.
  • the initial value of the internal parameters the pixel coordinates and three-dimensional coordinates of the feature points matched by the target video frame, the initial external parameters of the target video frame are obtained through the PnP algorithm.
  • Step 7 According to the initial extrinsic parameters, perform beam adjustment processing with all the captured images of the image data, that is, optimize the intrinsic and extrinsic parameters through the beam adjustment, and calculate the reprojection error.
  • Step 8 On the basis of the previous initial value of the internal reference, add a preset value b to obtain a new initial value of the internal reference.
  • b 5°. If the field of view of the camera is in the range of 30-150°, the initial value of the internal parameter is selected as 30°, which is equivalent to selecting an initial value of the internal parameter every 5°, and then repeating steps 6 and 7 to obtain the corresponding initial value of each internal parameter.
  • the initial external parameters of , and the reprojection error corresponding to each initial external parameter stop until the initial value of the new internal parameter exceeds 150°, that is, the initial value of the internal parameter is limited to the range of the camera’s field of view, so that in all these reprojections Find the smallest reprojection error among the errors.
  • Step 9 Repeat steps 6-8 until the current new initial value of the internal parameter exceeds the field of view of the camera to be positioned, then stop, and calculate the reprojection error corresponding to the initial value of each internal parameter, thereby obtaining the minimum reprojection error corresponding to The intrinsic parameters, pose and 3D position of .
  • the camera to be positioned is a still shooting camera, that is, a camera that cannot rotate to shoot, but can only shoot in a specified direction
  • the position and attitude of the camera can be directly obtained through the minimum reprojection error.
  • the camera to be positioned is a rotating camera, that is, a camera that can be rotated for shooting in multiple directions
  • the posture at PT0 is calculated according to the PTZ value corresponding to the video frame extracted in step 4, and the position remains unchanged. change, so as to obtain the position and attitude information of the camera to be positioned.
  • Step 10 According to the final calculated internal parameters, attitude and three-dimensional position of the camera to be positioned, map the internal parameters, attitude and three-dimensional position to the three-dimensional model in step 2, and process through the spatial analysis and visual field analysis of the three-dimensional map. , so that the specific height of the camera to be positioned in the three-dimensional electronic map and whether it is indoors or outdoors can be effectively determined, and whether a certain target area in the three-dimensional electronic map can be covered by the camera.
  • the invention can realize the automatic positioning of the monitoring camera on the high-precision three-dimensional visual map (that is, the electronic three-dimensional map), can maximize the value of the oblique photography of the drone and the data collected by the street view of the collecting vehicle, and solve the problem of the conventional three-dimensional map model in the defects in actual use.
  • surveillance video users can determine the exact positions, attitudes and internal parameters (such as FOV, distortion parameters) of all surveillance video cameras within the coverage area of the visual map, and then based on the application of the accurate visual field of the video, such as global perception, panorama Applications such as backtracking, alarm correlation, pointing where to play, humanoid track, high-low linkage, and gun-ball linkage can be implemented on a large scale.
  • the present invention further provides a processing terminal 100, which includes:
  • the processor 102 is configured to run the program instructions to execute the steps of the camera precise positioning method applied to the electronic map.
  • These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions
  • the apparatus implements the functions specified in the flow or flow of the flowcharts and/or the block or blocks of the block diagrams.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Software Systems (AREA)
  • Instructional Devices (AREA)
  • Image Processing (AREA)
  • Studio Devices (AREA)

Abstract

A method for camera precise positioning method applied to an electronic map, and a processing terminal. The method comprises: step 1: obtaining image data from a target region; step 2: generating the target region into a target model; step 3: establishing a bag of words of feature points to generate first image feature points into a sparse three-dimensional point cloud; step 4: obtaining the photographing material and position information of a camera to be positioned, obtaining a target video frame, and extracting second image feature points; step 5: matching feature points with positions to find the photographed picture; step 6: selecting an intrinsic parameter initial value to obtain the initial extrinsic parameter of the target video frame; step 7: according to the initial extrinsic parameter, calculating a reprojection error; step 8: adding a preset value b to the intrinsic parameter initial value to obtain a new intrinsic parameter initial value; and step 9: repeating step 6 to step 8 to obtain an intrinsic parameter, a pose and a three-dimensional position corresponding to the minimum reprojection error. The invention allows to determine a specific height of the camera, whether used indoors or outdoors, and a specific direction the camera is facing.

Description

一种应用于电子地图中的摄像头精准定位方法及处理终端A kind of camera precise positioning method and processing terminal applied in electronic map 技术领域technical field
本发明涉及电子地图技术领域,具体涉及一种应用于电子地图中的摄像头精准定位方法及处理终端。The invention relates to the technical field of electronic maps, in particular to a camera precise positioning method and a processing terminal applied in electronic maps.
背景技术Background technique
在一个城市中,监控摄像头数十万计,即使是由行政机关使用的用于公共监控的摄像头也是数以万计。对于诸如交通用的监控摄像头、警用安防用的监控摄像头等而言,往往在电子地图调阅这些监控摄像头时,需要确定目标区域能够被哪些摄像头覆盖,但由于摄像头准确的位置、姿态、焦距、畸变参数等难以确定,往往很难做到这一点。因此,需要能够准确定位摄像头,这一需求在安防行业成为了迫切需求。In a city, there are hundreds of thousands of surveillance cameras, even those used by administrative agencies for public surveillance in the tens of thousands. For surveillance cameras for traffic, surveillance cameras for police security, etc., it is often necessary to determine which cameras can cover the target area when accessing these surveillance cameras on the electronic map. , distortion parameters, etc. are difficult to determine, and it is often difficult to do this. Therefore, the need to be able to accurately locate the camera has become an urgent need in the security industry.
目前实际使用中,通常会采用地理测绘的方式确定摄像头的GPS坐标,例如,安防监控视频联网平台上的一机一档模块,一机一档模块内存储有摄像头的位置信息(经纬度信息)。当采用这种摄像头定位方式,其工作量是巨大的,需要使用到差分GPS仪、全站仪等专业的测绘装备,而且需要每个摄像头采用人工进行地理测绘,工作量巨大。这也导致在目前电子地图特别是三维电子地图中,往往是由人工在三维电子地图上手动标记出摄像头的经纬度。更有甚者,通过摄像头所在的目标物体(通常为大厦)的经纬度来代替摄像头的经纬度,实质上是摄像头的近似经纬度,无法做到精准定位。At present, in actual use, the GPS coordinates of the camera are usually determined by means of geographic surveying and mapping. For example, the one-machine-one-file module on the security surveillance video networking platform stores the camera's location information (latitude and longitude information). When this camera positioning method is adopted, the workload is huge, and professional surveying and mapping equipment such as differential GPS instrument and total station needs to be used, and each camera needs to be manually surveyed and mapped, and the workload is huge. This also leads to the fact that in the current electronic map, especially the three-dimensional electronic map, the longitude and latitude of the camera are often marked manually on the three-dimensional electronic map. What's more, the longitude and latitude of the camera is replaced by the longitude and latitude of the target object (usually a building) where the camera is located, which is essentially the approximate longitude and latitude of the camera, and accurate positioning cannot be achieved.
目前的摄像头定位方法,导致在电子地图中无法知道摄像头的具 体高度、处于室内还是室外、摄像头朝向的具体方向等信息,也无法通过算法以实现自动对于某一特定的目标区域,能够确定哪些摄像头可以覆盖该目标区域,以及无法基于算法以实现自动计算出某一个具体监控视频画面中是否有其他可用的监控摄像头资源,以实现可进行高低点联动视频跳转。The current camera positioning method makes it impossible to know the specific height of the camera, whether it is indoors or outdoors, the specific direction the camera is facing, etc. in the electronic map, and it is impossible to automatically determine which cameras for a specific target area through algorithms. The target area can be covered, and it is impossible to automatically calculate whether there are other available surveillance camera resources in a specific surveillance video screen based on algorithms, so as to realize high-low point linkage video jumping.
发明内容SUMMARY OF THE INVENTION
针对现有技术的不足,本发明的目的之一提供一种应用于电子地图中的摄像头精准定位方法,其能够解决应用于电子地图中基于视觉覆盖的目标区域范围内的摄像头的准确位置、姿态和内参数的问题;In view of the deficiencies of the prior art, one of the objectives of the present invention is to provide a method for precise positioning of a camera applied in an electronic map, which can solve the problem of the accurate position and attitude of the camera applied to the target area based on visual coverage in the electronic map. and the problem of internal parameters;
本发明的目的之二提供一种处理终端,其能够解决应用于电子地图中基于视觉覆盖的目标区域范围内的摄像头的准确位置、姿态和内参数的问题。The second object of the present invention is to provide a processing terminal, which can solve the problem of accurate position, attitude and internal parameters of a camera applied to a target area based on visual coverage in an electronic map.
实现本发明的目的之一的技术方案为:一种应用于电子地图中的摄像头精准定位方法,包括如下步骤:A technical solution for realizing one of the objectives of the present invention is: a method for precise positioning of a camera applied in an electronic map, comprising the following steps:
步骤1:根据目标区域的影像资料,将目标区域生成三维模型;Step 1: Generate a 3D model of the target area according to the image data of the target area;
步骤2:从影像资料中提取每一个拍摄画面的第一图像特征点,Step 2: Extract the first image feature point of each shot from the image data,
将第一图像特征点在所述三维模型中生成稀疏三维点云;generating a sparse three-dimensional point cloud in the three-dimensional model from the first image feature points;
步骤3:获得待定位摄像头的拍摄资料,并从拍摄资料中提取任一视频帧,记为目标视频帧,提取目标视频帧的第二图像特征点;Step 3: obtaining the shooting data of the camera to be positioned, and extracting any video frame from the shooting data, denoting it as the target video frame, and extracting the second image feature point of the target video frame;
步骤4:将目标视频帧与影像资料的每一个拍摄画面进行比较,将匹配程度最高的第一图像特征点对应的拍摄画面作为拍摄画面ref,Step 4: Compare the target video frame with each shooting picture of the image data, and use the shooting picture corresponding to the first image feature point with the highest matching degree as the shooting picture ref,
拍摄画面ref特征点的像素坐标分别与目标视频帧的每个特征 点的像素坐标和稀疏三维点云的三维坐标建立映射关系;The pixel coordinates of the ref feature points of the shooting picture are respectively established with the pixel coordinates of each feature point of the target video frame and the three-dimensional coordinates of the sparse three-dimensional point cloud;
步骤5:选定待定位摄像头的一个内参初值,根据内参初值确定目标视频帧的初始外参;Step 5: Select an initial value of an internal parameter of the camera to be positioned, and determine the initial external parameter of the target video frame according to the initial value of the internal parameter;
步骤6:根据初始外参,与影像资料的所有拍摄画面进行光束法平差处理,计算得到重投影误差;Step 6: According to the initial external parameters, perform beam method adjustment processing with all the shooting pictures of the image data, and calculate the reprojection error;
步骤7:更新内参初值,得到新的内参初值;Step 7: Update the initial value of the internal parameter to obtain a new initial value of the internal parameter;
步骤8:重复步骤5-步骤7,直至当前新的内参初值超过预设值,则停止,计算得到每一个内参初值对应的重投影误差,从而得到最小重投影误差对应的内参数、姿态和三维位置。Step 8: Repeat steps 5-7 until the current new initial value of the internal parameter exceeds the preset value, then stop, and calculate the reprojection error corresponding to the initial value of each internal parameter, so as to obtain the internal parameter and attitude corresponding to the minimum reprojection error. and 3D position.
进一步地,执行所述步骤1之前,还包括步骤0:对目标区域进行拍摄,获得包括目标区域的影像资料和拍摄相机的拍摄位置、姿态和内/外参数。Further, before performing the step 1, it also includes step 0: photographing the target area, and obtaining image data including the target area and the photographing position, posture and internal/external parameters of the photographing camera.
进一步地,所述步骤0中,对目标区域采用巡游采集拍摄,巡游采集拍摄包括空中的无人机倾斜拍摄和路面的采集车拍摄,或,无人机分别在高空和低空进行拍摄。Further, in the step 0, the target area is captured and photographed by a parade, and the parade capture and photography include aerial drone tilt photography and road surface capture vehicle photography, or the drone shoots at high altitude and low altitude respectively.
进一步地,所述步骤2中,第一图像特征点为道路标志标线角点或通用计算机视觉特征点。Further, in the step 2, the first image feature point is a corner point of a road sign marking or a general computer vision feature point.
进一步地,所述拍摄画面ref特征点的像素坐标分别与目标视频帧的每个特征点的像素坐标和稀疏三维点云的三维坐标建立映射关系,具体为,Further, a mapping relationship is established between the pixel coordinates of the ref feature points of the shooting screen and the pixel coordinates of each feature point of the target video frame and the three-dimensional coordinates of the sparse three-dimensional point cloud, specifically,
目标视频帧每个特征点的像素坐标对应为拍摄画面ref特征点的像素坐标,拍摄画面ref特征点的像素坐标还对应有稀疏三维点云 的三维坐标。The pixel coordinates of each feature point of the target video frame correspond to the pixel coordinates of the ref feature points of the shooting picture, and the pixel coordinates of the ref feature points of the shooting picture also correspond to the three-dimensional coordinates of the sparse three-dimensional point cloud.
进一步地,所述将目标视频帧与影像资料的每一个拍摄画面进行比较,找到与第二图像特征点匹配程度最高的第一图像特征点,具体为,Further, the target video frame is compared with each shooting picture of the image data, and the first image feature point with the highest matching degree with the second image feature point is found, specifically,
根据第一图像特征点为每一个拍摄画面建立特征点词袋,通过特征点词袋,将目标视频帧与影像资料中的每一个拍摄画面进行特征点和位置匹配,找到与第二图像特征点匹配程度最高的第一图像特征点。According to the feature points of the first image, a feature point word bag is established for each shooting picture, and through the feature point word bag, the feature points and positions of the target video frame and each shooting picture in the image data are matched, and the feature points corresponding to the second image are found. The first image feature point with the highest matching degree.
进一步地,所述步骤8中的预设值为待定位摄像头的视场角范围。Further, the preset value in step 8 is the range of the field of view of the camera to be positioned.
进一步地,所述内参初值的初始值为待定位摄像头的视场角范围的下限值,更新内参初值为每次增加b=5°。Further, the initial value of the initial value of the internal reference is the lower limit of the range of the field of view of the camera to be positioned, and the initial value of the updated internal reference is increased by b=5° each time.
进一步地,所述步骤8之后,还包括,Further, after the step 8, it also includes,
步骤9:根据最终计算得到的待定位摄像头的内参数、姿态和三维位置,将内参数、姿态和三维位置映射到步骤1中的三维模型中,通过三维地图的空间分析和可视域分析处理,确定待定位摄像头在三维电子地图中的具体高度和处于室内还是室外,以及三维电子地图中某一目标区域是否能被所述待定位摄像头覆盖。Step 9: According to the final calculated internal parameters, attitude and three-dimensional position of the camera to be positioned, map the internal parameters, attitude and three-dimensional position to the three-dimensional model in step 1, and process through the spatial analysis of the three-dimensional map and the field of view analysis , to determine the specific height of the camera to be positioned in the three-dimensional electronic map and whether it is indoors or outdoors, and whether a certain target area in the three-dimensional electronic map can be covered by the camera to be positioned.
实现本发明的目的之二的技术方案为:一种处理终端,,其包括:The technical solution for realizing the second object of the present invention is: a processing terminal, which includes:
存储器,用于存储程序指令;memory for storing program instructions;
处理器,用于运行所述程序指令,以执行所述的应用于电子地图中的摄像头精准定位方法的步骤。The processor is used for running the program instructions to execute the steps of the camera precise positioning method applied to the electronic map.
本发明的有益效果为:本发明能够在电子三维地图中确定出摄像头的具体高度、处于室内还是室外、摄像头朝向的具体方向等信息, 并且自动对于某一特定的目标区域,能够确定哪些摄像头可以覆盖该目标区域,以及自动计算出某一个具体监控视频画面中是否有其他可用的监控摄像头资源,以实现后续可进行高低点联动视频跳转的应用。The beneficial effects of the present invention are: the present invention can determine information such as the specific height of the camera, whether it is indoors or outdoors, the specific direction the camera is facing, etc. in the electronic three-dimensional map, and can automatically determine which cameras can be used for a specific target area. Cover the target area, and automatically calculate whether there are other available surveillance camera resources in a specific surveillance video screen, so as to realize the subsequent application of high-low point linkage video jumping.
附图说明Description of drawings
图1为较佳实施例的流程示意图;1 is a schematic flowchart of a preferred embodiment;
图2为处理终端的示意图。FIG. 2 is a schematic diagram of a processing terminal.
具体实施方式Detailed ways
为了使本申请的目的、技术方案和优点更加清楚,下面结合附图对本申请具体实施例作进一步的详细描述。可以理解的是,此处所描述的具体实施例仅仅用于解释本申请,而非对本申请的限定。另外还需要说明的是,为了便于描述,附图中仅示出了与本申请相关的部分而非全部内容。在更加详细地讨论示例性实施例之前应当提到的是,一些示例性实施例被描述成作为流程图描绘的处理或方法。虽然流程图将各项操作(或步骤)描述成顺序的处理,但是其中的许多操作可以被并行地、并发地或者同时实施。此外,各项操作的顺序可以被重新安排。当其操作完成时所述处理可以被终止,但是还可以具有未包括在附图中的附加步骤。所述处理可以对应于方法、函数、规程、子例程、子程序等等。In order to make the objectives, technical solutions and advantages of the present application clearer, the specific embodiments of the present application will be further described in detail below with reference to the accompanying drawings. It should be understood that the specific embodiments described herein are only used to explain the present application, but not to limit the present application. In addition, it should be noted that, for the convenience of description, the drawings only show some but not all of the contents related to the present application. Before discussing the exemplary embodiments in greater detail, it should be mentioned that some exemplary embodiments are described as processes or methods depicted as flowcharts. Although a flowchart depicts various operations (or steps) as a sequential process, many of the operations may be performed in parallel, concurrently, or concurrently. Additionally, the order of operations can be rearranged. The process may be terminated when its operation is complete, but may also have additional steps not included in the figures. The processes may correspond to methods, functions, procedures, subroutines, subroutines, and the like.
如图1所示,一种应用于电子地图中的摄像头精准定位方法,包括如下步骤:As shown in Figure 1, a method for precise positioning of a camera applied to an electronic map includes the following steps:
步骤1:对目标区域进行巡游采集拍摄,获得包括目标区域的影像资料和拍摄相机的内/外参数,目标区域通常选为一个城市的城区 或某一特定指定区域。Step 1: Take a tour of the target area to collect and shoot, and obtain the image data including the target area and the internal/external parameters of the shooting camera. The target area is usually selected as an urban area of a city or a specific designated area.
本步骤中,巡游采集拍摄包括空中的无人机倾斜拍摄和路面的采集车拍摄,例如街景采集车,从而能够获得目标区域内各目标物体的三维位置信息和路面信息,路面信息包括道路标志标线。采用这种结合拍摄方式,能够获得目标区域内各个目标物体(包括路面)的三维位置信息和路面信息,从而为后续将其转换为三维模型打下基础。当然,也可以只采用无人机或采集车进行拍摄,采集车也能够拍摄得到目标物体的三维位置信息和路面信息。当只采用无人机拍摄时,可以将无人机分别在高空和低空进行拍摄,从而更好获得目标物体的三维位置信息和清晰拍摄到路面信息。这里的高空和低空仅是指无人机拍摄位置的相对高度,并不限制无人机拍摄位置的具体高度。In this step, the parade collection and photography include aerial drone tilt photography and road surface collection vehicle photography, such as street view collection vehicles, so that the three-dimensional position information and road surface information of each target object in the target area can be obtained, and the road surface information includes road signs and signs. Wire. Using this combined shooting method, the three-dimensional position information and road surface information of each target object (including the road surface) in the target area can be obtained, thereby laying a foundation for subsequent conversion into a three-dimensional model. Of course, only a drone or a collection vehicle can be used for shooting, and the collection vehicle can also capture the three-dimensional position information and road surface information of the target object. When only the drone is used for shooting, the drone can be shot at high altitude and low altitude, so as to better obtain the three-dimensional position information of the target object and clearly shoot the road surface information. The high-altitude and low-altitude here only refer to the relative height of the drone's shooting position, and do not limit the specific height of the drone's shooting position.
步骤2:根据获得的影像资料,将目标区域生成三维模型。三维模型中能够获得目标区域内任意一个拍摄点处拍摄相机的拍摄位置、姿态和内/外参数。Step 2: Generate a three-dimensional model of the target area according to the obtained image data. In the 3D model, the shooting position, attitude and internal/external parameters of the shooting camera at any shooting point in the target area can be obtained.
拍摄相机也即对应是实际巡游采集拍摄时的摄像头,对无人机而言,是无人机上自身携带的摄像头,对街景采集车而言,是采集车上携带的摄像头。根据影像资料,将目标区域生成三维模型,可以通过现有的摄影测量建模软件进行处理得到。The shooting camera also corresponds to the camera during the actual cruise acquisition and shooting. For the drone, it is the camera carried on the drone itself, and for the street view acquisition vehicle, it is the camera carried on the acquisition vehicle. According to the image data, a three-dimensional model of the target area is generated, which can be processed by existing photogrammetric modeling software.
步骤3:从影像资料中提取每一个拍摄画面的第一图像特征点,并根据第一图像特征点为每一个拍摄画面建立特征点词袋,以及根据拍摄相机的内参数和外参数将第一图像特征点在三维模型中通过三角测量方法生成稀疏三维点云,这里的三维模型是指三维电子地图。Step 3: Extract the first image feature point of each shooting picture from the image data, establish a feature point word bag for each shooting picture according to the first image feature point, and classify the first image according to the internal parameters and external parameters of the shooting camera. The image feature points in the 3D model generate sparse 3D point cloud by triangulation method, and the 3D model here refers to the 3D electronic map.
本步骤中,第一图像特征点可以是基于图像识别提取到的道路标志标线角点,或者是ORB、SIFT、SURF中任一种的通用计算机视觉特征点。In this step, the first image feature point may be a corner point of a road sign and marking line extracted based on image recognition, or a general computer vision feature point of any one of ORB, SIFT, and SURF.
步骤4:获得待定位摄像头的拍摄资料和位置信息,拍摄资料通常为摄像头拍摄到的视频,并从拍摄资料中提取任一视频帧的第二图像特征点,从待定位摄像头的拍摄资料中提取的视频帧记为目标视频帧,待定位摄像头的位置信息可以通过一机一档模块获得,其位置信息是粗略的经纬度信息。通常待定位摄像头是固定在某个目标区域的某个建筑物上,其位置是固定而非移动的,因此,摄像头的拍摄角度通常是不变的,因此,可以从拍摄资料中找出任一视频帧,而不用提取每一帧视频帧,再从该视频帧中提取出第二图像特征点。Step 4: Obtain the shooting data and position information of the camera to be positioned. The shooting data is usually the video captured by the camera, and the second image feature point of any video frame is extracted from the shooting data, and extracted from the shooting data of the camera to be positioned. The video frame is recorded as the target video frame, and the position information of the camera to be positioned can be obtained through a one-machine-one-file module, and its position information is rough latitude and longitude information. Usually the camera to be positioned is fixed on a building in a certain target area, and its position is fixed rather than moving. Therefore, the shooting angle of the camera is usually unchanged. Therefore, any one can be found from the shooting data. It is not necessary to extract each video frame, and then extract the second image feature point from the video frame.
其中,第二图像特征点和第一图像特征点为相同类型的图像特征点,也即若第一图像特征点为道路标志标线角点,第二图像特征点也为道路标志标线角点,第一图像特征点为ORB,第二图像特征点也为ORB,其他也是一样。The second image feature point and the first image feature point are the same type of image feature point, that is, if the first image feature point is the corner point of the road sign marking, the second image feature point is also the corner point of the road marking marking , the first image feature point is ORB, the second image feature point is also ORB, and the others are the same.
步骤5:通过特征点词袋,将目标视频帧与影像资料中的每一个拍摄画面进行特征点和位置匹配,找到与第二图像特征点匹配程度最高的第一图像特征点,从而得到匹配程度最高的第一图像特征点对应的拍摄画面,匹配程度最高的第一图像特征点对应的拍摄画面记为拍摄画面ref。也即是,将第二图像特征点分别与特征点词袋中每一个拍摄画面对应的第一图像特征点进行匹配比较,找到匹配程度最高的第一图像特征点,该第一图像特征点对应的拍摄画面即是本步骤需要 找到的拍摄画面ref。Step 5: Through the feature point word bag, the target video frame is matched with the feature points and positions of each shooting picture in the image data, and the first image feature point with the highest matching degree with the second image feature point is found, so as to obtain the matching degree The shooting picture corresponding to the highest first image feature point and the shooting picture corresponding to the first image feature point with the highest matching degree are recorded as the shooting picture ref. That is, the second image feature point is matched and compared with the first image feature point corresponding to each captured picture in the feature point word bag, and the first image feature point with the highest matching degree is found, and the first image feature point corresponds to The shooting screen is the shooting screen ref that needs to be found in this step.
通过本步骤的匹配结果,目标视频帧每个特征点的像素坐标(即二维图像坐标)对应为拍摄画面ref特征点的像素坐标,拍摄画面ref特征点的像素坐标还对应有稀疏三维点云的三维坐标,三维坐标也即是指包括高度在内的地理坐标。Through the matching results of this step, the pixel coordinates of each feature point of the target video frame (ie, the two-dimensional image coordinates) correspond to the pixel coordinates of the ref feature points of the shooting screen, and the pixel coordinates of the ref feature points of the shooting screen also correspond to a sparse three-dimensional point cloud. The three-dimensional coordinates are the geographic coordinates including the height.
步骤6:对待定位摄像头的视场角范围内选定一个内参初值,内参初值的初始值优选为待定位摄像头的视场角范围的下限值。摄像头的视场角范围在出厂初时已被设定,目前,摄像头的视场角范围通常在30-150°,因此,内参初值的初始值可以选为30°。根据该内参初值、目标视频帧匹配出的特征点的像素坐标和三维坐标,通过PnP算法得到目标视频帧的初始外参。Step 6: Select an initial value of the internal parameter within the range of the angle of view of the camera to be positioned, and the initial value of the initial value of the internal parameter is preferably the lower limit of the range of the angle of view of the camera to be positioned. The field of view of the camera has been set at the beginning of the factory. At present, the field of view of the camera is usually in the range of 30-150°. Therefore, the initial value of the internal parameter can be selected as 30°. According to the initial value of the internal parameters, the pixel coordinates and three-dimensional coordinates of the feature points matched by the target video frame, the initial external parameters of the target video frame are obtained through the PnP algorithm.
步骤7:根据初始外参,与影像资料的所有拍摄画面进行光束法平差处理,也即通过光束法平差对内参和外参进行优化,,计算得到重投影误差。Step 7: According to the initial extrinsic parameters, perform beam adjustment processing with all the captured images of the image data, that is, optimize the intrinsic and extrinsic parameters through the beam adjustment, and calculate the reprojection error.
步骤8:在上一次的内参初值的基础上,增加一个预设值b,得到新的内参初值。Step 8: On the basis of the previous initial value of the internal reference, add a preset value b to obtain a new initial value of the internal reference.
本步骤中,b=5°。若摄像头的视场角范围在30-150°,内参初值选定为30°,相当于每隔5°选定一个内参初值,然后重复步骤6和步骤7,得到每个内参初值对应的初始外参,以及每个初始外参对应的重投影误差,直至新的内参初值超过150°则停止,也即内参初值限定在摄像头的视场角范围内,从而在这些所有重投影误差中找到最小的重投影误差。In this step, b=5°. If the field of view of the camera is in the range of 30-150°, the initial value of the internal parameter is selected as 30°, which is equivalent to selecting an initial value of the internal parameter every 5°, and then repeating steps 6 and 7 to obtain the corresponding initial value of each internal parameter. The initial external parameters of , and the reprojection error corresponding to each initial external parameter, stop until the initial value of the new internal parameter exceeds 150°, that is, the initial value of the internal parameter is limited to the range of the camera’s field of view, so that in all these reprojections Find the smallest reprojection error among the errors.
步骤9:重复步骤6-步骤8,直至当前新的内参初值超过待定位摄像头的视场角范围,则停止,计算得到每一个内参初值对应的重投影误差,从而得到最小重投影误差对应的内参数、姿态和三维位置。Step 9: Repeat steps 6-8 until the current new initial value of the internal parameter exceeds the field of view of the camera to be positioned, then stop, and calculate the reprojection error corresponding to the initial value of each internal parameter, thereby obtaining the minimum reprojection error corresponding to The intrinsic parameters, pose and 3D position of .
其中,若待定位的摄像头是静止拍摄摄像头,也即摄像头无法转动拍摄,而只能朝某一指定方向进行拍摄的摄像头,则可通过最小重投影误差直接得到该摄像头的位置和姿态。若待定位的摄像头是转动拍摄摄像头,也即可转动进行多个方向进行拍摄的摄像头,则根据步骤4中提取得到的视频帧对应的PTZ值计算出PT0点时的姿态,而位置一直保持不变,从而得到待定位摄像头的位置和姿态信息。Among them, if the camera to be positioned is a still shooting camera, that is, a camera that cannot rotate to shoot, but can only shoot in a specified direction, the position and attitude of the camera can be directly obtained through the minimum reprojection error. If the camera to be positioned is a rotating camera, that is, a camera that can be rotated for shooting in multiple directions, then the posture at PT0 is calculated according to the PTZ value corresponding to the video frame extracted in step 4, and the position remains unchanged. change, so as to obtain the position and attitude information of the camera to be positioned.
步骤10:根据最终计算得到的待定位摄像头的内参数、姿态和三维位置,将内参数、姿态和三维位置映射到步骤2中的三维模型中,通过三维地图的空间分析和可视域分析处理,从而能够有效确定待定位摄像头在三维电子地图中的具体高度和处于室内还是室外,以及三维电子地图中某一目标区域是否能被该摄像头覆盖到。Step 10: According to the final calculated internal parameters, attitude and three-dimensional position of the camera to be positioned, map the internal parameters, attitude and three-dimensional position to the three-dimensional model in step 2, and process through the spatial analysis and visual field analysis of the three-dimensional map. , so that the specific height of the camera to be positioned in the three-dimensional electronic map and whether it is indoors or outdoors can be effectively determined, and whether a certain target area in the three-dimensional electronic map can be covered by the camera.
本发明能实现将监控摄像头在高精度三维视觉地图(也即是电子三维地图)上实现自动定位,能够最大程度发挥无人机倾斜摄影和采集车街景采集数据的价值,解决常规三维地图模型在实际使用中所存在的缺陷。基于本发明,监控视频使用者可以确定视觉地图覆盖范围内所有监控视频摄像头的准确位置、姿态和内参数(如FOV,畸变参数),进而基于视频准确可视域的应用,如全域感知、全景回溯、告警关联、指哪打哪、人形轨、高低联动和枪球联动等应用,才可能大规模落地。The invention can realize the automatic positioning of the monitoring camera on the high-precision three-dimensional visual map (that is, the electronic three-dimensional map), can maximize the value of the oblique photography of the drone and the data collected by the street view of the collecting vehicle, and solve the problem of the conventional three-dimensional map model in the defects in actual use. Based on the present invention, surveillance video users can determine the exact positions, attitudes and internal parameters (such as FOV, distortion parameters) of all surveillance video cameras within the coverage area of the visual map, and then based on the application of the accurate visual field of the video, such as global perception, panorama Applications such as backtracking, alarm correlation, pointing where to play, humanoid track, high-low linkage, and gun-ball linkage can be implemented on a large scale.
如图2所示,本发明还提供一种处理终端100,其包括:As shown in FIG. 2, the present invention further provides a processing terminal 100, which includes:
存储器101,用于存储程序指令;a memory 101 for storing program instructions;
处理器102,用于运行所述程序指令,以执行所述应用于电子地图中的摄像头精准定位方法的步骤。The processor 102 is configured to run the program instructions to execute the steps of the camera precise positioning method applied to the electronic map.
本发明是参照根据本发明实施例的方法、设备(系统)、和计算机程序产品的流程图和/或方框图来描述的。应理解可由计算机程序指令实现流程图和/或方框图中的每一流程和/或方框、以及流程图和/或方框图中的流程和/或方框的结合。可提供这些计算机程序指令到通用计算机、专用计算机、嵌入式处理机或其他可编程数据处理设备的处理器以产生一个机器,使得通过计算机或其他可编程数据处理设备的处理器执行的指令产生用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的装置。The present invention is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the invention. It will be understood that each flow and/or block in the flowchart illustrations and/or block diagrams, and combinations of flows and/or blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to the processor of a general purpose computer, special purpose computer, embedded processor or other programmable data processing device to produce a machine such that the instructions executed by the processor of the computer or other programmable data processing device produce Means for implementing the functions specified in a flow or flow of a flowchart and/or a block or blocks of a block diagram.
这些计算机程序指令也可存储在能引导计算机或其他可编程数据处理设备以特定方式工作的计算机可读存储器中,使得存储在该计算机可读存储器中的指令产生包括指令装置的制造品,该指令装置实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能。These computer program instructions may also be stored in a computer-readable memory capable of directing a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory result in an article of manufacture comprising instruction means, the instructions The apparatus implements the functions specified in the flow or flow of the flowcharts and/or the block or blocks of the block diagrams.
这些计算机程序指令也可装载到计算机或其他可编程数据处理设备上,使得在计算机或其他可编程设备上执行一系列操作步骤以产生计算机实现的处理,从而在计算机或其他可编程设备上执行的指令提供用于实现在流程图一个流程或多个流程和/或方框图一个方框或多个方框中指定的功能的步骤。These computer program instructions can also be loaded on a computer or other programmable data processing device to cause a series of operational steps to be performed on the computer or other programmable device to produce a computer-implemented process such that The instructions provide steps for implementing the functions specified in the flow or blocks of the flowcharts and/or the block or blocks of the block diagrams.
尽管已描述了本发明的优选实施例,但本领域内的技术人员一旦得知了基本创造性概念,则可对这些实施例作出另外的变更和修改。 所以,所附权利要求意欲解释为包括优选实施例以及落入本发明范围的所有变更和修改。Although preferred embodiments of the present invention have been described, additional changes and modifications to these embodiments may occur to those skilled in the art once the basic inventive concepts are known. Therefore, the appended claims are intended to be construed to include the preferred embodiment and all changes and modifications that fall within the scope of the present invention.
显然,本领域的技术人员可以对本发明进行各种改动和变型而不脱离本发明的精神和范围。这样,倘若本发明的这些修改和变型属于本发明权利要求及其等同技术的范围之内,则本发明也意图包括这些改动和变型在内。It will be apparent to those skilled in the art that various modifications and variations can be made in the present invention without departing from the spirit and scope of the invention. Thus, provided that these modifications and variations of the present invention fall within the scope of the claims of the present invention and their equivalents, the present invention is also intended to include such modifications and variations.

Claims (10)

  1. 一种应用于电子地图中的摄像头精准定位方法,其特征在于,包括如下步骤:A method for precise positioning of a camera applied to an electronic map, characterized in that it comprises the following steps:
    步骤1:根据目标区域的影像资料,将目标区域生成三维模型;Step 1: Generate a 3D model of the target area according to the image data of the target area;
    步骤2:从影像资料中提取每一个拍摄画面的第一图像特征点,将第一图像特征点在所述三维模型中生成稀疏三维点云;Step 2: extracting the first image feature point of each shooting picture from the image data, and generating a sparse three-dimensional point cloud from the first image feature point in the three-dimensional model;
    步骤3:获得待定位摄像头的拍摄资料,并从拍摄资料中提取任一视频帧,记为目标视频帧,提取目标视频帧的第二图像特征点;Step 3: obtaining the shooting data of the camera to be positioned, and extracting any video frame from the shooting data, denoting it as the target video frame, and extracting the second image feature point of the target video frame;
    步骤4:将目标视频帧与影像资料的每一个拍摄画面进行比较,将匹配程度最高的第一图像特征点对应的拍摄画面作为拍摄画面ref,Step 4: Compare the target video frame with each shooting picture of the image data, and use the shooting picture corresponding to the first image feature point with the highest matching degree as the shooting picture ref,
    拍摄画面ref特征点的像素坐标分别与目标视频帧的每个特征点的像素坐标和稀疏三维点云的三维坐标建立映射关系;A mapping relationship is established between the pixel coordinates of the ref feature points of the captured image and the pixel coordinates of each feature point of the target video frame and the 3D coordinates of the sparse 3D point cloud;
    步骤5:选定待定位摄像头的一个内参初值,根据内参初值确定目标视频帧的初始外参;Step 5: Select an initial value of an internal parameter of the camera to be positioned, and determine the initial external parameter of the target video frame according to the initial value of the internal parameter;
    步骤6:根据初始外参,与影像资料的所有拍摄画面进行光束法平差处理,计算得到重投影误差;Step 6: According to the initial external parameters, perform beam method adjustment processing with all the shooting pictures of the image data, and calculate the reprojection error;
    步骤7:更新内参初值,得到新的内参初值;Step 7: Update the initial value of the internal parameter to obtain a new initial value of the internal parameter;
    步骤8:重复步骤5-步骤7,直至当前新的内参初值超过预设值,则停止,计算得到每一个内参初值对应的重投影误差,从而得到最小重投影误差对应的内参数、姿态和三维位置。Step 8: Repeat steps 5-7 until the current new initial value of the internal parameter exceeds the preset value, then stop, and calculate the reprojection error corresponding to the initial value of each internal parameter, so as to obtain the internal parameter and attitude corresponding to the minimum reprojection error. and 3D position.
  2. 根据权利要求1所述的应用于电子地图中的摄像头精准定位方法,其特征在于,执行所述步骤1之前,还包括步骤0:对目标区域进行拍摄,获得包括目标区域的影像资料和拍摄相机的拍摄位置、 姿态和内/外参数。The method for accurate positioning of a camera applied to an electronic map according to claim 1, wherein before performing the step 1, the method further comprises step 0: photographing the target area, obtaining image data including the target area and the photographing camera The shooting position, attitude and internal/external parameters.
  3. 根据权利要求2所述的应用于电子地图中的摄像头精准定位方法,其特征在于,所述步骤0中,对目标区域采用巡游采集拍摄,巡游采集拍摄包括空中的无人机倾斜拍摄和路面的采集车拍摄,或,无人机分别在高空和低空进行拍摄。The method for precise positioning of a camera applied to an electronic map according to claim 2, wherein, in the step 0, the target area is captured and photographed by a parade, and the parade capture and photography include aerial drone tilt photography and road surface photography. Capture vehicle shooting, or drone shooting at high and low altitudes, respectively.
  4. 根据权利要求1所述的应用于电子地图中的摄像头精准定位方法,其特征在于,所述步骤2中,第一图像特征点为道路标志标线角点或通用计算机视觉特征点。The method for precise positioning of a camera applied to an electronic map according to claim 1, wherein, in the step 2, the first image feature point is a corner point of a road sign marking or a general computer vision feature point.
  5. 根据权利要求1所述的应用于电子地图中的摄像头精准定位方法,其特征在于,所述拍摄画面ref特征点的像素坐标分别与目标视频帧的每个特征点的像素坐标和稀疏三维点云的三维坐标建立映射关系,具体为,The method for accurate positioning of a camera applied to an electronic map according to claim 1, wherein the pixel coordinates of the ref feature points of the shooting screen are respectively the pixel coordinates of each feature point of the target video frame and the sparse three-dimensional point cloud. The three-dimensional coordinates of , establish a mapping relationship, specifically,
    目标视频帧每个特征点的像素坐标对应为拍摄画面ref特征点的像素坐标,拍摄画面ref特征点的像素坐标还对应有稀疏三维点云的三维坐标。The pixel coordinates of each feature point of the target video frame correspond to the pixel coordinates of the ref feature points of the shooting picture, and the pixel coordinates of the ref feature points of the shooting picture also correspond to the three-dimensional coordinates of the sparse three-dimensional point cloud.
  6. 根据权利要求1所述的应用于电子地图中的摄像头精准定位方法,其特征在于,所述将目标视频帧与影像资料的每一个拍摄画面进行比较,找到与第二图像特征点匹配程度最高的第一图像特征点,具体为,The method for precise positioning of a camera applied to an electronic map according to claim 1, wherein the target video frame is compared with each shot of the image data to find the one with the highest matching degree with the second image feature point. The first image feature point, specifically,
    根据第一图像特征点为每一个拍摄画面建立特征点词袋,通过特征点词袋,将目标视频帧与影像资料中的每一个拍摄画面进行特征点和位置匹配,找到与第二图像特征点匹配程度最高的第一图像特征点。According to the feature points of the first image, a feature point word bag is established for each shooting picture, and through the feature point word bag, the feature points and positions of the target video frame and each shooting picture in the image data are matched, and the feature points corresponding to the second image are found. The first image feature point with the highest matching degree.
  7. 根据权利要求1所述的应用于电子地图中的摄像头精准定位方法,其特征在于,所述步骤8中的预设值为待定位摄像头的视场角范围。The method for precise positioning of a camera applied to an electronic map according to claim 1, wherein the preset value in the step 8 is the range of the field of view of the camera to be positioned.
  8. 根据权利要求1所述的应用于电子地图中的摄像头精准定位方法,其特征在于,所述内参初值的初始值为待定位摄像头的视场角范围的下限值,更新内参初值为每次增加b=5°。The method for precise positioning of a camera applied to an electronic map according to claim 1, wherein the initial value of the initial value of the internal reference is the lower limit of the field of view range of the camera to be positioned, and the initial value of the updated internal reference is every Second increase b = 5°.
  9. 根据权利要求1所述的应用于电子地图中的摄像头精准定位方法,其特征在于,所述步骤8之后,还包括,The method for precise positioning of a camera applied in an electronic map according to claim 1, wherein after the step 8, the method further comprises:
    步骤9:根据最终计算得到的待定位摄像头的内参数、姿态和三维位置,将内参数、姿态和三维位置映射到步骤1中的三维模型中,通过三维地图的空间分析和可视域分析处理,确定待定位摄像头在三维电子地图中的具体高度和处于室内还是室外,以及三维电子地图中某一目标区域是否能被所述待定位摄像头覆盖。Step 9: According to the final calculated internal parameters, attitude and three-dimensional position of the camera to be positioned, map the internal parameters, attitude and three-dimensional position to the three-dimensional model in step 1, and process through the spatial analysis of the three-dimensional map and the field of view analysis , to determine the specific height of the camera to be positioned in the three-dimensional electronic map and whether it is indoors or outdoors, and whether a certain target area in the three-dimensional electronic map can be covered by the camera to be positioned.
  10. 一种处理终端,其特征在于,其包括:A processing terminal, characterized in that it includes:
    存储器,用于存储程序指令;memory for storing program instructions;
    处理器,用于运行所述程序指令,以执行如权利要求1、4-9任一项所述的应用于电子地图中的摄像头精准定位方法的步骤。The processor is configured to run the program instructions to execute the steps of the method for precise positioning of a camera applied in an electronic map according to any one of claims 1, 4-9.
PCT/CN2021/122607 2020-10-14 2021-10-08 Camera precise positioning method applied to electronic map, and processing terminal WO2022078240A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202011100418.1A CN112184890B (en) 2020-10-14 2020-10-14 Accurate positioning method of camera applied to electronic map and processing terminal
CN202011100418.1 2020-10-14

Publications (1)

Publication Number Publication Date
WO2022078240A1 true WO2022078240A1 (en) 2022-04-21

Family

ID=73950229

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/122607 WO2022078240A1 (en) 2020-10-14 2021-10-08 Camera precise positioning method applied to electronic map, and processing terminal

Country Status (2)

Country Link
CN (1) CN112184890B (en)
WO (1) WO2022078240A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115439469A (en) * 2022-10-12 2022-12-06 东南大学 Unmanned aerial vehicle-based building defect detection method and device and electronic equipment
CN115861546A (en) * 2022-12-23 2023-03-28 四川农业大学 Crop geometric perception and three-dimensional phenotype reconstruction method based on nerve body rendering
CN116630598A (en) * 2023-07-19 2023-08-22 齐鲁空天信息研究院 Visual positioning method and device under large scene, electronic equipment and storage medium
CN117459688A (en) * 2023-12-26 2024-01-26 海纳云物联科技有限公司 Camera angle marking method, device and medium based on map system

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112184890B (en) * 2020-10-14 2023-06-30 佳都科技集团股份有限公司 Accurate positioning method of camera applied to electronic map and processing terminal
CN115222808B (en) * 2021-06-30 2023-10-20 达闼机器人股份有限公司 Positioning method and device based on unmanned aerial vehicle, storage medium and electronic equipment
CN113920263A (en) * 2021-10-18 2022-01-11 浙江商汤科技开发有限公司 Map construction method, map construction device, map construction equipment and storage medium
CN114782550A (en) * 2022-04-25 2022-07-22 高德软件有限公司 Camera calibration method, device, electronic equipment and program product
CN115883812A (en) * 2022-11-30 2023-03-31 重庆大学 Multi-sensor cooperative station distribution method based on particle swarm optimization

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105869136A (en) * 2015-01-22 2016-08-17 北京雷动云合智能技术有限公司 Collaborative visual SLAM method based on multiple cameras
US20180297207A1 (en) * 2017-04-14 2018-10-18 TwoAntz, Inc. Visual positioning and navigation device and method thereof
CN109003305A (en) * 2018-07-18 2018-12-14 江苏实景信息科技有限公司 A kind of positioning and orientation method and device
CN110796706A (en) * 2019-11-08 2020-02-14 四川长虹电器股份有限公司 Visual positioning method and system
CN112184890A (en) * 2020-10-14 2021-01-05 佳都新太科技股份有限公司 Camera accurate positioning method applied to electronic map and processing terminal

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104596502B (en) * 2015-01-23 2017-05-17 浙江大学 Object posture measuring method based on CAD model and monocular vision
JP6641163B2 (en) * 2015-12-02 2020-02-05 日本放送協会 Object tracking device and its program
CN107507243A (en) * 2016-06-14 2017-12-22 华为技术有限公司 A kind of camera parameters method of adjustment, instructor in broadcasting's video camera and system
CN109945853B (en) * 2019-03-26 2023-08-15 西安因诺航空科技有限公司 Geographic coordinate positioning system and method based on 3D point cloud aerial image

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105869136A (en) * 2015-01-22 2016-08-17 北京雷动云合智能技术有限公司 Collaborative visual SLAM method based on multiple cameras
US20180297207A1 (en) * 2017-04-14 2018-10-18 TwoAntz, Inc. Visual positioning and navigation device and method thereof
CN109003305A (en) * 2018-07-18 2018-12-14 江苏实景信息科技有限公司 A kind of positioning and orientation method and device
CN110796706A (en) * 2019-11-08 2020-02-14 四川长虹电器股份有限公司 Visual positioning method and system
CN112184890A (en) * 2020-10-14 2021-01-05 佳都新太科技股份有限公司 Camera accurate positioning method applied to electronic map and processing terminal

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115439469A (en) * 2022-10-12 2022-12-06 东南大学 Unmanned aerial vehicle-based building defect detection method and device and electronic equipment
CN115439469B (en) * 2022-10-12 2024-03-22 东南大学 Unmanned aerial vehicle-based building defect detection method and device and electronic equipment
CN115861546A (en) * 2022-12-23 2023-03-28 四川农业大学 Crop geometric perception and three-dimensional phenotype reconstruction method based on nerve body rendering
CN115861546B (en) * 2022-12-23 2023-08-08 四川农业大学 Crop geometric perception and three-dimensional phenotype reconstruction method based on nerve volume rendering
CN116630598A (en) * 2023-07-19 2023-08-22 齐鲁空天信息研究院 Visual positioning method and device under large scene, electronic equipment and storage medium
CN116630598B (en) * 2023-07-19 2023-09-29 齐鲁空天信息研究院 Visual positioning method and device under large scene, electronic equipment and storage medium
CN117459688A (en) * 2023-12-26 2024-01-26 海纳云物联科技有限公司 Camera angle marking method, device and medium based on map system
CN117459688B (en) * 2023-12-26 2024-05-03 海纳云物联科技有限公司 Camera angle marking method, device and medium based on map system

Also Published As

Publication number Publication date
CN112184890B (en) 2023-06-30
CN112184890A (en) 2021-01-05

Similar Documents

Publication Publication Date Title
WO2022078240A1 (en) Camera precise positioning method applied to electronic map, and processing terminal
CN112767391B (en) Power grid line part defect positioning method integrating three-dimensional point cloud and two-dimensional image
CN109387186B (en) Surveying and mapping information acquisition method and device, electronic equipment and storage medium
CN106485785B (en) Scene generation method and system based on indoor three-dimensional modeling and positioning
CN109523471B (en) Method, system and device for converting ground coordinates and wide-angle camera picture coordinates
CN109520500B (en) Accurate positioning and street view library acquisition method based on terminal shooting image matching
CN115439424A (en) Intelligent detection method for aerial video image of unmanned aerial vehicle
KR102200299B1 (en) A system implementing management solution of road facility based on 3D-VR multi-sensor system and a method thereof
CN111540048A (en) Refined real scene three-dimensional modeling method based on air-ground fusion
CN112113542A (en) Method for checking and accepting land special data for aerial photography construction of unmanned aerial vehicle
KR101852368B1 (en) Method for underground information based on vrs geometric-correction used by uav taking picture
WO2022077296A1 (en) Three-dimensional reconstruction method, gimbal load, removable platform and computer-readable storage medium
CN113012292B (en) AR remote construction monitoring method and system based on unmanned aerial vehicle aerial photography
JP2017201261A (en) Shape information generating system
CN115423863B (en) Camera pose estimation method and device and computer readable storage medium
CN113066112A (en) Indoor and outdoor fusion method and device based on three-dimensional model data
CN108801225A (en) A kind of unmanned plane tilts image positioning method, system, medium and equipment
CN115004273A (en) Digital reconstruction method, device and system for traffic road
WO2020051208A1 (en) Method for obtaining photogrammetric data using a layered approach
Muji et al. Assessment of Digital Elevation Model (DEM) using onboard GPS and ground control points in UAV image processing
CN114659499B (en) Smart city 3D map model photography establishment method based on unmanned aerial vehicle technology
CN116129064A (en) Electronic map generation method, device, equipment and storage medium
US20160086339A1 (en) Method of providing cartograic information of an eletrical component in a power network
CN110617800A (en) Emergency remote sensing monitoring method, system and storage medium based on civil aircraft
Chen et al. 3D model construction and accuracy analysis based on UAV tilt photogrammetry

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21879285

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 21879285

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 12.09.2023)