CN106204656A - Target based on video and three-dimensional spatial information location and tracking system and method - Google Patents
Target based on video and three-dimensional spatial information location and tracking system and method Download PDFInfo
- Publication number
- CN106204656A CN106204656A CN201610579894.3A CN201610579894A CN106204656A CN 106204656 A CN106204656 A CN 106204656A CN 201610579894 A CN201610579894 A CN 201610579894A CN 106204656 A CN106204656 A CN 106204656A
- Authority
- CN
- China
- Prior art keywords
- video
- camera
- dimensional
- coordinate
- target
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30241—Trajectory
Landscapes
- Closed-Circuit Television Systems (AREA)
Abstract
The present invention provides a kind of and registrates the target location merged based on video and three-dimensional spatial information and follow the tracks of system and method, and the method includes: step 1, obtains scene and the video information of target by video camera;Step 2, sets up three-dimensional virtual scene, the video information of camera acquisition is fused in three-dimensional spatial information, forms the destination virtual scene with space coordinates;Step 3, gathers the information of moving target, obtains the space coordinates of target based on the scaling method demarcating thing, and in three-dimensional virtual scene, imaging shows.The target based on video with three-dimensional spatial information registration fusion that the application provides positions and follows the tracks of system and method, video information can be merged with three-dimensional spatial information, show with three-dimensional display visual pattern, observes for people and analyze.
Description
Technical field
The present invention relates to information fusion technology field, more particularly, to a kind of based on video and three-dimensional spatial information
Target location and tracking system and method.
Background technology
At present, the target location merged and tracking are registrated based on video and three-dimensional spatial information so that user is straight
See ground, from a virtual real platform, obtain spatial information positioning and the movement locus of moving target exactly.
Primary limitation is gone back from the application in every profession and trade of merging of current technology, video and three-dimensional spatial information technology
In to the management of various geography information key elements and two three-dimensional spatial information displaying aspects, analyzing, in terms of decision-making and professional service
Still in the stage of fumbling.How true to nature with tracked information for the space orientation of moving object complicated for macroscopic view is illustrated in people's eye
Before, how by spatial information wide range of services polynary for magnanimity in daily life, how by advanced video and space
The problem being to need to solve applied effectively by the platform of information fusion.
In recent years, location technology based on geospatial information is at public safety service, dynamic resource management, boats and ships pipe
The many-sides such as reason, navigation and intelligent transportation system (ITS) are all widely used, and location technology worldwide receives height
Degree attention and conduct in-depth research.
Common location algorithm is studied and is all based on the two dimension without video fusion or three-dimensional spatial information platform, but
Not only it needs to be determined that the coordinate of moving target in actual application, in addition it is also necessary to know that the concrete space of moving target positions and chases after
Track, particularly under the downtown area that skyscraper stands in great numbers or the yo-yo mountain environment of height, this demand is more urgent.
Summary of the invention
The present invention provide a kind of overcome the problems referred to above or at least in part solve the problems referred to above based on video and three-dimensional
The target that spatial information registration merges positions and follows the tracks of system and method.In order to solve the problems referred to above, according to perspective imaging with take the photograph
Shadow measures agreement principle, proposes based on three-dimensional scenic simulation imaging, and relies on the corresponding pass between video image with emulating image
System carries out the space-location method of target in geographic scenes.
According to an aspect of the present invention, it is provided that a kind of target location registrating fusion based on video and three-dimensional spatial information
It is characterized in that with tracking, including: step 1, obtain scene and the video information of target by video camera;Step 2, sets up
Three-dimensional virtual scene, is fused to the video information of camera acquisition in three-dimensional spatial information, forms the mesh with space coordinates
Mark virtual scene;Step 3, gathers the information of moving target, obtains the space coordinates of target based on the scaling method demarcating thing,
In three-dimensional virtual scene, imaging shows.
Another aspect according to the application, it is provided that a kind of target location registrating fusion based on video and three-dimensional spatial information
With the system of tracking, position including camera acquisition module, three-dimensional virtual scene module, three dimensions imaging wooden unit and moving target
With tracing module;Described camera acquisition module, for recorded video and transmission video, obtains scene and the video information of target;
Described three-dimensional virtual scene module is used for setting up three-dimensional virtual scene, and the video-splicing of camera acquisition is fused to three-dimensional space
Between, form the virtual scene of carrying space coordinate;Described three dimensions image-forming module, for all moving targets data acquisition with
And the imaging at three-dimensional virtual scene;Described coordinates of targets location and tracing module, for moving object at video and three-dimensional space
Between merge after platform on space orientation and movement locus follow the trail of.
The target based on video with three-dimensional spatial information registration fusion that the application provides positions and follows the tracks of system and method,
Video information can be merged with three-dimensional spatial information, show with three-dimensional display visual pattern, see for people
Examine, use and analyze.For example, it is possible to see from the platform that video information and three-dimensional spatial information merge certain moving object
The three-dimensional spatial information in certain moment and video information, it is also possible to follow the tracks of its movement locus.
The application can use various technological means, collects data through various channels, specifically includes adopting of spatial data
Collection and the collection of attribute data.The system of the application also has data management function, using the teaching of the invention it is possible to provide video data and three dimensions
The unification of data stores, retrieves, inquires about, revises function, and can edit platform in different range, has several and regards
Frequency splicing fusion function, and be automatically fused on three-dimensional spatial information.
The system of the application can realize calculating and the analytic function of data, and this function is video and three-dimensional spatial information is joined
The Core Feature of quasi-emerging system, by the distance measurements calculation etc. between computing function and the locus of GIS-Geographic Information System, comes
Calculate location and the track following of moving object.Further, to some extent solve on VR-Platform target
The problem of locating and tracking.
Accompanying drawing explanation
Fig. 1 is the space orientation schematic diagram of the monitoring objective according to the embodiment of the present invention;
Fig. 2 is registrating the target location merged based on video and three-dimensional spatial information and following the tracks of according to the embodiment of the present invention
The flow chart of method.
Detailed description of the invention
Below in conjunction with the accompanying drawings and embodiment, the detailed description of the invention of the present invention is described in further detail.Hereinafter implement
Example is used for illustrating the present invention, but is not limited to the scope of the present invention.
It is appreciated that three dimensions is the space that the position given directions is determined by three coordinates, the realistic space of objective reality
It is exactly three dimensions, there is length three kinds tolerance.Target location mentioned by the application refers to that target is in video and three-dimensional
The positional information location in system after spatial information registration fusion.The tracking that the application mentions refers to that following the tracks of target is regarding
The method of the location information in system after frequency and three-dimensional spatial information registration fusion.
Fig. 1 illustrate according to embodiments of the present invention registrate based on video and three-dimensional spatial information the target location merged and with
The overview flow chart of track method, as it is shown in figure 1, described method includes: step 1, obtains regarding of scene and target by video camera
Frequently information;Step 2, sets up three-dimensional virtual scene based on digital elevation model (Digital Elevation Model, DEM), will
The video fusion of camera acquisition, to three dimensions, forms the destination virtual scene with space coordinates;Step 3, gathers motion
The information of target, obtains the space coordinates of target based on the scaling method demarcating thing, and in three-dimensional virtual scene, imaging shows.Institute
Method of stating also includes step 4, is fused in three-dimensional digital earth system by multi-channel video registration.
Further, step 3 also includes: according to the space coordinates of each moment moving target, obtains moving target each
Displacement, speed and acceleration on three directions of moment, location and the track of pursuit movement target.
Wherein, in step 3, calculating in the three dimensions absolute coordinate of moving target, this method uses video camera based on mark
The scaling method of earnest.
Further, step 3 also includes, calculates principal point coordinate and aspect ratio first with plane template scaling method, then
Use Tsai video camera to carry out camera calibration, and calculate its projection matrix;In order to obtain accurate calibrating parameters, utilize first
Three dimensional space coordinate is projected to the difference of two dimensional image coordinate and true two dimensional image coordinate by the projection matrix obtained, by repeatedly
Generation optimizes tries to achieve the most accurate camera parameter;Finally utilize back projection's matrix and need the movement destination image coordinate of location
Calculate the three dimensions absolute coordinate position of moving target.
Wherein, in step 2, use and carry out target location based on video and three-dimensional information registration fusion.Fig. 2 is illustrated based on
The target that video and three-dimensional spatial information registration merge positions the flow chart with tracking, as in figure 2 it is shown, step 2 includes: enter
Row camera calibration;Solve the parameter that the coordinate system of camera model external parameter, i.e. video camera is relevant;Solve camera model
The focal length of inner parameter, i.e. video camera, distortion coefficients of camera lens etc.;Calculate the projection matrix between two video cameras;Calculate based on
The 3 d space coordinate of the platform after video and three-dimensional spatial information fusion, and export spatial orientation information and tracked information.
Further, in order to determine the three-dimensional geometry position of space object surface point and they are in the picture between corresponding point
Mutual relation, it is necessary to set up the geometric model of video camera imaging, these geometric model parameters are exactly camera parameters.
Utilize given object reference point coordinate (x, y, z) and the image coordinate of given object (u v) determines video camera
Internal geometry and optical characteristics (inner parameter) and video camera coordinate relation (external parameter) in three-dimensional world.Internal
Parameter includes the ginsengs such as lens focus f, distortion coefficients of camera lens (k, s, p), coordinate warping factor s and image coordinate initial point (u0, v0)
Number.External parameter includes the parameters such as camera coordinate system spin matrix R and translation vector T relative to world coordinate system.
Specifically, solve overdetermined linear system first with method of least square, try to achieve model external parameter;Then asking
By the horizontal spacing of the photosensitive unit of CCD (charge coupled device) array and longitudinal pitch as known parameters in solution preocess, solve takes the photograph
Camera inner parameter.
Position relationship between above-described calculating process, intrinsic parameters of the camera and two video cameras is the most
Through knowing, next calculate the projection matrix of two video cameras.
Make the position initial point as world coordinate system at first video camera place, the then projection matrix of two video cameras
For:
Wherein, K1,K2For camera intrinsic parameter matrix, R, T are the rotation between two video cameras and translation matrix, because
First video camera position is the initial point of world coordinate system, so R, T namely the second lattice video camera is for world coordinates
The rotation of system and translation matrix.
It follows that utilize in space the distance between known 2 to calculate scale factor, by triangle positioning principle meter
The moving target 3 d space coordinate calculated is multiplied by scale factor and obtains real relative spatial co-ordinates position.By same method
Obtain the space coordinates of each moment moving target, then calculate moving target displacement on three directions in each moment,
Speed, acceleration, just can trace into the track of moving target according to these data.
Wherein, in step 4, the method that multi-channel video registration is fused to three-dimensional digital earth system is included: by shooting
Machine calibration algorithm obtains the parameter of video camera, then utilizes camera parameter to ask for projection matrix, finally utilizes the former of triangle
Reason calculates spatial three-dimensional position.The method has a following characteristics: (1), be applicable between video camera the scene that has the overlapping ken, the most extremely
There are two video cameras less, and the space coordinates calculated is first video camera world coordinates initial point that be relative and that set
Position;(2), method be applicable to all scenes in indoor and outdoor because not using GPS, do not limited by indoor and outdoor;(3)、
This method also has the error of existence: camera calibration uses the method for self-calibration, and camera model does not consider abnormal
Situation about becoming, because the method that this method uses tradition demarcation, camera model uses band single order radial distortion
Model, so can bring error of coordinate, if needing higher precision, in addition it is also necessary to improve further.
Another embodiment according to the application, it is provided that a kind of target registrating fusion based on video and three-dimensional spatial information
Location and the system of tracking, this system includes camera acquisition module, three-dimensional virtual scene module, three dimensions imaging wooden unit and fortune
Moving targets location and tracing module.This system also includes three-dimensional digital earth Fusion Module.
Specifically, described camera acquisition module, for recorded video and transmission video, obtains scene and the video of target
Information.
Three-dimensional virtual scene module, is used for setting up three-dimensional virtual scene, and the video-splicing of camera acquisition is fused to three
Dimension space, forms the virtual scene of a carrying space coordinate.
Three dimensions image-forming module, the data acquisition for all moving targets and the imaging at three-dimensional virtual scene.
Coordinates of targets location and tracing module, the sky on moving object platform after video merges with three dimensions
Between location with movement locus follow the trail of.
Wherein, described three dimensions image-forming module is additionally operable to, based on video camera scaling method based on demarcation thing, calculate fortune
In the three dimensions absolute coordinate of moving-target.
Further, described three dimensions image-forming module also includes, calculates principal point first with plane template scaling method and sits
Mark and aspect ratio, then use Tsai video camera to carry out camera calibration, and calculate its projection matrix;In order to be marked accurately
Determine parameter, utilize the projection matrix obtained for the first time that three dimensional space coordinate projects to two dimensional image coordinate and sit with true two dimensional image
Target difference, tries to achieve the most accurate camera parameter by iteration optimization;Finally utilize back projection's matrix and need location
Movement destination image coordinate calculates the three dimensions absolute coordinate position of moving target.
Wherein, described coordinates of targets location is additionally operable to tracing module: carry out camera calibration;Solve outside camera model
The parameter that the coordinate system of portion's parameter, i.e. video camera is relevant;Solve focal length and the camera lens of camera model inner parameter, i.e. video camera
Distortion factor etc.;Calculate the projection matrix between two video cameras;Calculate putting down after merging with three-dimensional spatial information based on video
The 3 d space coordinate of platform, and export spatial orientation information and tracked information.
Further, in order to determine the three-dimensional geometry position of space object surface point and they are in the picture between corresponding point
Mutual relation, it is necessary to set up the geometric model of video camera imaging, these geometric model parameters are exactly camera parameters.
Utilize given object reference point coordinate (x, y, z) and the image coordinate of given object (u v) determines video camera
Internal geometry and optical characteristics (inner parameter) and video camera coordinate relation (external parameter) in three-dimensional world.Internal
Parameter includes the ginsengs such as lens focus f, distortion coefficients of camera lens (k, s, p), coordinate warping factor s and image coordinate initial point (u0, v0)
Number.External parameter includes the parameters such as camera coordinate system spin matrix R and translation vector T relative to world coordinate system.
Finally, the present processes is only preferably embodiment, is not intended to limit protection scope of the present invention.All
Within the spirit and principles in the present invention, any modification, equivalent substitution and improvement etc. made, should be included in the protection of the present invention
Within the scope of.
Claims (10)
1. target location and the tracking registrating fusion based on video and three-dimensional spatial information, it is characterised in that including:
Step 1, obtains scene and the video information of target by video camera;
Step 2, sets up three-dimensional virtual scene, the video information of camera acquisition is fused in three-dimensional spatial information, forms tool
There is the destination virtual scene of space coordinates;
Step 3, gathers the information of moving target, obtains the space coordinates of target based on the scaling method demarcating thing, three-dimensional empty
Intend imaging in scene to show.
Method the most according to claim 1, it is characterised in that described method also includes:
Step 4, is fused to multi-channel video registration in three-dimensional digital earth system.
Method the most according to claim 1, it is characterised in that step 3 also includes:
According to the space coordinates of each moment moving target, obtain moving target displacement on three directions of each moment, speed
Degree and acceleration, location and the track of pursuit movement target.
Method the most according to claim 1, it is characterised in that described step 3 also includes:
Plane template scaling method is utilized to calculate principal point coordinate and aspect ratio;
Use Tsai video camera to carry out camera calibration, and calculate its projection matrix;
Utilize the projection matrix obtained for the first time that three dimensional space coordinate projects to two dimensional image coordinate and true two dimensional image coordinate
Difference, try to achieve accurate camera parameter by iteration optimization;
Utilize back projection's matrix and need the movement destination image coordinate of location definitely to sit to the three dimensions calculating moving target
Cursor position.
Method the most according to claim 1, it is characterised in that described step 3 also includes:
Carry out camera calibration;
Solve camera model external parameter and inner parameter;
Calculate the projection matrix between two video cameras;
Calculate the 3 d space coordinate of the platform after merging with three-dimensional spatial information based on video, and export spatial orientation information
With tracked information.
Method the most according to claim 5, it is characterised in that obtain camera model in described step 3 and include:
The given reference point coordinate of object and the image coordinate of given object is utilized to determine the geometry within video camera and optics
Characteristic and the video camera coordinate relation in three-dimensional world;
Described inner parameter includes lens focus, distortion coefficients of camera lens, coordinate warping factor and image coordinate initial point parameter, described
External parameter includes that camera coordinate system is relative to the spin matrix of world coordinate system and translation vector.
Method the most according to claim 6, it is characterised in that the solution procedure bag of described inner parameter and external parameter
Include:
Utilize method of least square to solve overdetermined linear system, obtain model external parameter;
Horizontal spacing and the longitudinal pitch of photosensitive for charge coupled device array unit are joined as known parameters, the video camera solved inside
Number.
Method the most according to claim 7, it is characterised in that the solution procedure of described projection matrix includes:
Using the position at first video camera place as the initial point of world coordinate system, then the projection matrix of two video cameras is:
Wherein, K1,K2For camera intrinsic parameter matrix, R, T are the rotation between two video cameras and translation matrix;
Utilize the distance calculating scale factor between known 2, triangle positioning principle the motion mesh calculated in space
Mark 3 d space coordinate is multiplied by scale factor and obtains real relative spatial co-ordinates position.
9. target location and the system of tracking registrating fusion based on video and three-dimensional spatial information, it is characterised in that include
Camera acquisition module, three-dimensional virtual scene module, three dimensions imaging wooden unit and moving target location and tracing module;
Described camera acquisition module, for recorded video and transmission video, obtains scene and the video information of target;Described three
Dimension virtual scene module is used for setting up three-dimensional virtual scene, and the video-splicing of camera acquisition is fused to three dimensions, is formed
The virtual scene of carrying space coordinate;Described three dimensions image-forming module, for the data acquisition of all moving targets and three
The imaging of dimension virtual scene;Described coordinates of targets location and tracing module, merge with three dimensions at video for moving object
After platform on space orientation and movement locus follow the trail of.
System the most according to claim 9, it is characterised in that described three dimensions image-forming module is additionally operable to:
Plane template scaling method is utilized to calculate principal point coordinate and aspect ratio;
Use Tsai video camera to carry out camera calibration, and calculate its projection matrix;
Utilize the projection matrix obtained for the first time that three dimensional space coordinate projects to two dimensional image coordinate and true two dimensional image coordinate
Difference, try to achieve accurate camera parameter by iteration optimization;
Utilize back projection's matrix and need the movement destination image coordinate of location definitely to sit to the three dimensions calculating moving target
Cursor position.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610579894.3A CN106204656A (en) | 2016-07-21 | 2016-07-21 | Target based on video and three-dimensional spatial information location and tracking system and method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610579894.3A CN106204656A (en) | 2016-07-21 | 2016-07-21 | Target based on video and three-dimensional spatial information location and tracking system and method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106204656A true CN106204656A (en) | 2016-12-07 |
Family
ID=57491204
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610579894.3A Pending CN106204656A (en) | 2016-07-21 | 2016-07-21 | Target based on video and three-dimensional spatial information location and tracking system and method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106204656A (en) |
Cited By (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107645646A (en) * | 2017-07-07 | 2018-01-30 | 罗普特(厦门)科技集团有限公司 | A kind of regional monitoring system implementation method |
CN107680159A (en) * | 2017-10-16 | 2018-02-09 | 西北工业大学 | A kind of space non-cooperative target three-dimensional rebuilding method based on projection matrix |
CN108182218A (en) * | 2017-12-21 | 2018-06-19 | 深圳先进技术研究院 | A kind of video character recognition method, system and electronic equipment based on GIS-Geographic Information System |
CN108573522A (en) * | 2017-03-14 | 2018-09-25 | 腾讯科技(深圳)有限公司 | A kind of methods of exhibiting and terminal of flag data |
CN108572663A (en) * | 2017-03-08 | 2018-09-25 | 通用汽车环球科技运作有限责任公司 | Target following |
CN109598794A (en) * | 2018-11-30 | 2019-04-09 | 苏州维众数据技术有限公司 | The construction method of three-dimension GIS dynamic model |
CN109714567A (en) * | 2018-11-08 | 2019-05-03 | 中国船舶重工集团公司七五0试验场 | A kind of real-time construction method of three-dimensional virtual scene based on infrared viewing device and device |
CN109977853A (en) * | 2019-03-25 | 2019-07-05 | 太原理工大学 | A kind of mine group overall view monitoring method based on more identifiers |
CN110009561A (en) * | 2019-04-10 | 2019-07-12 | 南京财经大学 | A kind of monitor video target is mapped to the method and system of three-dimensional geographical model of place |
CN110148223A (en) * | 2019-06-03 | 2019-08-20 | 南京财经大学 | Monitor video target concentration expression and system in three-dimensional geography model of place |
CN110176022A (en) * | 2019-05-23 | 2019-08-27 | 广西交通科学研究院有限公司 | A kind of tunnel overall view monitoring system and method based on video detection |
CN110544302A (en) * | 2019-09-06 | 2019-12-06 | 广东工业大学 | Human body action reconstruction system and method based on multi-view vision and action training system |
CN110567460A (en) * | 2018-12-05 | 2019-12-13 | 昆明北理工产业技术研究院有限公司 | Unmanned platform indoor positioning system and positioning method |
CN110930508A (en) * | 2019-11-27 | 2020-03-27 | 西安应用光学研究所 | Two-dimensional photoelectric video and three-dimensional scene fusion method |
CN111046725A (en) * | 2019-10-24 | 2020-04-21 | 宝略科技(浙江)有限公司 | Spatial positioning method based on face recognition and point cloud fusion of surveillance video |
CN111062966A (en) * | 2019-11-05 | 2020-04-24 | 东北大学 | Method for optimizing camera tracking based on L-M algorithm and polynomial interpolation |
CN111273763A (en) * | 2018-12-05 | 2020-06-12 | 扩展知识公司 | Augmented reality implementation method and computer system for implementing augmented reality |
CN111836012A (en) * | 2020-06-28 | 2020-10-27 | 航天图景(北京)科技有限公司 | Video fusion and video linkage method based on three-dimensional scene and electronic equipment |
CN112033284A (en) * | 2020-08-28 | 2020-12-04 | 北京睿呈时代信息科技有限公司 | Memory, interactive measurement method, system and equipment based on monitoring video |
CN112037159A (en) * | 2020-07-29 | 2020-12-04 | 长安大学 | Cross-camera road space fusion and vehicle target detection tracking method and system |
CN112396633A (en) * | 2020-10-19 | 2021-02-23 | 北京理工大学 | Target tracking and track three-dimensional reproduction method and device based on single camera |
CN112489121A (en) * | 2019-09-11 | 2021-03-12 | 丰图科技(深圳)有限公司 | Video fusion method, device, equipment and storage medium |
CN113532390A (en) * | 2021-06-18 | 2021-10-22 | 广州领汇信息科技有限公司 | Target positioning method and device based on artificial intelligence technology and storage medium |
CN113553986A (en) * | 2021-08-02 | 2021-10-26 | 浙江索思科技有限公司 | Method and system for detecting moving target on ship |
CN113627005A (en) * | 2021-08-02 | 2021-11-09 | 成都视安创新科技有限公司 | Intelligent visual monitoring method |
CN113870163A (en) * | 2021-09-24 | 2021-12-31 | 埃洛克航空科技(北京)有限公司 | Video fusion method and device based on three-dimensional scene, storage medium and electronic device |
CN114460942A (en) * | 2022-02-09 | 2022-05-10 | 中国农业银行股份有限公司 | Indoor robot navigation control method, device, equipment and medium |
CN114872055A (en) * | 2022-07-12 | 2022-08-09 | 广东隆崎机器人有限公司 | SCARA robot assembly control method and system |
CN114979611A (en) * | 2022-05-19 | 2022-08-30 | 国网智能科技股份有限公司 | Binocular sensing system and method |
CN115393442A (en) * | 2022-03-02 | 2022-11-25 | 深圳市深视智能科技有限公司 | Calibration method, device and system of three-dimensional camera |
CN116580099A (en) * | 2023-07-14 | 2023-08-11 | 山东艺术学院 | Forest land target positioning method based on fusion of video and three-dimensional model |
CN116609124A (en) * | 2023-07-19 | 2023-08-18 | 北京建工环境修复股份有限公司 | Device and method for taking columnar mud sample underwater |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101951502A (en) * | 2010-10-19 | 2011-01-19 | 北京硅盾安全技术有限公司 | Three-dimensional intelligent video monitoring method |
CN102567989A (en) * | 2011-11-30 | 2012-07-11 | 重庆大学 | Space positioning method based on binocular stereo vision |
CN103226830A (en) * | 2013-04-25 | 2013-07-31 | 北京大学 | Automatic matching correction method of video texture projection in three-dimensional virtual-real fusion environment |
CN103716586A (en) * | 2013-12-12 | 2014-04-09 | 中国科学院深圳先进技术研究院 | Monitoring video fusion system and monitoring video fusion method based on three-dimension space scene |
-
2016
- 2016-07-21 CN CN201610579894.3A patent/CN106204656A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101951502A (en) * | 2010-10-19 | 2011-01-19 | 北京硅盾安全技术有限公司 | Three-dimensional intelligent video monitoring method |
CN102567989A (en) * | 2011-11-30 | 2012-07-11 | 重庆大学 | Space positioning method based on binocular stereo vision |
CN103226830A (en) * | 2013-04-25 | 2013-07-31 | 北京大学 | Automatic matching correction method of video texture projection in three-dimensional virtual-real fusion environment |
CN103716586A (en) * | 2013-12-12 | 2014-04-09 | 中国科学院深圳先进技术研究院 | Monitoring video fusion system and monitoring video fusion method based on three-dimension space scene |
Non-Patent Citations (1)
Title |
---|
石玉洁: ""视频监控中基于摄像机标定的运动目标三维空间定位技术的研究"", 《中国优秀硕士学位论文全文数据库 信息科技辑》 * |
Cited By (50)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108572663A (en) * | 2017-03-08 | 2018-09-25 | 通用汽车环球科技运作有限责任公司 | Target following |
CN108572663B (en) * | 2017-03-08 | 2021-12-21 | 通用汽车环球科技运作有限责任公司 | Target tracking |
CN108573522A (en) * | 2017-03-14 | 2018-09-25 | 腾讯科技(深圳)有限公司 | A kind of methods of exhibiting and terminal of flag data |
CN108573522B (en) * | 2017-03-14 | 2022-02-25 | 腾讯科技(深圳)有限公司 | Display method of mark data and terminal |
CN107645646A (en) * | 2017-07-07 | 2018-01-30 | 罗普特(厦门)科技集团有限公司 | A kind of regional monitoring system implementation method |
CN107680159A (en) * | 2017-10-16 | 2018-02-09 | 西北工业大学 | A kind of space non-cooperative target three-dimensional rebuilding method based on projection matrix |
CN108182218A (en) * | 2017-12-21 | 2018-06-19 | 深圳先进技术研究院 | A kind of video character recognition method, system and electronic equipment based on GIS-Geographic Information System |
CN109714567B (en) * | 2018-11-08 | 2021-05-18 | 中国船舶重工集团公司七五0试验场 | Real-time three-dimensional virtual scene construction method and device based on infrared night vision device |
CN109714567A (en) * | 2018-11-08 | 2019-05-03 | 中国船舶重工集团公司七五0试验场 | A kind of real-time construction method of three-dimensional virtual scene based on infrared viewing device and device |
CN109598794A (en) * | 2018-11-30 | 2019-04-09 | 苏州维众数据技术有限公司 | The construction method of three-dimension GIS dynamic model |
CN109598794B (en) * | 2018-11-30 | 2022-12-27 | 苏州维众数据技术有限公司 | Construction method of three-dimensional GIS dynamic model |
CN110567460A (en) * | 2018-12-05 | 2019-12-13 | 昆明北理工产业技术研究院有限公司 | Unmanned platform indoor positioning system and positioning method |
CN111273763A (en) * | 2018-12-05 | 2020-06-12 | 扩展知识公司 | Augmented reality implementation method and computer system for implementing augmented reality |
CN109977853B (en) * | 2019-03-25 | 2023-07-14 | 太原理工大学 | Underground worker panoramic monitoring method based on multiple identification devices |
CN109977853A (en) * | 2019-03-25 | 2019-07-05 | 太原理工大学 | A kind of mine group overall view monitoring method based on more identifiers |
CN110009561B (en) * | 2019-04-10 | 2023-04-18 | 南京财经大学 | Method and system for mapping surveillance video target to three-dimensional geographic scene model |
CN110009561A (en) * | 2019-04-10 | 2019-07-12 | 南京财经大学 | A kind of monitor video target is mapped to the method and system of three-dimensional geographical model of place |
CN110176022B (en) * | 2019-05-23 | 2023-03-28 | 广西交通科学研究院有限公司 | Tunnel panoramic monitoring system and method based on video detection |
CN110176022A (en) * | 2019-05-23 | 2019-08-27 | 广西交通科学研究院有限公司 | A kind of tunnel overall view monitoring system and method based on video detection |
CN110148223A (en) * | 2019-06-03 | 2019-08-20 | 南京财经大学 | Monitor video target concentration expression and system in three-dimensional geography model of place |
CN110148223B (en) * | 2019-06-03 | 2023-03-14 | 南京财经大学 | Method and system for concentrating and expressing surveillance video target in three-dimensional geographic scene model |
CN110544302A (en) * | 2019-09-06 | 2019-12-06 | 广东工业大学 | Human body action reconstruction system and method based on multi-view vision and action training system |
CN112489121A (en) * | 2019-09-11 | 2021-03-12 | 丰图科技(深圳)有限公司 | Video fusion method, device, equipment and storage medium |
CN111046725A (en) * | 2019-10-24 | 2020-04-21 | 宝略科技(浙江)有限公司 | Spatial positioning method based on face recognition and point cloud fusion of surveillance video |
CN111062966A (en) * | 2019-11-05 | 2020-04-24 | 东北大学 | Method for optimizing camera tracking based on L-M algorithm and polynomial interpolation |
CN111062966B (en) * | 2019-11-05 | 2024-01-02 | 东北大学 | Method for optimizing camera tracking based on L-M algorithm and polynomial interpolation |
CN110930508A (en) * | 2019-11-27 | 2020-03-27 | 西安应用光学研究所 | Two-dimensional photoelectric video and three-dimensional scene fusion method |
CN110930508B (en) * | 2019-11-27 | 2023-03-24 | 西安应用光学研究所 | Two-dimensional photoelectric video and three-dimensional scene fusion method |
CN111836012A (en) * | 2020-06-28 | 2020-10-27 | 航天图景(北京)科技有限公司 | Video fusion and video linkage method based on three-dimensional scene and electronic equipment |
CN111836012B (en) * | 2020-06-28 | 2022-05-13 | 航天图景(北京)科技有限公司 | Video fusion and video linkage method based on three-dimensional scene and electronic equipment |
CN112037159B (en) * | 2020-07-29 | 2023-06-23 | 中天智控科技控股股份有限公司 | Cross-camera road space fusion and vehicle target detection tracking method and system |
CN112037159A (en) * | 2020-07-29 | 2020-12-04 | 长安大学 | Cross-camera road space fusion and vehicle target detection tracking method and system |
CN112033284A (en) * | 2020-08-28 | 2020-12-04 | 北京睿呈时代信息科技有限公司 | Memory, interactive measurement method, system and equipment based on monitoring video |
CN112033284B (en) * | 2020-08-28 | 2022-05-17 | 北京睿呈时代信息科技有限公司 | Memory, interactive measurement method, system and equipment based on monitoring video |
CN112396633A (en) * | 2020-10-19 | 2021-02-23 | 北京理工大学 | Target tracking and track three-dimensional reproduction method and device based on single camera |
CN112396633B (en) * | 2020-10-19 | 2023-02-28 | 北京理工大学 | Target tracking and track three-dimensional reproduction method and device based on single camera |
CN113532390A (en) * | 2021-06-18 | 2021-10-22 | 广州领汇信息科技有限公司 | Target positioning method and device based on artificial intelligence technology and storage medium |
CN113553986A (en) * | 2021-08-02 | 2021-10-26 | 浙江索思科技有限公司 | Method and system for detecting moving target on ship |
CN113627005A (en) * | 2021-08-02 | 2021-11-09 | 成都视安创新科技有限公司 | Intelligent visual monitoring method |
CN113627005B (en) * | 2021-08-02 | 2024-03-26 | 成都视安创新科技有限公司 | Intelligent vision monitoring method |
CN113870163B (en) * | 2021-09-24 | 2022-11-29 | 埃洛克航空科技(北京)有限公司 | Video fusion method and device based on three-dimensional scene, storage medium and electronic device |
CN113870163A (en) * | 2021-09-24 | 2021-12-31 | 埃洛克航空科技(北京)有限公司 | Video fusion method and device based on three-dimensional scene, storage medium and electronic device |
CN114460942A (en) * | 2022-02-09 | 2022-05-10 | 中国农业银行股份有限公司 | Indoor robot navigation control method, device, equipment and medium |
CN115393442A (en) * | 2022-03-02 | 2022-11-25 | 深圳市深视智能科技有限公司 | Calibration method, device and system of three-dimensional camera |
CN114979611A (en) * | 2022-05-19 | 2022-08-30 | 国网智能科技股份有限公司 | Binocular sensing system and method |
CN114872055B (en) * | 2022-07-12 | 2022-11-18 | 广东隆崎机器人有限公司 | SCARA robot assembly control method and system |
CN114872055A (en) * | 2022-07-12 | 2022-08-09 | 广东隆崎机器人有限公司 | SCARA robot assembly control method and system |
CN116580099A (en) * | 2023-07-14 | 2023-08-11 | 山东艺术学院 | Forest land target positioning method based on fusion of video and three-dimensional model |
CN116609124A (en) * | 2023-07-19 | 2023-08-18 | 北京建工环境修复股份有限公司 | Device and method for taking columnar mud sample underwater |
CN116609124B (en) * | 2023-07-19 | 2023-09-29 | 北京建工环境修复股份有限公司 | Device and method for taking columnar mud sample underwater |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106204656A (en) | Target based on video and three-dimensional spatial information location and tracking system and method | |
CN108765328B (en) | High-precision multi-feature plane template and distortion optimization and calibration method thereof | |
CN106767706B (en) | A kind of unmanned plane reconnoitres the Aerial Images acquisition method and system of the scene of a traffic accident | |
CN110136208A (en) | A kind of the joint automatic calibration method and device of Visual Servoing System | |
CN108227929B (en) | Augmented reality lofting system based on BIM technology and implementation method | |
CN106990776B (en) | Robot homing positioning method and system | |
CN109872401B (en) | Unmanned aerial vehicle video augmented reality implementation method | |
CN105913439A (en) | Large-view-field camera calibration method based on laser tracker | |
CN103226838A (en) | Real-time spatial positioning method for mobile monitoring target in geographical scene | |
CN109712232B (en) | Object surface contour three-dimensional imaging method based on light field | |
CN109903382B (en) | Point cloud data fusion method and device | |
CN110146030A (en) | Side slope surface DEFORMATION MONITORING SYSTEM and method based on gridiron pattern notation | |
CN114283203B (en) | Calibration method and system of multi-camera system | |
CN102693543B (en) | Method for automatically calibrating Pan-Tilt-Zoom in outdoor environments | |
CN104318604A (en) | 3D image stitching method and apparatus | |
CN106370160A (en) | Robot indoor positioning system and method | |
CN105374067A (en) | Three-dimensional reconstruction method based on PAL cameras and reconstruction system thereof | |
TW201317544A (en) | Ground target geolocation system and method | |
CN110274612A (en) | A kind of pitching of remote optical turntable and orientation calibration method based on ground remote sensing photomap | |
CN114820924A (en) | Method and system for analyzing museum visit based on BIM and video monitoring | |
CN103260008B (en) | A kind of image position is to the projection conversion method of physical location | |
CN116309851B (en) | Position and orientation calibration method for intelligent park monitoring camera | |
US6922484B2 (en) | Information presentation apparatus and information presentation method | |
Zhao et al. | An ORB-SLAM3 Autonomous Positioning and Orientation Approach using 360-degree Panoramic Video | |
CN108592789A (en) | A kind of steel construction factory pre-assembly method based on BIM and machine vision technique |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20161207 |
|
RJ01 | Rejection of invention patent application after publication |