US20170138732A1 - Surveying by mobile vehicles - Google Patents

Surveying by mobile vehicles Download PDF

Info

Publication number
US20170138732A1
US20170138732A1 US15/348,886 US201615348886A US2017138732A1 US 20170138732 A1 US20170138732 A1 US 20170138732A1 US 201615348886 A US201615348886 A US 201615348886A US 2017138732 A1 US2017138732 A1 US 2017138732A1
Authority
US
United States
Prior art keywords
mobile vehicle
mobile
surveying
algorithm
light pattern
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US15/348,886
Other languages
English (en)
Inventor
Bo Pettersson
Knut Siercks
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hexagon Technology Center GmbH
Original Assignee
Hexagon Technology Center GmbH
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hexagon Technology Center GmbH filed Critical Hexagon Technology Center GmbH
Assigned to HEXAGON TECHNOLOGY CENTER GMBH reassignment HEXAGON TECHNOLOGY CENTER GMBH ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SIERCKS, KNUT, PETTERSSON, BO
Publication of US20170138732A1 publication Critical patent/US20170138732A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C11/00Photogrammetry or videogrammetry, e.g. stereogrammetry; Photographic surveying
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01CMEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
    • G01C15/00Surveying instruments or accessories not provided for in groups G01C1/00 - G01C13/00
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/0094Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots involving pointing a payload, e.g. camera, weapon, sensor, towards a fixed or moving target
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64CAEROPLANES; HELICOPTERS
    • B64C39/00Aircraft not otherwise provided for
    • B64C39/02Aircraft not otherwise provided for characterised by special use
    • B64C39/024Aircraft not otherwise provided for characterised by special use of the remote controlled vehicle type, i.e. RPV
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/24Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
    • G01B11/25Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/24Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
    • G01B11/25Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
    • G01B11/2513Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object with several lines being projected in more than one direction, e.g. grids, patterns
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/0011Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots associated with a remote control arrangement
    • G05D1/0022Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots associated with a remote control arrangement characterised by the communication link
    • H04N13/0253
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/254Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/18Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast
    • H04N7/181Closed-circuit television [CCTV] systems, i.e. systems in which the video signal is not broadcast for receiving images from a plurality of remote sources
    • B64C2201/123
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U10/00Type of UAV
    • B64U10/10Rotorcrafts
    • B64U10/13Flying platforms
    • B64U10/14Flying platforms with four distinct rotor axes, e.g. quadcopters
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64UUNMANNED AERIAL VEHICLES [UAV]; EQUIPMENT THEREFOR
    • B64U2101/00UAVs specially adapted for particular uses or applications
    • B64U2101/30UAVs specially adapted for particular uses or applications for imaging, photography or videography
    • G05D2201/0207

Definitions

  • Some embodiments of the present invention relates generally to a system for three dimensional surveying and to a method for three dimensional surveying.
  • unmanned mobile vehicles For example, unmanned airborne vehicles—also called UAVs or drones, unmanned ground vehicles—also called UGVs or rovers, unmanned surface vehicles—also called USVs or unmanned marine vessels, unmanned underwater vehicles—also called UUVs or underwater drones, or the like.
  • unmanned airborne vehicles also called UAVs or drones
  • unmanned ground vehicles also called UGVs or rovers
  • unmanned surface vehicles also called USVs or unmanned marine vessels
  • unmanned underwater vehicles also called UUVs or underwater drones, or the like.
  • mobile surviving vehicles which are at least partially or preferably fully autonomous.
  • Such an at least partially or fully autonomous usage means for example, that the vehicles can be controlled by high level commands like, “move to this position”, “survey this room or area”, “follow this object” “survey this track”, etc., wherein the mobile vehicle is capable to automatically complete this task and to avoid obstructions and crashes, automatically navigate in an area even when it is partially or fully unknown, automatically complement gaps in already gained surveying information, etc.
  • a human operator is not required to control every detail of the movement of the mobile vehicle, but only needs to provide general directives.
  • EP 2 511 656 shows a measurement system for determination of 3D coordinates of measurement points on an industrial product.
  • a scanning apparatus carried in an unmanned, controllable, automobile aircraft determines measurement points in an inner scanning coordinate system.
  • a referencing arrangement is providing referencing information of the scanning apparatus, for referencing the inner measurement point coordinates in the outer object coordinate system.
  • An evaluation unit determines 3D coordinates of the measurement points in the outer object coordinate system, such that the inner measurement point coordinates are in the form of 3D coordinates in the outer object coordinate system.
  • EP 2 511 659 shows a geodetic marking system for marking a known target point, with a self-propelled, unmanned, remotely controllable sighting unit with a geodetic position determining arrangement for a determination of the actual position of the sighting unit.
  • the sighting unit can be positioned at least temporarily in a hovering fashion largely fixed in position.
  • the sighting unit carries a marking unit for marking the target point, and a control unit, so that the sighting unit can be positioned in a defined desired position relative to the target point position as a function of the external actual position.
  • the control unit takes into account the actual position, the desired position and a defined marking direction from the marking unit to the target point, so that the target point can be marked with geodetic accuracy.
  • the patent application EP 13162632.7 shows a method for air image capture with an unmanned and controllable aircraft comprising a camera.
  • EP 0 251 178 shows an example of a geodetic measurement system having a geodetic measurement unit that comprises an emission unit with at least one radiation source for a collimated optical beam.
  • the measurement system also has a self-propelled, unmanned, controllable flight vehicle with an optical module which vehicle can be moved under control and can be positioned in a substantially fixed position.
  • An evaluation determines an actual state of the flight vehicle by a position, an alignment and/or a change in position in a coordinate system from an interaction of the optical beam with the optical module.
  • the measurement system has a control unit for controlling the flight vehicle in such a way that control data can be produced with an algorithm as a function of the actual state and a defined desired state, and an automatic control can bring the flight vehicle to the desired state.
  • EP 2 811 255, IL 193486 US 2010/139995, KR 101431383, KR 20030069709, JP 2002307340, CN 104080579 or CN 102819263 are showing examples of similar unmanned ground vehicles.
  • KR 2009 0069535, KR 2011 0052102, WO 2010/123380 or U.S. Pat. No. 6,802,236 are showing examples of similar unmanned water vehicles.
  • Such air, ground and/or water vehicles can all be mobile vehicles according to the invention.
  • a problem with above mentioned prior art is the fact, that a surveying and/or navigation by visual information from a camera, which for example comprises a structure from motion and/or a simultaneous location and mapping algorithm or the like depends on automatically detectable and distinguishable visual features within the camera image. If such features are rare, missing or ambiguous, the visual surveying algorithms have difficulties, become inaccurate or even produce wrong surveying results.
  • Some embodiments of the present invention improve a surveying by mobile vehicle, in particular in view of unfavorable conditions of the target surface to be surveyed.
  • Some embodiments of the invention gain more flexibility in a three dimensional surveying system based on a camera vision system, in particular in view making it more autonomous and adaptable to its environment.
  • Some embodiments improve the surveying result by unmanned mobile vehicle which are navigating at least partially through an at least partially unknown territory, which's surface has to be surveyed to gain three dimensional information.
  • Some embodiments of the invention provide a method and a surveying system by mobile vehicles which can survey a topology of optically uniform surfaces by imaging with one or more cameras, wherein the mobile vehicles are preferably moving at least partially autonomous.
  • Some embodiments of the present invention relate to a system for three dimensional surveying of a surface, which system comprises at least a first mobile vehicle with at least one camera unit having its field of view at least partially directed towards the surface for imaging the surface.
  • the system also comprises a computer-vision unit built to execute a visual surveying algorithm based on images from the camera unit for determining a topology of the surface and/or a location relative to the surface, in particular in form of a point cloud.
  • the system also comprises at least a second mobile vehicle, comprising a light projection unit built to emit a defined structured light pattern onto the surface.
  • the camera unit of the first mobile vehicle at least partially images the structured light pattern projected from the second mobile vehicle to the surface.
  • the computer-vision unit executes the visual surveying algorithm with an evaluation of the structured light pattern, in particular with an evaluation the first mobile vehicles images of the structured light pattern that is emitted from the second vehicle.
  • this means that the visual surveying at least partially bases on the fact that the camera unit images the projected pattern distorted according to the topographical relief of surface that is to be determined and whereof the three dimensional information of the surface can be determined.
  • the first and second mobile vehicles are separate from each other, which means that a relative position from at least one of the first mobile vehicle with respect to at least one of the second mobile vehicle can varied and controlled by at least one of the first mobile vehicles. Therefore, according to the invention, the origin of the projection at the second mobile vehicle can freely move with respect to the surface and/or with respect to the surveying camera at the first mobile vehicle. In other words, camera unit and projection unit are not mechanically fixed with respect to each other.
  • the visual surveying algorithm can comprise a structure from motion (SFM) algorithm, determining the topology of the surface.
  • SFM structure from motion
  • the visual surveying algorithm can alternatively or in addition also comprise a simultaneous location and mapping (SLAM) algorithm, determining the topology of the surface and the location relative to the surface.
  • SLAM simultaneous location and mapping
  • the first and the second mobile vehicle can be unmanned mobile vehicles, which in particular can be built to move autonomous under control of a comprised computation unit based on the determined topology of the surface and/or a location of the mobile vehicle relative to the surface.
  • least one of the first and/or second mobile vehicle can be an unmanned aerial vehicle (UAV).
  • UAV unmanned aerial vehicle
  • at least one of the first and at least one of the second mobile vehicles can be an UAV or all mobile vehicles can be UAVs.
  • At least one of the first and/or second mobile vehicles can comprise a location referencing unit, built to provide a position information of the mobile vehicle, in particular a relative position information with respect to another mobile vehicle and/or absolute position information with respect to geodetic-coordinates.
  • a location referencing unit built to provide a position information of the mobile vehicle, in particular a relative position information with respect to another mobile vehicle and/or absolute position information with respect to geodetic-coordinates.
  • the surveying system can thereby be aided by a GNSS location and navigation system, such as a navigation satellite receiver like GPS or Glonas, or another radio navigation system based on similar principles.
  • Technical implementation details can for example be handled similar as described in the above sited prior art or in a combination thereof.
  • At least one of the first and/or second mobile vehicles can also emit guidance light beam for guiding one or more of the remaining mobile vehicles, which are evaluating the guidance light beam or its projection on the surface, in particular wherein the guidance light beam can be coded by a modulation for transmitting remote control data for the evaluating mobile vehicle.
  • the emitted structured light pattern can be established by the light projection unit.
  • the light projection unit can therefore comprise a fixed and/or dynamic masking of the light projection, like a fixed mask, a hologram, a diffraction grating, a LCD or DLP projector, etc.
  • a moving beam can be used to establish the pattern, which can be done by scanning the beam along a desired path either with and/or without modulating the intensity of the moving beam during moving.
  • the structured light pattern can therein comprise a plurality of specks of different light intensity on the surface it is emitted to.
  • the structured light pattern can be changing over time, for example, the structured light pattern can comprise a predefined sequence of predefined patterns or sub-patterns, which can be known to the visual surveying algorithm or wherein the structured light pattern can comprise a sequence of random or pseudo-random structured light patterns.
  • the projected pattern can comprise fiducial markers, in particular uniformly seeded checkboard markers, which can be comprised in a random or pseudo-random speckle pattern or in a coded light pattern. Thereby a correspondence between projected and recovered pattern determined and considered in the evaluation.
  • Some embodiments of the present invention relate to an according method for three dimensional surveying of a surface comprising an imaging of at least part of the surface by a camera unit at a first mobile vehicle with the cameras field of view at least partially directed towards the surface, and a calculating of a visual surveying algorithm based on the images from the camera unit by a computer-vision unit and determining a topology of the surface and/or a location relative to the surface by the visual surveying algorithm, in particular in form of a point cloud.
  • an emitting of a defined structured light pattern onto the surface takes place by a light projection unit of at least one separate second mobile vehicle.
  • the imaging of the first mobile vehicle at least partially perceives the structured light pattern on the surface that is emitted from the second mobile vehicle.
  • the visual surveying algorithm is at least partially evaluating the structured light pattern.
  • At least one of the first and/or the second mobile vehicles can be unmanned aerial vehicles (UAV) which are airborne for the execution of the method, in particular wherein the first and/or the second mobile vehicle is at least partially autonomously moving and controlled by a comprised computation unit according to three dimensional information at least partially based on the results of the visual surveying algorithm.
  • UAV unmanned aerial vehicles
  • the visual surveying algorithm can comprises a structure from motion (SFM) algorithm determining the topology of the surface and/or a simultaneous location and mapping (SLAM) algorithm determining the topology of the surface and the location relative to the surface.
  • SFM structure from motion
  • SLAM simultaneous location and mapping
  • the first and second mobile vehicles can be moving with respect to each other. This moving can be controlled at least partially based on the results of the visual surveying algorithm.
  • the method according to the present invention can also be embodied as a computer program product, in particular as a computer program product that is stored on a machine readable medium or a computer-data-signal embodied as electromagnetic wave (such as wired or wireless data signal).
  • the computer program product implements a visual surveying algorithm which is at least partially evaluating digital images from a camera at a first mobile vehicle and which is evaluating a light pattern in the digital images that is resulting from a projection of a separate, second mobile vehicle which is emitting a defined structured light pattern at least partially into the field of view of the camera at the first mobile vehicle.
  • the computer program product calculates three dimensional spatial surveying information of a surface.
  • the computer program product can therein comprise code implementing a Structure From Motion (SFM) and/or Simultaneous Location And Mapping (SLAM) algorithm, in particular similar as known in the art but supplemented according to the invention by the marginal conditions and constraints given by the flexibly mobile projection and imaging of the separated vehicles, preferably wherein conditions given by a detection of the projection within the digital image is dissolving ambiguities of the determined three dimensional surveying data.
  • SFM Structure From Motion
  • SLAM Simultaneous Location And Mapping
  • the computer program product can implement a control of the movement of the first mobile vehicle and/or second mobile vehicle.
  • the computer program product can comprise code which implements an autonomous or semi-autonomous control of the first mobile vehicle and/or second mobile vehicle.
  • Some embodiments of the present invention relate to a computation means built to run a computer program providing functionality according to the invention, with or without the computer program actually loaded, in particular if comprised in a three dimensional surveying system with multiple mobile vehicles as described herein.
  • FIG. 1 shows an example of a first embodiment of a three dimensional surveying system according to the invention with two mobile air vehicles in an outdoor application;
  • FIG. 2 shows an example of a second embodiment of a three dimensional surveying system according to the invention with two mobile ground vehicles in an outdoor application;
  • FIG. 3 shows an example of a third embodiment of a three dimensional surveying system according to the invention with two mobile ground and two mobile air vehicles in an indoor application;
  • FIG. 4 shows an example of a fourth embodiment of a three dimensional surveying system according to the invention.
  • FIG. 5 shows an example of a fifth embodiment of a three dimensional surveying system according to the invention.
  • FIG. 6 shows an example of a simplified block diagram of a three dimensional surveying method according to the invention.
  • FIG. 7 shows an example of an embodiment of light pattern according to the invention.
  • FIG. 1 illustrates an example of an embodiment of a surveying system 1 according to the invention.
  • a first mobile vehicle 2 which is here shown as an unmanned airborne drone, also called unmanned aerial vehicle (UAV).
  • the shown first mobile vehicle 2 is built to be mobile in air, for example comprising on or more rotor blades 20 and a body 21 comprising sensors and a computation unit, e.g. a quadcopter capable of floating in midair into any desired direction or another airborne vehicle.
  • the first mobile vehicle 2 is equipped with a camera unit 3 that is according to the invention used for surveying purpose. The surveying is done at least partially based on images from the camera unit 3 , which are evaluated by a computer vision algorithm.
  • This evaluation is preferably done by the before mentioned onboard computation unit of the first mobile vehicle 2 , but can in another embodiment also at least partly be done by a remote, e.g. a ground based, computation unit.
  • a remote e.g. a ground based, computation unit.
  • An example of a field of view 24 of the camera unit 3 is shown by thick dashed lines.
  • the field of view 24 is here exemplary shown to be substantially cone-shaped, whereas in another embodiment, the field of view 24 can be shaped differently (e.g. substantially in form of a pyramid or the like).
  • the field of view 24 can not only be fixed but can also be variable in its direction with respect to the first mobile vehicle 2 and/or in its size.
  • the field of view 24 of the camera unit 3 is here directed toward ground, as the ground surface is the surface 5 which is desired to be surveyed in three dimensions, which in particular means to determine a topography of the surface 5 in form of three dimensional data.
  • it can be a task to survey the road 6 for potholes, settlements, etc. or to capture a topography of the houses 7 , of the hills 8 , of the vegetation 9 , and or other objects of which the surface topology is of interest.
  • the surface 5 to be surveyed is not static but variable, e.g. it would also be possible to follow a car 10 or another object for surveillance purposes or the like.
  • the surface-topology is surveyed according to images from the camera unit 3 by means of a visual surveying algorithm.
  • a visual surveying algorithm This means that, based on the images from the camera unit 3 , three dimensional information with respect to the portion of the surface 5 in the field of view 24 of the camera unit 3 or of a part of this field of view 24 is calculated by a computer vision unit.
  • the computer vision unit is preferably comprised in the mobile vehicle 2 , but can also be at least partially remote from the mobile vehicle 2 . As the first mobile vehicle 2 moves, the portion of the surface 5 in the field of view 24 of the camera unit 3 changes and information with respect to this movement can be provided to the computer vision unit to be incorporated in the visual surveying algorithm calculation.
  • SFM Structure From Motion
  • SLAM Simultaneous Localisation And Mapping
  • the surveying system 1 comprises at least one additional and separate second mobile vehicle 12 .
  • the second mobile vehicle 12 is also embodied as a UAV, e.g. like discussed above.
  • the first as well as the second mobile vehicles 2 , 12 can also comprise a random mixture of air, ground and/or water vehicles.
  • This second UAV 12 comprises a light projection unit 4 built to emit a defined structured light pattern 25 onto the surface 5 to be surveyed or at least on part of the surface 5 to be surveyed.
  • the part of the surface 5 where the light pattern is projected to is roughly about the same size as the field of view 24 of the camera unit 3 .
  • the projection is preferably kept to be substantially comprised within the field of view 24 of the camera unit 3 of the first UAV, in particular as an emitted pattern out of a surveying camera view 24 cannot contribute to the surveying.
  • the emission of the light pattern 25 is substantially directed onto a part of the surface 5 to be surveyed, where the surveying by a first mobile vehicle's camera unit 3 is providing unsatisfiable or ambiguous results when the light pattern is not present.
  • the invention can be described to concern surveying a target surface by imaging the surface with a camera unit at a first vehicle, which surface gets projected by a light pattern from a projection unit at a second vehicle, wherein the second vehicle is separate from the first vehicle and both first and second vehicle can move freely an independent with respect to each other.
  • the present invention there is no physical link between the camera unit and the projection unit.
  • the first and/or second mobile vehicle 2 , 12 can be moving during surveying.
  • both the first and the second UAVs 2 , 12 can progress along a desired track (e.g. the road 6 ) or can raster a desired area during the surveying.
  • at least one of the first and/or second mobile vehicles 2 , 12 can be a master vehicle, commanding one or more of the remaining mobile vehicles in their movements.
  • the computation unit which is processing the surveying can be embodied to also provide navigation information to the first and/or second mobile vehicles 2 , 12 .
  • the first and/or second moving vehicles 2 , 12 can comprise a, preferably wireless, communication interface to communicate in-between each other and/or with one or more remote stations.
  • the camera unit 3 of the first mobile vehicle 2 at least partially images the emitted light 25 from the second mobile vehicle 12 on the surface 5 or on part of this surface 5 .
  • the image of the projected light pattern 25 from the second mobile vehicle 12 on the surface, which is taken by the camera unit 3 can vary.
  • the computer-vision unit executes the visual surveying algorithm, which is built to at least partially evaluate a three dimensional survey or topography of the surface 5 .
  • FIG. 2 shows another example of an embodiment of a system 1 for three dimensional surveying of a surface according to the invention.
  • the mobile vehicles 2 , 12 are embodied as rover units moving on ground, like in form of wheeled or tracked vehicles.
  • the system 1 has the task to survey the shown pipeline 30 .
  • the task can e.g. comprise to check for movements, damages, deformation, displacements, etc. by surveying the shape, size and/or position of the outer surface of the pipeline 30 , which surveying results can be compared to desired values or values of previous measurements. Illegal tapping, overgrowing by vegetation or the like can also be detected in the surveying data.
  • the mobile vehicles 2 , 12 can for example move along a maintenance and emergency track running along the pipeline 30 .
  • the mobile vehicles 2 , 12 can e.g. also run inside of the pipeline 30 , and/or airborne mobile vehicles 2 , 20 can be used in addition or alternatively.
  • the mobile vehicles 2 , 12 can be autonomous or semi-autonomous, which means that for example they can be provided with the general task of following the pipeline 30 , whereby the mobile vehicles 2 , 12 will autonomously execute this task and automatically compute how they have to move to accomplish this task.
  • the uniformly painted outer surface of the pipeline 30 can be difficult to survey by image processing according to prior art, as the uniform surface can comprise too little visual features to for a classic visual surveying algorithm without an illumination pattern emitted towards it, as it is done according to the present invention.
  • the system 1 can have the task to survey the building 32 from outside and/or inside, e.g. a factory or a power plant, which it is potentially too hazardous for human workers after a disaster or in a war-scenery.
  • this is done by a system 1 of at least one first mobile vehicle 2 comprising a camera unit 3 and at least a second mobile vehicle 12 comprising a projection unit 4 for emitting light in a defined structured pattern.
  • the first and second mobile vehicles 2 , 12 are moving inside and/or outside of the building 32 , preferably autonomous—but alternatively also assisted or navigated by a human operator via a wired or wireless data link.
  • a computer-vision unit then executes a visual surveying algorithm based the images from the camera unit and determines three dimensional information, wherein the light pattern 25 emitted by the second mobile vehicle 12 which is at least partly comprised in the imaged field of view 24 of the camera unit 3 is evaluated by the visual surveying algorithm.
  • the visual surveying algorithm can therein gain additional information based on image of the projected pattern on the surface, and based on this additional information spatial surveying information can be determined and/or otherwise present ambiguities can be resolved, which both can result in faster and more accurate three dimensional surveying results.
  • the field of view of the camera unit 3 and/or the emission field of the projection unit 4 can be fixed with respect to the corresponding mobile vehicle 2 resp. 12 , but it can also be variable in its direction, size and/or shape, e.g. by a tilting, rotating and/or zooming unit.
  • FIG. 3 shows another an example of an embodiment of a three dimensional surveying system 1 according to the invention in an indoor application.
  • There are multiple mobile vehicles 2 a , 2 b , 12 a , 12 b wherein the mobile vehicles 2 a and 2 b are first mobile vehicles characterized by comprising at least one camera unit 3 for surveying and wherein the mobile vehicles 12 a , 12 b are second mobile vehicles characterized by comprising at least one projection unit 4 according to the above used terminology.
  • the mobile vehicles 2 a and 12 a are airborn, whereas the mobile vehicles 2 b and 12 b are ground vehicles. According to the invention, those mobile vehicles 2 a , 2 b , 12 a , 12 b collaborate in the surveying system and procedure.
  • a room 5 a with a table 5 b , furniture 5 c , a pole 5 d , a wall 5 e and a hallway 5 f shown, which's outer hulls—forming the three dimensional environment—can be considered to be the surface targeted to be surveyed.
  • the system 1 will thereby survey a 3D model of the room or of part of it, preferably with colour textures based on information from images from the camera units 3 .
  • the airborne mobile vehicles 2 a , 12 a are in particular predestined to survey top and side views, but might be unfavourable for view from below, like the floor beneath the table, the underside of the table, the ceiling, etc.
  • the flying drone 12 a as a second mobile vehicle comprises a projection unit 4 for emitting a light pattern, as indicated by the symbolized emission cone 25 a .
  • the light pattern 25 a can be a structure of lit and unlit areas, by projecting specks of different light intensity. This can for example be achieved by masking of the emitted light (like in an LCD-Projector), by directing the light (line in a DLP-Projector), by individually controlling multiple collimated light sources (like a laser projector), by controlling a deflection of a light beam along a desired path (like in a laser scanning projector), wherein in latter the light source can emit continuously and the deflection path defines the pattern or wherein the light source is modulated in its intensity (resp.
  • the optical axis of the projection unit 4 can be moveable with respect to the mobile vehicle 12 a which is carrying it, so that the emission can be directed towards a desired target area, independent of the pose and/or movement of the mobile vehicle 12 a.
  • the optical axis of the projection unit 4 and/or the projection size can also be moveable by moving the mobile vehicle 12 which comprises the projection unit 4 with respect to the targeted surface 5 .
  • the rover unit 12 b also comprises a projection unit 4 for emitting a light pattern indicated by 25 b.
  • the flying drone 2 a as a first mobile vehicle comprises a camera unit 3 built to capture electronic images or video sequences, e.g. comprising a CCD- or CMOS-sensor with an array of photosensitive pixels and an imaging optics.
  • the camera unit's images can be monochrome or comprise multiple colour channels, which can, in addition and/or alternatively to the visual spectrum, also comprise an infrared and/or ultraviolet range.
  • the camera unit 3 has a field of view 24 a in the direction of its optical axis, which field of view 24 a will be directed to substantially comprise at least partially the target to be surveyed. This directing can be achieved by moving the mobile vehicle 2 a with the camera unit 3 and/or by moving the optical axis of the camera unit 3 with respect to the mobile vehicle 2 a.
  • the mobile vehicles can be built to locate and/or identify each other.
  • the mobile vehicles can comprise tags 27 , which can be read out optically, e.g. by the camera unit 3 and by which an identification and/or location of the mobile vehicle can be established.
  • An identification and/or location of the mobile vehicles with respect to each other by means of radio signals is another option.
  • FIG. 4 shows an example of an embodiment of a mobile vehicle surveying system 1 according to the invention.
  • the mobile vehicles 2 a and 2 b comprise a camera unit 3 for surveying the topology of the surface 5 by a visual surveying algorithm based on images from the camera units 3 .
  • a corresponding coordinate system 17 is symbolized, in particular wherein the coordinate system 17 for the target surface 5 can be defined as an origin of the desired surveying coordinate system.
  • the relative and/or absolute location of those coordinate systems 17 with respect to each other are exemplary indicated by the shown dimensional lines. Those locations can for example be known by absolute and/or relative position tracking units at one or more of the mobile vehicles 2 a , 2 b , 12 .
  • those locations can also be determined according to the surveying capabilities of the system 1 according to the invention, in particular comprising a usage of a SLAM algorithm as visual surveying algorithm for navigation of a mobile vehicles and/or an image processing of one or more markers on another mobile vehicle in view of the camera unit 3 can be used to determine the locations.
  • a SLAM algorithm as visual surveying algorithm for navigation of a mobile vehicles and/or an image processing of one or more markers on another mobile vehicle in view of the camera unit 3 can be used to determine the locations.
  • the mobile vehicle 12 comprises a projection unit 4 , built for emitting a light pattern to the surface 5 , as indicated by the textured area 25 on the surface 5 and the corresponding emission cone from the second mobile vehicle 12 .
  • the guidance-beam 18 can comprise a modulation information for data transmission, whereby for example a command like a desired location relative to the guidance beam can be transmitted.
  • the guidance-beam following mobile vehicles 12 and/or 2 b can receive and evaluate the modulated data, for example by its camera unit. Thereby, the receiving mobile vehicle can for example be remote controlled in its location and/or movement by the emitting mobile vehicle.
  • the following mobile vehicle can be commanded to automatically move in such a way, to keep the light-spot 19 within a certain area or at a certain spot in its cameras field of view and/or to keep the light-spot 19 at a defined shape and/or size in the image of its cameras field of view.
  • the guidance beam can be formed by visible and/or invisible light and can be coded by colour, modulation, polarity, shape, size, etc. for example to address one specific mobile client vehicle.
  • the guidance-beam emitting mobile vehicle can therein be considered as a master vehicle, which is commanding one or more slave-vehicles.
  • the roles of master vehicle respectively slave vehicle can be associated either to a first mobile vehicle or to a second mobile vehicle according to the invention.
  • radio signals can exclusively or in addition be used for determining a location of the mobile vehicles with respect to one another like a multilateration as known from WIFI or GSM networks.
  • FIG. 5 an example of an embodiment according to the invention is shown, with three mobile vehicles 2 a , 2 b , 12 .
  • the second mobile vehicle 12 comprises a projection unit emitting a defined light pattern 25 toward the surface 5 to be three dimensionally surveyed.
  • the first mobile vehicles 2 a , 2 b each comprise camera units. In the shown example there are two camera units at each first mobile vehicle 2 a , 2 b and optionally also at the second mobile vehicle 12 .
  • the fields of view of the vehicles camera units are at least partially overlapping.
  • the overlapping of the fields of view can be used to calculate a stereo vision algorithm, either based on images of the cameras of only a single mobile vehicle and/or based on images from cameras of multiple of the mobile vehicles, with or without exact knowledge of their locations with respect to each other that gives a stereo basis.
  • This stereo vision algorithm can also be comprised in the visual surveying algorithm according to the invention, in particular wherein the projected pattern 25 from the second mobile vehicle 12 is incorporated in the computation of the three dimensional surveying.
  • the camera unit can only comprise a single field of view, or the fields of view of the cameras can be substantially non-overlapping.
  • a base station 41 emitting a guidance beam 43 towards at least one of the mobile vehicles 12 , and the receiving mobile vehicle is built to follow this guidance beam 43 .
  • the guidance beam can also be used to survey the location of the targeted mobile vehicle with respect to the base station 41 or transformed to another coordinate system, for example comprising electro-optical distance measurement in direction of the beam 43 and a determination angular coordinates of the emission direction of the beam 43 .
  • the beam 43 can also be modulated and comprise information as discussed below.
  • the computation of the spatial information and/or at least part of the surveying algorithms can for example also be done by the computation unit 42 .
  • navigation satellites 40 wherein one or more of the mobile vehicles 2 a , 2 b , 12 can be built to determine it's location on basis of electromagnetic signals from those satellites 40 . This can for example be done in addition or alternatively to the guidance by the base station 41 .
  • a guidance light beam from one mobile vehicle 12 to another of the mobile vehicles 2 b, respectively vice-versa.
  • the mobile vehicle 12 emits a beam 44 in the direction where the mobile client device 2 b should be.
  • the mobile vehicle 2 b comprises a receiver for this beam 44 and is built to follow this beam, for example by keeping the beams striking point and/or direction within a defined tolerance range.
  • This beam 44 can be modulated and comprise information, for example beside general data, also data for remote controlling the mobile vehicle 2 b in it's movements, in particular to navigate the mobile vehicle 2 b along the axis of the light beam—which is not defined by the direction of the emission alone.
  • the mobile vehicle 12 can be considered in the role of a guiding master and the mobile vehicle 2 b in the role of a following slave.
  • the guidance beam can also embody an electro-optic distance meter, which determines the distance between the mobile vehicles based on a time of flight or phase measurement.
  • the roles of the mobile vehicles 12 and 2 a , 2 b as first and second mobile vehicles could also be the other way round.
  • FIG. 6 an example of a block diagram of an embodiment of the invention is shown.
  • the three dimensional surveying of a surface 5 comprises an imaging of at least part of the surface 5 —as target to be surveyed—by a camera unit 3 at a first mobile vehicle 2 a , 2 b with its field of view 24 at least partially directed towards the surface 5 in block 71 .
  • an emitting a defined structured light pattern 25 onto the surface 5 by a light projection unit 4 is done from at least one second mobile vehicle 12 , as indicated by block 72 .
  • the second mobile vehicle 12 is separate and freely movable with respect to the first mobile vehicle 2 a , 2 b.
  • a calculating of a visual surveying algorithm based on images from the camera unit 3 by a computer-vision unit is done for determining a topology of the surface 5 and/or a location relative to the surface 5 , for example resulting in a point cloud representing a three dimensional topography of the surface 5 .
  • the imaging of the first mobile vehicle 2 a , 2 b at least partially perceives the structured light pattern 25 from the second mobile vehicle 12 as indicated in block 74
  • the visual surveying algorithm is at least partially evaluating the structured light pattern 25 for determining three dimensional information of the target in the visual surveying algorithm, as indicated in block 75 .
  • a commanding of a movement of the first mobile vehicle 2 a , 2 b and/or of the second mobile vehicle 12 can be done as indicated by block 76 and block 77 .
  • the projection of the pattern can exclusively target areas of the surface 5 which are otherwise difficult or ambiguous in a visual surveying algorithm
  • the direction and/or size of the projected pattern can be changed by moving the second mobile vehicle to fit to the shape of the surface 5
  • the projected pattern can vary within the image to gain additional information while the pattern itself is static but by movements of the second mobile vehicle, etc.
  • FIG. 7 an example of an embodiment according to the invention is shown, in which an exemplary structured pattern 25 c, which is emitted by the second mobile vehicle 12 c.
  • the second mobile vehicle 12 c comprises also a camera unit that is adopted to have the field of view 24 c .
  • the structured pattern 25 c comprises visual fiducial markers 50 and/or a checkboard pattern, which can be non-ambiguously recognized and surveyed in a camera image by a visual surveying algorithm, preferably also when the projected pattern 25 c is distorted by an uneven projection surface.
  • this visual surveying is done by the mobile vehicle 2 a and/or the mobile vehicle 2 b, which comprise camera units with the shown field of view 24 b respectively the field of view 24 a.
  • the second mobile vehicle 12 c can comprise a camera, as indicated by the field of view 24 c.
  • the projected pattern 25 c can be static or it can be dynamically changing, either according to a defined sequence of defined patterns or in random or pseudo-random.
  • the mobile vehicles can move in a desired formation and/or along desired paths during the surveying.
  • the formation and/or the paths of movement of each of the vehicles can be adapted dynamically, preferably by an autonomous or semi-autonomous control computed by a computation unit, in particular wherein the control is built in such a way to improve and/or complete the surveying result.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Multimedia (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Signal Processing (AREA)
  • Automation & Control Theory (AREA)
  • Electromagnetism (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
  • Length Measuring Devices By Optical Means (AREA)
US15/348,886 2015-11-12 2016-11-10 Surveying by mobile vehicles Abandoned US20170138732A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP15194351.1 2015-11-12
EP15194351.1A EP3168704B1 (de) 2015-11-12 2015-11-12 3d überwachung einer fläche durch mobile fahrzeuge

Publications (1)

Publication Number Publication Date
US20170138732A1 true US20170138732A1 (en) 2017-05-18

Family

ID=54542065

Family Applications (1)

Application Number Title Priority Date Filing Date
US15/348,886 Abandoned US20170138732A1 (en) 2015-11-12 2016-11-10 Surveying by mobile vehicles

Country Status (3)

Country Link
US (1) US20170138732A1 (de)
EP (1) EP3168704B1 (de)
CN (1) CN107024199A (de)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140210663A1 (en) * 2011-04-14 2014-07-31 Hexagon Technology Center Gmbh Measuring system and method for determining new points
US20180197300A1 (en) * 2015-07-17 2018-07-12 Nec Corporation Irradiation system, irradiation method, and program storage medium
US10152891B2 (en) * 2016-05-02 2018-12-11 Cnh Industrial America Llc System for avoiding collisions between autonomous vehicles conducting agricultural operations
US20190019051A1 (en) * 2016-03-23 2019-01-17 JVC Kenwood Corporation Unmanned mobile apparatus capable of transferring imaging, method of transferring
EP3435030A1 (de) * 2017-07-27 2019-01-30 Testo SE & Co. KGaA Verfahren zur erstellung eines 3d-modells von einem objekt
CN109658450A (zh) * 2018-12-17 2019-04-19 武汉天乾科技有限责任公司 一种基于无人机的快速正射影像生成方法
US20190127083A1 (en) * 2017-10-27 2019-05-02 Drone Delivery Canada Corp. Unmanned aerial vehicle and method for indicating a landing zone
US10382668B2 (en) * 2017-11-14 2019-08-13 Conary Enterprise Co., Ltd. Laser collimator module on a mobile device for image measurement
US10417469B2 (en) * 2016-05-07 2019-09-17 Morgan E. Davidson Navigation using self-describing fiducials
US20190295423A1 (en) * 2018-03-26 2019-09-26 D2, Llc Method and system for generating aerial imaging flight path
US11036946B2 (en) * 2016-05-07 2021-06-15 Canyon Navigation, LLC Navigation using self-describing fiducials
US20210311205A1 (en) * 2016-05-07 2021-10-07 Canyon Navigation, LLC Navigation Using Self-Describing Fiducials
US11294456B2 (en) * 2017-04-20 2022-04-05 Robert C. Brooks Perspective or gaze based visual identification and location system
US11327149B2 (en) * 2017-07-10 2022-05-10 Aurora Flight Sciences Corporation Laser speckle system and method for an aircraft
US11346665B2 (en) 2018-11-21 2022-05-31 Guangzhou Xaircraft Technology Co., Ltd Method and apparatus for planning sample points for surveying and mapping, control terminal, and storage medium
US11513029B2 (en) * 2016-04-25 2022-11-29 Siemens Aktiengesellschaft Moving flying object for scanning an object, and system for analyzing damage to the object

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109196441A (zh) * 2016-05-16 2019-01-11 深圳市大疆创新科技有限公司 用于协调设备动作的系统和方法
WO2018080969A1 (en) * 2016-10-24 2018-05-03 Agco International Gmbh Land mapping and guidance system
US10495421B2 (en) 2017-08-25 2019-12-03 Aurora Flight Sciences Corporation Aerial vehicle interception system
US11064184B2 (en) * 2017-08-25 2021-07-13 Aurora Flight Sciences Corporation Aerial vehicle imaging and targeting system
US10890920B2 (en) 2018-02-15 2021-01-12 Aptiv Technologies Limited Vehicle map-data gathering system and method
DE102018002499A1 (de) 2018-03-22 2019-09-26 Fachhochschule Dortmund Verfahren und Vorrichtung zur Bestimmung der Position von Objekten in einer dreidimensionalen Umgebung
US11958579B2 (en) * 2019-07-30 2024-04-16 Schlumberger Technology Corporation System and method for autonomous exploration for mapping underwater environments
CN111338383B (zh) * 2020-04-24 2023-10-13 北京泛化智能科技有限公司 基于gaas的自主飞行方法及系统、存储介质
CN113377106B (zh) * 2021-06-09 2022-03-15 安徽信息工程学院 基于平板电脑的室内图像测绘系统、方法

Family Cites Families (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE3789634T2 (de) 1986-06-27 1994-08-04 Fuji Photo Film Co Ltd Verfahren zum Zuführen von Regenerationsflüssigkeit in einem automatischen Entwicklungsgerät.
JP2002307340A (ja) 2001-04-19 2002-10-23 Sony Corp 脚式移動ロボット及びその制御方法
KR100425695B1 (ko) 2002-02-22 2004-04-01 엘지전자 주식회사 추락방지용 보행로봇 및 그의 제어방법
US6802236B1 (en) 2003-01-21 2004-10-12 The United States Of America As Represented By The Secretary Of The Navy System for in-stride identification of minelike contacts for surface countermeasures
EP1607710A1 (de) * 2004-06-18 2005-12-21 Saab Ab System zur Zielentfernungsermittlung für eine Laserlenkungswaffe
WO2007051972A1 (en) * 2005-10-31 2007-05-10 Qinetiq Limited Navigation system
SE530384C2 (sv) 2006-02-17 2008-05-20 Totalfoersvarets Forskningsins Metod för fjärrstyrning av en obemannad markfarkost med rörlig kamera samt en sådan markfarkost
KR20090069535A (ko) 2007-12-26 2009-07-01 한국해양연구원 수중-수상 연계 통신 플랫폼
US7926598B2 (en) 2008-12-09 2011-04-19 Irobot Corporation Mobile robotic vehicle
NO20091637L (no) 2009-04-24 2010-10-25 Sperre As Undervannsfartoy med forbedrede fremdrifts- og handteringsmuligheter
KR101128032B1 (ko) 2009-11-12 2012-03-29 한국해양대학교 산학협력단 다자유도 무인 수상 로봇 기반의 수중 작업 로봇
US9522595B2 (en) 2011-01-27 2016-12-20 Irobot Defense Holdings, Inc. Small unmanned ground vehicle
EP2511659A1 (de) 2011-04-14 2012-10-17 Hexagon Technology Center GmbH Geodätisches Markierungssystem zur Markierung von Zielpunkten
EP2511656A1 (de) 2011-04-14 2012-10-17 Hexagon Technology Center GmbH Vermessungssystem zur Bestimmung von 3D-Koordinaten einer Objektoberfläche
KR101327975B1 (ko) 2012-05-17 2013-11-13 한국해양과학기술원 해저 로봇의 기능 시험용 테스트 베드
CN102819263B (zh) 2012-07-30 2014-11-05 中国航天科工集团第三研究院第八三五七研究所 多摄像头无人车视觉感知系统
US9185392B2 (en) * 2012-11-12 2015-11-10 Spatial Integrated Systems, Inc. System and method for 3-D object rendering of a moving object using structured light patterns and moving window imagery
KR101431383B1 (ko) 2013-06-04 2014-08-18 서강대학교산학협력단 다관절 구동장치 및 이를 구비한 다족 주행로봇

Cited By (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9772185B2 (en) * 2011-04-14 2017-09-26 Hexagon Technology Center Gmbh Measuring system and method for determining new points
US20140210663A1 (en) * 2011-04-14 2014-07-31 Hexagon Technology Center Gmbh Measuring system and method for determining new points
US10497132B2 (en) * 2015-07-17 2019-12-03 Nec Corporation Irradiation system, irradiation method, and program storage medium
US20180197300A1 (en) * 2015-07-17 2018-07-12 Nec Corporation Irradiation system, irradiation method, and program storage medium
US10846866B2 (en) 2015-07-17 2020-11-24 Nec Corporation Irradiation system, irradiation method, and program storage medium
US20190019051A1 (en) * 2016-03-23 2019-01-17 JVC Kenwood Corporation Unmanned mobile apparatus capable of transferring imaging, method of transferring
US11513029B2 (en) * 2016-04-25 2022-11-29 Siemens Aktiengesellschaft Moving flying object for scanning an object, and system for analyzing damage to the object
US10152891B2 (en) * 2016-05-02 2018-12-11 Cnh Industrial America Llc System for avoiding collisions between autonomous vehicles conducting agricultural operations
US11828859B2 (en) * 2016-05-07 2023-11-28 Canyon Navigation, LLC Navigation using self-describing fiducials
US10417469B2 (en) * 2016-05-07 2019-09-17 Morgan E. Davidson Navigation using self-describing fiducials
US20210311205A1 (en) * 2016-05-07 2021-10-07 Canyon Navigation, LLC Navigation Using Self-Describing Fiducials
US11036946B2 (en) * 2016-05-07 2021-06-15 Canyon Navigation, LLC Navigation using self-describing fiducials
US11294456B2 (en) * 2017-04-20 2022-04-05 Robert C. Brooks Perspective or gaze based visual identification and location system
US11327149B2 (en) * 2017-07-10 2022-05-10 Aurora Flight Sciences Corporation Laser speckle system and method for an aircraft
EP3435030A1 (de) * 2017-07-27 2019-01-30 Testo SE & Co. KGaA Verfahren zur erstellung eines 3d-modells von einem objekt
US11053021B2 (en) * 2017-10-27 2021-07-06 Drone Delivery Canada Corp. Unmanned aerial vehicle and method for indicating a landing zone
US20190127083A1 (en) * 2017-10-27 2019-05-02 Drone Delivery Canada Corp. Unmanned aerial vehicle and method for indicating a landing zone
US10382668B2 (en) * 2017-11-14 2019-08-13 Conary Enterprise Co., Ltd. Laser collimator module on a mobile device for image measurement
US10741086B2 (en) * 2018-03-26 2020-08-11 D2, Llc Method and system for generating aerial imaging flight path
US20190295423A1 (en) * 2018-03-26 2019-09-26 D2, Llc Method and system for generating aerial imaging flight path
US20230121363A1 (en) * 2018-03-26 2023-04-20 D2, Llc Method and system for generating aerial imaging flight path
US11346665B2 (en) 2018-11-21 2022-05-31 Guangzhou Xaircraft Technology Co., Ltd Method and apparatus for planning sample points for surveying and mapping, control terminal, and storage medium
CN109658450A (zh) * 2018-12-17 2019-04-19 武汉天乾科技有限责任公司 一种基于无人机的快速正射影像生成方法

Also Published As

Publication number Publication date
EP3168704B1 (de) 2021-02-24
EP3168704A1 (de) 2017-05-17
CN107024199A (zh) 2017-08-08

Similar Documents

Publication Publication Date Title
EP3168704B1 (de) 3d überwachung einer fläche durch mobile fahrzeuge
US9758239B2 (en) System and method for controlling an unmanned air vehicle
JP6843773B2 (ja) 環境の走査及び無人航空機の追跡
CA2832956C (en) System and method for controlling an unmanned aerial vehicle
AU2019217205B2 (en) Method of and apparatus for analyzing images
AU2015340110B2 (en) Underwater positioning system
AU2012376428B2 (en) Map data creation device, autonomous movement system and autonomous movement control device
US9482524B2 (en) Measuring system for determining 3D coordinates of an object surface
CA2328227C (en) Method of tracking and sensing position of objects
US9142063B2 (en) Positioning system utilizing enhanced perception-based localization
US20080262718A1 (en) Landmark Navigation for Vehicles Using Blinking Optical Beacons
US10527423B1 (en) Fusion of vision and depth sensors for navigation in complex environments
CN114585875A (zh) 计量系统
US8582195B2 (en) Systems and methods for relative positioning
US20230064071A1 (en) System for 3d surveying by an autonomous robotic vehicle using lidar-slam and an estimated point distribution map for path planning
CN110945510A (zh) 借助于测量车辆进行空间测量的方法
CN115718298A (zh) Ugv和uav自动提供其激光雷达数据参照进行3d探测的系统
JP7401192B2 (ja) 車両位置提示システム、これに利用する車載器、車両位置提示方法、および車両位置提示用プログラム
JP2021182177A (ja) 車両操縦システムと車両操縦方法
CN112146627A (zh) 在无特征表面上使用投影图案的飞行器成像系统
US11762398B1 (en) Multimodal beacon based precision landing system for autonomous aircraft
Harris et al. Assessment of a visually guided autonomous exploration robot
Krátký et al. Documentation of Historical Buildings with the Usage of Model Predictive Control

Legal Events

Date Code Title Description
AS Assignment

Owner name: HEXAGON TECHNOLOGY CENTER GMBH, SWITZERLAND

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:PETTERSSON, BO;SIERCKS, KNUT;SIGNING DATES FROM 20160126 TO 20160311;REEL/FRAME:040291/0346

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION