EP2602588A1 - Positions- und Ausrichtungsbestimmung in 6-DOF - Google Patents
Positions- und Ausrichtungsbestimmung in 6-DOF Download PDFInfo
- Publication number
- EP2602588A1 EP2602588A1 EP11192220.9A EP11192220A EP2602588A1 EP 2602588 A1 EP2602588 A1 EP 2602588A1 EP 11192220 A EP11192220 A EP 11192220A EP 2602588 A1 EP2602588 A1 EP 2602588A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- known shape
- orientation
- range
- camera
- information
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C15/00—Surveying instruments or accessories not provided for in groups G01C1/00 - G01C13/00
- G01C15/002—Active optical surveying means
-
- E—FIXED CONSTRUCTIONS
- E02—HYDRAULIC ENGINEERING; FOUNDATIONS; SOIL SHIFTING
- E02F—DREDGING; SOIL-SHIFTING
- E02F3/00—Dredgers; Soil-shifting machines
- E02F3/04—Dredgers; Soil-shifting machines mechanically-driven
- E02F3/76—Graders, bulldozers, or the like with scraper plates or ploughshare-like elements; Levelling scarifying devices
- E02F3/80—Component parts
- E02F3/84—Drives or control devices therefor, e.g. hydraulic drive systems
- E02F3/841—Devices for controlling and guiding the whole machine, e.g. by feeler elements and reference lines placed exteriorly of the machine
- E02F3/842—Devices for controlling and guiding the whole machine, e.g. by feeler elements and reference lines placed exteriorly of the machine using electromagnetic, optical or photoelectric beams, e.g. laser beams
-
- E—FIXED CONSTRUCTIONS
- E02—HYDRAULIC ENGINEERING; FOUNDATIONS; SOIL SHIFTING
- E02F—DREDGING; SOIL-SHIFTING
- E02F9/00—Component parts of dredgers or soil-shifting machines, not restricted to one of the kinds covered by groups E02F3/00 - E02F7/00
- E02F9/26—Indicating devices
- E02F9/261—Surveying the work-site to be treated
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C15/00—Surveying instruments or accessories not provided for in groups G01C1/00 - G01C13/00
- G01C15/02—Means for marking measuring points
- G01C15/06—Surveyors' staffs; Movable markers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
- G06T7/74—Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
Definitions
- the present invention relates generally to a method for a six degree of freedom position and orientation determination of a three dimensional known shape according to the preamble of claim 1, and a six degrees of freedom spatial position and orientation determination device according to the preamble of claim 15.
- a determination of the position and orientation of a target object in 3D space is required.
- the spatial posture of a target object in space by measuring its spatial coordinates and its spatial orientation has to be determined, resulting in up to six degrees of freedom which needs to be evaluated.
- the target object can be the object of interest itself or a dedicated target object which is attached thereto.
- Controlling of construction machinery is another application, e.g. as described in US 5,771,978 , wherein a Tracking-Station is tracking a retro reflector as a target mark which is attached to a working tool of the machine, in this example a blade of a dozer for moving earth into a desired shape.
- the document EP 1 171 752 (- honoured with the EPO inventors award in 2010 -) describes another application, wherein the position and orientation of a measurement probe for coordinate measurement is determined in six degrees of freedom by a tracker-device, which is tracking multiple dedicated discrete singular reference points that are attached to the probe for determining the posture of the measuring probe-tip in six degrees of freedom.
- Pattern projection systems like the DAVID laser scanner project from TU-Braunschweig are also known.
- a laser pattern e.g. a line
- a digital video camera records the resulting images, whereof a 3D-point-cloud-model of the scanned object is created.
- a photographic image of the camera can also be used to texture the resulting 3D-Model. Due to the scanning, a digitalisation of the full field of view of the camera takes time and is therefore improper for evaluating or observing non-steady objects.
- Range Imaging Modules For a 3D measurement by gathering 3D point cloud, also the usage of Range Imaging Modules (RIM) is a known technique, which can digitize a whole scenery in "one shot".
- a RIM-camera comprises in principle an array of pixels functioning as optoelectronic distance meters, e.g. based on a time of flight measurement (TOF) of optical radiation, e.g. by pulse, phase, signal-shape or interferometrical principles which are known in the art, e.g. from " Electronic Distance Measurement” from J.M. Rüeger, Ed. 4, Springer-Verlag, Berlin, Heidelberg 1996 .
- TOF time of flight measurement
- a range image taken by such a RIM-camera comprises distance information for each of the pixels of the camera, resulting in a 3D image of the scenery taken from a single point of view.
- the measurement is thereby done in polar coordinates according to the angular field of view of a camera-pixel and the therewith determined distance in a unit of length.
- a graphical view of the measured RIM-image can e.g. be presented in a two dimensional representation by luminescence- or colour keying of the distances into the 2D picture of the view of the RIM-image. Alternatively, it can be presented in an axonometric view or by a real 3D display.
- an intensity of the scattered back distance measurement radiation can be determined by the RIM camera an provided in an image.
- drawbacks like again prolonging the measurement time which is an important parameter, in particular in the case of measuring a potentially moving target.
- WO 2006/075017 presents a method and geodetic apparatus for surveying at least one target.
- a range imaging module comprising sensors in a matrix arrangement, e.g. 32 x 32 sensors are used for providing a range image.
- the range image provides a so called cluster of points or point cloud information comprising the range of the target points imaged by the respective pixels of the sensor.
- range images of details subsequently reduced in size can be taken.
- this may improve the accuracy of the range image in a certain manner, due to the relative low resolution of the range image, it is still difficult to exactly address distinctive target points of the target which can be extremely important in case of smaller and/or moving targets which are changing their orientation and position while being surveyed.
- Another drawback is the prolonged measurement time for measuring the full scenario with the respective high resolution and accuracy.
- RIM-Imaging often is the low image resolution of the available RIM-cameras in particular compared to the state of the art digital photographic cameras having resolutions of up to tens of Mega-Pixels and more.
- US 2010/0046802 describes an approach which is using a combination of a RIM-camera and a picture camera for enhancing 3D resolution for an enhanced depth feel for a movie or still camera by the presented distance estimation apparatus.
- the document comprises different aspects and embodiments of such an approach and can serve as a reference for some of the underlying principles of the present invention.
- edge-extraction techniques and other aspects of matching a range image and a corresponding visual picture are elaborated therein.
- An object of the present invention is to provide an improved method and apparatus to determine position and orientation of an object in a viewed scenery, in particular in 3D space in six degrees of freedom.
- Another object of the invention is to provide a method for determining the position and orientation in six degrees of freedom of an object of known shape inside of an evaluated scenery with high accuracy.
- Yet another object is a 6-DOF measurement with a reduced measurement time, in particular to allow measurements of moving objects.
- a particular object of the invention is to provide a method of 3D measurement which can in particular be used to precisely measure a distinct point of the measurement object, preferably by tactile means, and thereby enhance a recorded point cloud in positional measurement accuracy.
- Another object of the invention is also to provide a measurement method which is capable of measuring parts of the measurement object which are shaded from the measurement devices point of view.
- the present invention relates to a method for a six degree of freedom position and orientation determination of a three dimensional known shape in a scenery.
- the method comprises a taking of a range image by means of a range imaging camera.
- the range imaging camera comprises a range image module (RIM) having a sensor array with a first number of pixels, wherein for each of the first pixels a range information from the sensor to a point of the scenery is determined, resulting in a 3D cluster of points.
- RIM range image module
- the RIM can be embodied by a light emitting means, emitting modulated optical radiation in at least one visible or preferably invisible spectral range, directed to a measurement target and an array of electro-optical sensor means for converting the part of the optical radiation scattered back from the measurement target into electrical signals.
- the distance can be determined, in particular according the delay of a light pulse from transmit to receive or according to phase differences of a modulated light signal or bursts of modulated light signals.
- a known reference distance can serve as calibration means.
- Ambiguities can for example be resolved by the usage of different modulation frequencies.
- the receiving sensor elements of the array are built in such a way to cover each an angular field of view in direction to the target object, preferably in a non overlapping manner, so the field of view of the RIM is resolved in pixels with a resolution equal to the rows and columns of the sensor array.
- the method further comprises a taking of a visual picture by means of a digital camera (CAM).
- the digital camera comprises an image sensor (e.g. a CCD or CMOS array or the like) having a second number of pixels, wherein the second number of pixels of the digital camera can be higher than the first number of pixels, resulting in a 2D picture with an second angular resolution of the cameras field of view which can been higher than the one of the RIM.
- the picture is a digital photographic image, in particular comprising intensity and spectral information for an array of pixels, comprising angular information of each pixel according to the pixels field of view together with luminosity and chromatic information.
- optical patterns or textures of the target in the spectral range of the digital camera can be observed, as well as shading effects, different coloured sections of the target object.
- the field of view of the digital camera and also of the RIM can be shaped by some optics, e.g. comprising lenses, filters, mirrors, etc. as known in the art of optics. Possible optical or geometrical distortions of the pixel arrays field of view can be compensated by the either optical means or numerical means in the evaluation software.
- the relation of the first and the second pixels fields of view is known, or made known by a calibration with a reference object observed by both the RIM and the photographic camera.
- the scenery can be defined a the area of interest, seen from a point of view of the measurement instrument comprising RIM and CAM, wherein at least a part of the target object resides and which can be covered by the RIMs and CAMs field of view.
- the RIMs and/or CAMs field of view can be moveable, wherein the amount of movement has to be determinable, either by a dedicated measurement means or by extracting the movement information from the RIM- and/or CAM-data.
- the mentioned three dimensional known shape is known in such a way that a 3D digital representation of the known shape is available in a computation system involved in measurement.
- the digital representation can be provided as geometric CAD-model represented by numerical information of the shapes shape and size. Alternatively, this information can be gathered by so called calibration measurement of the shape for "learning" its shape and making it known.
- the digital representation of the object is aligned, e.g. moved and rotated in virtual space to result in the same posture as the known shape has in real space, which is defined by a match of the virtual view of the shape in the calculated digital numerical representation in virtual space and the view as seen from the CAM and/or RIM which is reproduced in form of the 2D/3D information taken by them.
- the six degree of freedom position and orientation of the known shape in the scenery is determined, by the information taken from the virtual alignment of the geometrically manipulated digital representation of the known shape in virtual space, which then corresponds to the real position and orientation of the known shape in the real scenery observed by the method.
- Fig. 1 shows an embodiment of a six degrees of freedom spatial position and orientation determination device 1 according to the present invention, comprising a RIM-unit 4, comprising a range imaging camera for determining a range image as a three dimensional point cloud information with a first resolution, a CAM-unit 5, comprising a visual picture camera for determining a two dimensional visual picture of the scenery with a second resolution, in particular in multiple colours.
- the second resolution is greater than the first resolution, which can also be the other way round or the two resolutions can also be equal.
- Another option is a skipping of pixels in the readout of the CAM- and/or RIM-sensor, wherein the physical sensor-resolution might be different from the actually evaluated resolution, for example for the sake of increased evaluation speed.
- the device 1 also comprises a digital image processing unit 66, in particular built in such a way to determine edge and face information in the visual picture and a six dimensional known shape matching unit 67 built in such a way to determine a three dimensional best fit of a 3D digital representation of a known shape within the range image and the visual picture and extracting position and orientation of the known shape.
- a digital image processing unit 66 in particular built in such a way to determine edge and face information in the visual picture
- a six dimensional known shape matching unit 67 built in such a way to determine a three dimensional best fit of a 3D digital representation of a known shape within the range image and the visual picture and extracting position and orientation of the known shape.
- the device 1 shows a system according to the present invention comprising the device 1 and a reference object as the known shape 2.
- the reference object is formed in it's outer shape in such a way, to achieve an unambiguous identification of its position and orientation in space from a single point of view.
- the reference object being reflective for the measurement-wavelength of the range imaging camera 4 and having characteristic identifiable optical features in the wavelength(s) captured by the visual camera 5.
- the visual picture 14 and the range image 13 can be taken from a single measurement apparatus 1, in particular with a single line of sight 9, preferably wherein the taking of the range image and the visual picture is synchronized in time.
- the illustration shows the light source for RIM measurement 3 and the RIM-Sensor 4 which are guided through the same objective lens 8 of the device 1 as the photographic image pickup means 5 is using.
- the wavelengths of the range imaging and the visual picturing are in this example separated by the dichromatic mirror 7.
- the picture comprises spectral or colour information 10,11,12 for each pixel and the range image comprises range information 13 of each pixel which can be combined with intensity information 14 of the scattered back radiation.
- This evaluation of intensity information determined from the reflected intensity of the range imaging light received by the range camera, in particular as monochrome intensity values of the distance measurement light scattered back by the known shape, can be included in the virtual matching and fitting.
- Fig. 4 illustrates the gathered data according to the invention.
- the known shape 2 is represented by it's digital representation 11 - shown as a CAD-model.
- the photographical CAM-unit takes a picture 14 as shown, which preferably is a 2D colour image.
- the RIM-unit 3,4 determines a point cloud representation 13 of the object as illustrated by the dots for which a distance has been evaluated.
- edges 15 and or faces 16 can be extracted. Due to the high resolution of the visual picture from the CAM-unit in this embodiement, the edges and faces can be determined with a spatial accuracy, higher than the one of the RIM-unit.
- Those extracted data can be matched to the digital representation 11, by manipulations in virtual 3D space to achieve a best fit. Multiple information can thereby be processed and combined by a maximum likelihood algorithm or the like.
- the position and orientation of the best fit in virtual space reveals the position and orientation of the known shape 2 in the scenery.
- Fig. 5 illustrates a schematic sketch of the functional principle of a method for spatial location and posture or six degree of freedom position and orientation determination according to the present invention.
- a three dimensional known shape 2 as volumetric object of known geometrical shape, which is placed in a scenery, is observed by a range imaging camera 4.
- the range camera 4 is taking a range image of the scenery by means of a range image module RIM, having a sensor array with a first number of pixels. For each of the first pixels, a range information from the sensor to a point of the scenery is determined, resulting in a cluster of three dimensional point information 13, representing the 3D surface information seen from range camera by the RIMs field of view.
- the object of known geometrical shape 2 is further observed by taking a visual photographic picture with a digital camera 5.
- the camera comprises an image sensor with a second number of pixels, which can record a 2D picture 14, in particular a color picture comprising intensity and spectral information for each of the pixels, according to the light received within the image sensors field of view.
- the second number of pixels of the 2D picture camera 5 can therein be higher than the first number of pixels of the RIM-camera, whereby the angular resolution of the field of view of the scenery and the therein comprised known shape 2 in the 2D representation 14 is higher than the one of the 3D point cloud 13.
- the relation of the first and the second pixels fields of view is known and the angular fields of view of the pixels of the RIM and CAM pixels can be assigned to each other.
- the know shape 2 is known is such a way that a 3D digital representation 11 of the known shape 2 is available or can be generated, e.g. in form of a stored CAD-model.
- This digital representation 11 can be moved and rotated in a virtual space in a computing system and virtual views of the digital representation of the known shape can be numerically generated as two and/or three dimensional data.
- the 3D digital representation 11 of the known shape 2 is geometrically manipulated in virtual space by translation and rotation in six degrees of freedom in such a way, to match 12 its virtual view with the reproduction of the known object in the 2D picture and the 3D cluster of points.
- determining the six degree of freedom position and orientation of the known shape in the scenery is achieved and the known shape 2 can thereby be survived in six degrees of freedom an the known shape's position and orientation in the scenery can be measured.
- the RIM-camera 4 can also be a colour RIM, with a sensor working in for example the RGB-colour space - similar to a 2D digital imaging device- and an RGB-Illumination which be embodied by an emission of red green and blur light pulses, which can be done subsequent or in parallel.
- the RIM-resolution would be equal to the CAM-resolution, or - e.g. when interpreting a group of three colour CAM-pixels as a single image pixel coloured with a mixture of the three-colours - the 3D-RIM resolution could also be higher than the one of the colour picture. This results in three range images - one for each colour. The summed to colourize achieve a colourized 3D Image.
- the ranges of the three RIM-images can be combined (determining a mean value if all three images have the same field of view or geometrically interpolated, if the fields of view of the pixels of the three images are shifted in a sub pixel range for enhancing 3D resolution.
- Beside RGB- also other colour spaces are known, e.g. also with more than three colours.
- Another option can be to add a "fourth colour", which is dedicated to the RIM-measurement, e.g. a single R+G+B+Infrared sensor, whereby a single piece of hardware embodies both of an infrared-RIM- and colour-CAM-unit, e.g. with equal resolution.
- 3D-RIM-evaluation can (e.g. due to higher computational effort) achieve only lower frame-rates than CAM-picture-evaluation, some RIM-pixels can be skipped during evaluation. In particular when leaving out pixels, the order of the evaluated/skipped RIM-pixels can be alternated in evaluation.
- An extracting of a 3D geometrical information by combining information from the visual picture 14 and information from the range image 13 can be accomplished, in particular wherein the combining comprises a matching of at least one geometrical feature of the known shape 2 in both of the range image 13 and the visual picture 14 and a further matching of the geometrical feature in the 3D digital representation 11 within the combined information.
- the matching 12 can comprise identifying a face by combining edge/face extraction 15,16 of the face in the visual picture 14 and/or edge/face detection 15,16 in the range image 13, preferably by fitting planes in the cluster of points.
- a matching of the identified face and/or edge information with a corresponding face or edge of the 3D digital representation 11, in particular according to a least square or maximum likelihood algorithm, can be done in the matching process for fitting the image and RIM information 14,13.
- multiple faces and/or edges are three dimensionally matched and the results are combined and interpolated to increase measurement accuracy.
- a recognition of textural structures within the visual picture, in particular on the known shape can be included in the determination.
- the method can further comprise an additional position and orientation determination according to the appearing size of the known shape 2 and/or its texture in the visual picture 14.
- the known textures and or colours of the known object can be used for deriving uniqueness in the position and orientation determination.
- an inclusion of the textural information can be used.
- a directional dithering of an optical axis of the range camera relative to an optical axis of the digital camera in a sub-resolution of the first pixels can be used to enhance resolution.
- This dithering of the optical axis, e.g. of the RIM 4 compared to the CAM 5 can in a simple way be introduced by means of a reversely used (or so to say misused) optical image stabilization unit, such as known from digital photographic and video cameras, (like SONY "steady-shot"-technology, or similar approaches from other vendors).
- a reversely used optical image stabilization unit such as known from digital photographic and video cameras, (like SONY "steady-shot"-technology, or similar approaches from other vendors).
- the RIM-unit 4 and CAM-unit 5 can be movable, in particular motorized movable, relative to it's mounting such as the shown tripod 6. Thereby the field of view of the device can be moved, e.g. to keep the known shape in the field of view or to alternatively achieve the above mentioned dithering.
- Fig. 6 shows a simplified flowchart of the method according the present invention.
- a RIM-image 200 and a CAM-picture 201 are taken and a visual picture 203 - comprising an angular array of intensity and optionally spectral data - and a point cloud 202 - comprising an angular array of range and optionally intensity data - are captured.
- the takings are preferably synchronized in time.
- a reproduction of the known shape is generated of the captured information and is matched in virtual space in step 208 to the digital representation 209 of the known shape 2, e.g. in form of stored CAD-data 210.
- the virtual match is then used to determine the 6-DOF position and orientation of the known shape in the scenery captured by the device 1, as shown in step 211.
- the above metioned steps can be described as a three dimensional identification and fitting of a digital representation of the known shape in the geometrical information gathered by a RIM- and CAM-unit and processed by a digital image processing unit, in particular built in such a way to determine edge and face information in the visual picture which can involve extracting of vectorial information of the scenery, e.g. comprising an extracting of vector-information from the visual picture by edge detection and a combining the edge information with the point cloud information from the RIM image and/or a vectorial face detection.
- a consolidation module can be used for combining information from the RIM-unit and the CAM-unit for determining an accurate three dimensional geometrical information of the scenery, preferably comprising a match of the vectorial information and the point cloud information with the known digital representation.
- geometrical object of higher order can be better, in particular faster, matched than points, an increased measurement rate and/or matching accuracy can be achieved.
- a matching and fitting can be explained in a simplified manner as utilizations of supporting points from the range image in combination with the 2D visual image and a determination of a plane on a face of the reference object by at least three supporting points from the range image.
- the plane can then be fine adjusted according to edges of a face detected in the visual picture and/or textural structures on the face of the reference object.
- a deriving of range information of the reference object dependant on the appearing size of the object in the visual picture as additional information can be used to gain additional information.
- a camera calibration as e.g. presented in the above mentioned Microsoft document can be used therewith.
- an inclusion of greyscale values of shaded or semi-shaded surfaces of the target object can be included in the evaluation for improving a 3D reconstruction.
- the digital representation of the known shape can also be established by a teaching-in of the reference object by means of determining the known shape by the apparatus itself.
- the teaching-in of the reference object can also be done by defining the known shape by a numerical input of geometrical data or a numerical parameterisation of a virtual reference object.
- the method according to the present invention can be used for determining a dependent position and orientation of an item attached to the known shape 2.
- the known shape - as a known three dimensional solid volumetric body having a three dimensional outer configuration facilitating a precisely and uniquely determinable spatial position and orientation of the known shape 2 in six degrees of freedom - can be used for determining 6-DOF spatial information of the item it is attached to.
- the item or part of it can also be the known shape 2 itself.
- the attached item is a measurement probe, in particular a tactile coordinate-measurement probe 106, and the method is measuring spatial coordinates of a measurement object approaches by the measurement probe.
- the shown arm can also be actuated robot arm, which's posture in space, in particular of the tip of the arm is determined according to the method of the present invention.
- the robot arms posture in space is determined according to the position and orientation of the known shape.
- the upper right illustration shows a worker carrying a handheld measurement probe 102 which is, comprises or is attached to an object of known shape 2.
- the measurement object illustrated by a car body 101 can be measured.
- this reveals an additional method of three dimensional modelling by digitalizing an outer form of a measurement object.
- the rough modelling comprises a taking of a range image of the measurement object 101 and optionally a taking of a visual picture of the measurement object 101 and generating a rough 3D model of the gathered data, in particular by combining information of the range and the visual image.
- the additional method also involves a fine 3D modelling.
- the fine modelling comprises a measuring of at least one surface point of the measurement object with a tactile probe 102 by the 6-DOF determination of a known shape 2 by combined RIM and CAM information as described above. Thereby the exact spatial position and orientation of the surface point measured by the tactile probe is determined according to the position and orientation of the known shape with increased accuracy compared to the rough modelling.
- the additional method refines measurement points of the rough model according to the at least one surface point, by a fine adjusting at least one point of the rough 3D model according to the at least one surface point of the fine modelling.
- Fig. 7 shows a working tool 105 of a construction site machinery 104 as the item attached to the known shape 2, which's spatial coordinates are determined.
- the worksite machinery shown is an earth moving machine embodied as an excavator 104, but it can also be a dozer, grader, digger, etc.
- the device 1 according to the invention is shown which is used for the 6-DOF determination of the known shape 2 within its field of view.
- Fig. 8 shows a special usage of the method according to the present invention to achieve a so called look behind measurement.
- This special method of generating a three dimensional model of a measurement object comprises the steps of:
- This arrangement determines a secondary field of view of the range and visual camera, which is a subset of the primary field of view.
- the secondary field of view will be redirected by the mirror according to the position and orientation of the first known shape 2A, Thereby a measuring in the secondary field of view by the range-imaging means 4 and/or visual camera 5 can be executed, in particular by a position and orientation determination of a second known shape in the secondary field of view (e.g. behind the target object 2B wherefore the second known shape is not shown).
- a transforming of the measurements taken in the secondary field of view into a coordinate system of the first field of view is the calculated according to the determined position and orientation of the mirror 73, which is determined by its frame 2A. Then a measurement - or generating of the three dimensional model according to measurements - from the primary and secondary field of view is combined.
- the measurement apparatus 1 and the measurement object 2B are steady with respect to each other and the mirror 73 is movable.
- the mirror can be movable by a motor or by hand (as a handheld item) or it can be fixed to a tripod 75 if rearrangement of the mirror is not or not often required.
- a lower resolution RIM-camera can be zoomed (and optionally also positional aligned - or tracked) to the mirror's field of view for achieving a higher 3D resolution, while the CAM-unit at higher resolution is evaluating a broader scenery comprising the mirror and determining it's position and orientation according to its (or specifically its frame's) known shape.
- the view of the RIM-camera can preferably also cover the mirror's frame or parts of it, for improved position and orientation determination of the mirror according to the invention.
- a digitalisation or modelling of the measurement object's shape from side, behind or inside can be achieved this way.
- the methods according to the invention can be at least partly embodied as a computer program product with program code being stored on a machine readable medium or embodied as an electromagnetic wave, the program code being configured to automatically execute and operate the method for a six degree of freedom position and orientation determination of a known shape as described above, in particular the program is carried out on the digital image processing unit 66 as a digital computation unit of a six degree of freedom spatial position and orientation determination device 1.
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP11192220.9A EP2602588A1 (de) | 2011-12-06 | 2011-12-06 | Positions- und Ausrichtungsbestimmung in 6-DOF |
CN201280059710.0A CN103959012B (zh) | 2011-12-06 | 2012-12-05 | 6自由度位置和取向确定 |
PCT/EP2012/074552 WO2013083650A1 (en) | 2011-12-06 | 2012-12-05 | Position and orientation determination in 6-dof |
US14/363,102 US9443308B2 (en) | 2011-12-06 | 2012-12-05 | Position and orientation determination in 6-DOF |
EP12797905.2A EP2788717B8 (de) | 2011-12-06 | 2012-12-05 | Positions- und ausrichtungsbestimmung in 6-dof |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP11192220.9A EP2602588A1 (de) | 2011-12-06 | 2011-12-06 | Positions- und Ausrichtungsbestimmung in 6-DOF |
Publications (1)
Publication Number | Publication Date |
---|---|
EP2602588A1 true EP2602588A1 (de) | 2013-06-12 |
Family
ID=47297276
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP11192220.9A Withdrawn EP2602588A1 (de) | 2011-12-06 | 2011-12-06 | Positions- und Ausrichtungsbestimmung in 6-DOF |
EP12797905.2A Active EP2788717B8 (de) | 2011-12-06 | 2012-12-05 | Positions- und ausrichtungsbestimmung in 6-dof |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP12797905.2A Active EP2788717B8 (de) | 2011-12-06 | 2012-12-05 | Positions- und ausrichtungsbestimmung in 6-dof |
Country Status (4)
Country | Link |
---|---|
US (1) | US9443308B2 (de) |
EP (2) | EP2602588A1 (de) |
CN (1) | CN103959012B (de) |
WO (1) | WO2013083650A1 (de) |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2518673A (en) * | 2013-09-30 | 2015-04-01 | Ortery Technologies Inc | A method using 3D geometry data for virtual reality presentation and control in 3D space |
EP3457080A1 (de) * | 2017-09-13 | 2019-03-20 | Topcon Corporation | Vermessungsinstrument |
IT201700112107A1 (it) * | 2017-10-06 | 2019-04-06 | A D R I S R L S | Kit di rilievo architettonico certificato di un immobile o di un terreno e metodo associato |
CN109934869A (zh) * | 2017-12-19 | 2019-06-25 | 欧姆龙株式会社 | 位置姿势推断装置、位置姿势推断方法及记录介质 |
EP3392612A4 (de) * | 2015-12-14 | 2019-07-10 | Nikon-Trimble Co., Ltd. | Fehlererkennungsvorrichtung und programm |
WO2020049272A1 (en) * | 2018-09-03 | 2020-03-12 | University Of Northumbria At Newcastle | Object tracking |
CN111144478A (zh) * | 2019-12-25 | 2020-05-12 | 电子科技大学 | 一种穿帮镜头的自动检测方法 |
CN114378825A (zh) * | 2022-01-21 | 2022-04-22 | 四川长虹智能制造技术有限公司 | 一种多相机视觉定位方法、系统及电子设备 |
Families Citing this family (49)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9113154B2 (en) | 2013-07-10 | 2015-08-18 | Faro Technologies, Inc. | Three-dimensional measurement device having three-dimensional overview camera |
US9229674B2 (en) | 2014-01-31 | 2016-01-05 | Ebay Inc. | 3D printing: marketplace with federated access to printers |
EP2930462B1 (de) * | 2014-04-08 | 2017-09-13 | Hexagon Technology Center GmbH | Verfahren zur Erzeugung von Informationen über eine Sensorkette einer Koordinatenmessmaschine (CMM) |
US9507995B2 (en) * | 2014-08-29 | 2016-11-29 | X Development Llc | Combination of stereo and structured-light processing |
EP3190380B1 (de) * | 2014-09-03 | 2022-03-23 | Nikon Corporation | Bildaufnahmevorrichtung, informationsverarbeitungsverfahren und bildaufnahmesystem |
JP6426968B2 (ja) * | 2014-10-08 | 2018-11-21 | キヤノン株式会社 | 情報処理装置およびその方法 |
US9595037B2 (en) | 2014-12-16 | 2017-03-14 | Ebay Inc. | Digital rights and integrity management in three-dimensional (3D) printing |
US20160167307A1 (en) * | 2014-12-16 | 2016-06-16 | Ebay Inc. | Systems and methods for 3d digital printing |
CN104539928B (zh) * | 2015-01-05 | 2016-08-24 | 武汉大学 | 一种光栅立体印刷图像合成方法 |
CN104833342B (zh) * | 2015-03-09 | 2017-04-19 | 深圳市魔眼科技有限公司 | 一种多点测距建立立体模型的移动终端及方法 |
DE102015209896B3 (de) * | 2015-05-29 | 2016-08-18 | Kuka Roboter Gmbh | Ermittlung der Roboterachswinkel und Auswahl eines Roboters mit Hilfe einer Kamera |
DE102015109612B4 (de) * | 2015-06-16 | 2020-08-13 | Carl Zeiss Industrielle Messtechnik Gmbh | Koordinatenmessgerät und Verfahren zum Kalibrieren desselben mit einer Time-of-Flight-Kamera |
CN106228167B (zh) * | 2015-07-30 | 2018-04-27 | 全维智造(北京)技术开发有限公司 | 特征采集、标记和识别方法 |
KR20170105627A (ko) | 2015-09-30 | 2017-09-19 | 가부시키가이샤 고마쓰 세이사쿠쇼 | 촬상 장치 |
US10352689B2 (en) * | 2016-01-28 | 2019-07-16 | Symbol Technologies, Llc | Methods and systems for high precision locationing with depth values |
EP3428877A4 (de) | 2016-03-09 | 2019-10-30 | Nikon Corporation | Detektionsvorrichtung, informationsverarbeitungsvorrichtung, detektionsverfahren, detektionsprogramm und detektionssystem |
EP3485112B1 (de) | 2016-07-15 | 2021-08-25 | Fastbrick IP Pty Ltd | Fahrzeug mit eingebauter ziegelsteinlegemaschine |
BR112019000722B1 (pt) | 2016-07-15 | 2023-03-28 | Fastbrick Ip Pty Ltd | Lança extensível telescópica para transportar item e lança dobrável |
CN106392773B (zh) * | 2016-09-12 | 2019-10-25 | 清华大学 | 一种五轴联动机床主轴头姿态角测量装置及测量方法 |
CN106482665B (zh) * | 2016-09-21 | 2018-05-08 | 大连理工大学 | 一种组合点群高精度三维信息视觉测量方法 |
US10486060B2 (en) | 2016-11-23 | 2019-11-26 | Microsoft Technology Licensing, Llc | Tracking core for providing input to peripherals in mixed reality environments |
EP3351899B1 (de) * | 2017-01-24 | 2020-06-17 | Leica Geosystems AG | Verfahren und vorrichtung zum "colorising" von dreidimensionalen punktwolken |
EP3364374A1 (de) * | 2017-02-20 | 2018-08-22 | My Virtual Reality Software AS | Verfahren zur visualisierung von dreidimensionalen daten |
US10379606B2 (en) | 2017-03-30 | 2019-08-13 | Microsoft Technology Licensing, Llc | Hologram anchor prioritization |
US10466953B2 (en) | 2017-03-30 | 2019-11-05 | Microsoft Technology Licensing, Llc | Sharing neighboring map data across devices |
CN107194962B (zh) * | 2017-04-01 | 2020-06-05 | 深圳市速腾聚创科技有限公司 | 点云与平面图像融合方法及装置 |
CN107170345B (zh) * | 2017-04-11 | 2019-07-19 | 广东工业大学 | 面向工业机器人的基于机器视觉和陀螺仪的示教方法及装置 |
US10977818B2 (en) * | 2017-05-19 | 2021-04-13 | Manor Financial, Inc. | Machine learning based model localization system |
WO2019006511A1 (en) | 2017-07-05 | 2019-01-10 | Fastbrick Ip Pty Ltd | REAL-TIME POSITION TRACKING AND ORIENTATION DEVICE |
CN111213098B (zh) | 2017-08-17 | 2024-03-15 | 快砖知识产权私人有限公司 | 用于交互系统的通信系统 |
EP3669138B1 (de) | 2017-08-17 | 2023-05-24 | Fastbrick IP Pty Ltd | Lasertracker mit verbesserter wankwinkelmessung |
AU2018348785A1 (en) | 2017-10-11 | 2020-05-07 | Fastbrick Ip Pty Ltd | Machine for conveying objects and multi-bay carousel for use therewith |
TWI672207B (zh) * | 2017-11-03 | 2019-09-21 | 財團法人工業技術研究院 | 機械設備之定位系統及其方法 |
TW201944365A (zh) | 2018-04-16 | 2019-11-16 | 香港商電競方程式控股有限公司 | 增強第一身視角體驗的方法 |
US11182914B2 (en) | 2018-05-21 | 2021-11-23 | Facebook Technologies, Llc | Dynamic structured light for depth sensing systems based on contrast in a local area |
CN110555878B (zh) * | 2018-05-31 | 2021-04-13 | 上海微电子装备(集团)股份有限公司 | 物体空间位置形态的确定方法、装置、存储介质及机器人 |
US20200034985A1 (en) * | 2018-07-30 | 2020-01-30 | The Boeing Company | Method and system for measuring the orientation of one rigid object relative to another |
US11137247B2 (en) * | 2018-07-30 | 2021-10-05 | The Boeing Company | Method and system for measuring the orientation of one rigid object relative to another |
EP3864570A1 (de) * | 2018-10-10 | 2021-08-18 | DeLaval Holding AB | Identifizierung von tieren mittels sichtverfahren |
US11812171B2 (en) | 2018-10-31 | 2023-11-07 | Sony Semiconductor Solutions Corporation | Electronic device, method and computer program |
TWI680670B (zh) * | 2018-11-05 | 2019-12-21 | 國立臺北科技大學 | 擴增實境比對現實物體之方法及系統 |
US11630197B2 (en) * | 2019-01-04 | 2023-04-18 | Qualcomm Incorporated | Determining a motion state of a target object |
CN109816778B (zh) * | 2019-01-25 | 2023-04-07 | 北京百度网讯科技有限公司 | 物料堆三维重建方法、装置、电子设备和计算机可读介质 |
JP2021008747A (ja) * | 2019-07-01 | 2021-01-28 | 株式会社小松製作所 | 学習済みの作業分類推定モデルの製造方法、学習用データ、コンピュータによって実行される方法、および作業機械を含むシステム |
CN110531706A (zh) * | 2019-09-05 | 2019-12-03 | 肇庆学院 | 用于智能化车间安全围网与钣金外罩模块化3d建模系统 |
CN111470070B (zh) * | 2020-05-09 | 2022-03-22 | 清华大学 | 一种空间非合作目标无接触消旋辨识装置及方法 |
CN115037866A (zh) * | 2021-03-03 | 2022-09-09 | 中兴通讯股份有限公司 | 成像方法、装置、设备及存储介质 |
WO2023183659A1 (en) * | 2022-03-25 | 2023-09-28 | Atheraxon, Inc. | Method, platform, and system of electromagnetic marking of objects and environments for augmented reality |
CN117718985A (zh) * | 2024-02-07 | 2024-03-19 | 西安中科光电精密工程有限公司 | 一种基于智能三维视觉的搜排爆机器人 |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5771978A (en) | 1996-06-05 | 1998-06-30 | Kabushiki Kaisha Topcon | Grading implement elevation controller with tracking station and reference laser beam |
EP1171752A1 (de) | 1999-04-19 | 2002-01-16 | Leica Geosystems AG | Indirekte positionsbestimmung mit hilfe eines trackers |
US6608913B1 (en) * | 2000-07-17 | 2003-08-19 | Inco Limited | Self-contained mapping and positioning system utilizing point cloud data |
US20050002555A1 (en) * | 2003-05-12 | 2005-01-06 | Fanuc Ltd | Image processing apparatus |
WO2005039836A2 (de) | 2003-10-20 | 2005-05-06 | Isra Vision Systems Ag | Verfahren zur einrichtung einer bewegung eines handhabungsgeräts und bildverarbeitung |
WO2006075017A1 (de) | 2005-01-14 | 2006-07-20 | Leica Geosystems Ag | Verfahren und geodätisches gerät zur vermessung wenigstens eines zieles |
US20100046802A1 (en) | 2008-08-19 | 2010-02-25 | Tatsumi Watanabe | Distance estimation apparatus, distance estimation method, storage medium storing program, integrated circuit, and camera |
WO2011104199A1 (de) * | 2010-02-23 | 2011-09-01 | Ifm Electronic Gmbh | Überwachungssystem |
JP2011179980A (ja) * | 2010-03-01 | 2011-09-15 | Saxa Inc | 立体マーカを利用した位置計測システム |
CN102252653A (zh) * | 2011-06-27 | 2011-11-23 | 合肥工业大学 | 基于tof无扫描三维成像的位姿测量方法 |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5805275A (en) * | 1993-04-08 | 1998-09-08 | Kollmorgen Corporation | Scanning optical rangefinder |
US6173066B1 (en) * | 1996-05-21 | 2001-01-09 | Cybernet Systems Corporation | Pose determination and tracking by matching 3D objects to a 2D sensor |
US6166744A (en) * | 1997-11-26 | 2000-12-26 | Pathfinder Systems, Inc. | System for combining virtual images with real-world scenes |
US7003134B1 (en) * | 1999-03-08 | 2006-02-21 | Vulcan Patents Llc | Three dimensional object pose estimation which employs dense depth information |
GB2411532B (en) * | 2004-02-11 | 2010-04-28 | British Broadcasting Corp | Position determination |
US7860301B2 (en) * | 2005-02-11 | 2010-12-28 | Macdonald Dettwiler And Associates Inc. | 3D imaging system |
WO2008036354A1 (en) * | 2006-09-19 | 2008-03-27 | Braintech Canada, Inc. | System and method of determining object pose |
FR2911707B1 (fr) * | 2007-01-22 | 2009-07-10 | Total Immersion Sa | Procede et dispositifs de realite augmentee utilisant un suivi automatique, en temps reel, d'objets geometriques planaires textures, sans marqueur, dans un flux video. |
US7844106B2 (en) * | 2007-04-23 | 2010-11-30 | Mitsubishi Electric Research Laboratories, Inc | Method and system for determining poses of objects from range images using adaptive sampling of pose spaces |
US20080310757A1 (en) * | 2007-06-15 | 2008-12-18 | George Wolberg | System and related methods for automatically aligning 2D images of a scene to a 3D model of the scene |
DE102007048320A1 (de) * | 2007-10-09 | 2008-05-15 | Daimler Ag | Verfahren zur Anpassung eines Objektmodells an eine dreidimensionale Punktwolke |
IL196162A (en) * | 2008-12-24 | 2013-02-28 | Rafael Advanced Defense Sys | A system that uses 3D models to enable image comparison regardless of the source of the images |
US8442304B2 (en) * | 2008-12-29 | 2013-05-14 | Cognex Corporation | System and method for three-dimensional alignment of objects using machine vision |
EP2385483B1 (de) * | 2010-05-07 | 2012-11-21 | MVTec Software GmbH | Erkennung und Haltungsbestimmung von 3D-Objekten in 3D-Szenen mittels Deskriptoren von Punktpaaren und der verallgemeinerten Hough Transformation |
US8872851B2 (en) * | 2010-09-24 | 2014-10-28 | Intel Corporation | Augmenting image data based on related 3D point cloud data |
WO2013083730A1 (en) * | 2011-12-06 | 2013-06-13 | Hexagon Technology Center Gmbh | Coordinate measuring machine having a camera |
US20130329012A1 (en) * | 2012-06-07 | 2013-12-12 | Liberty Reach Inc. | 3-d imaging and processing system including at least one 3-d or depth sensor which is continually calibrated during use |
-
2011
- 2011-12-06 EP EP11192220.9A patent/EP2602588A1/de not_active Withdrawn
-
2012
- 2012-12-05 US US14/363,102 patent/US9443308B2/en active Active
- 2012-12-05 CN CN201280059710.0A patent/CN103959012B/zh active Active
- 2012-12-05 EP EP12797905.2A patent/EP2788717B8/de active Active
- 2012-12-05 WO PCT/EP2012/074552 patent/WO2013083650A1/en active Application Filing
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5771978A (en) | 1996-06-05 | 1998-06-30 | Kabushiki Kaisha Topcon | Grading implement elevation controller with tracking station and reference laser beam |
EP1171752A1 (de) | 1999-04-19 | 2002-01-16 | Leica Geosystems AG | Indirekte positionsbestimmung mit hilfe eines trackers |
US6608913B1 (en) * | 2000-07-17 | 2003-08-19 | Inco Limited | Self-contained mapping and positioning system utilizing point cloud data |
US20050002555A1 (en) * | 2003-05-12 | 2005-01-06 | Fanuc Ltd | Image processing apparatus |
WO2005039836A2 (de) | 2003-10-20 | 2005-05-06 | Isra Vision Systems Ag | Verfahren zur einrichtung einer bewegung eines handhabungsgeräts und bildverarbeitung |
WO2006075017A1 (de) | 2005-01-14 | 2006-07-20 | Leica Geosystems Ag | Verfahren und geodätisches gerät zur vermessung wenigstens eines zieles |
US20100046802A1 (en) | 2008-08-19 | 2010-02-25 | Tatsumi Watanabe | Distance estimation apparatus, distance estimation method, storage medium storing program, integrated circuit, and camera |
WO2011104199A1 (de) * | 2010-02-23 | 2011-09-01 | Ifm Electronic Gmbh | Überwachungssystem |
JP2011179980A (ja) * | 2010-03-01 | 2011-09-15 | Saxa Inc | 立体マーカを利用した位置計測システム |
CN102252653A (zh) * | 2011-06-27 | 2011-11-23 | 合肥工业大学 | 基于tof无扫描三维成像的位姿测量方法 |
Non-Patent Citations (3)
Title |
---|
3DPVT'08 - THE FOURTH INTERNATIONAL SYMPOSIUM ON 3D DATA PROCESSING, VISUALIZATION AND TRANSMISSION, June 2008 (2008-06-01) |
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, vol. 22, no. 11, November 2000 (2000-11-01) |
J.M. RUEGER: "Electronic Distance Measurement", 1996, SPRINGER-VERLAG |
Cited By (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2518673A (en) * | 2013-09-30 | 2015-04-01 | Ortery Technologies Inc | A method using 3D geometry data for virtual reality presentation and control in 3D space |
EP3392612A4 (de) * | 2015-12-14 | 2019-07-10 | Nikon-Trimble Co., Ltd. | Fehlererkennungsvorrichtung und programm |
US11893721B2 (en) | 2015-12-14 | 2024-02-06 | Nikon-Trimble Co., Ltd. | Defect detection apparatus and medium |
US11536567B2 (en) | 2017-09-13 | 2022-12-27 | Topcon Corporation | Surveying instrument |
EP3457080A1 (de) * | 2017-09-13 | 2019-03-20 | Topcon Corporation | Vermessungsinstrument |
IT201700112107A1 (it) * | 2017-10-06 | 2019-04-06 | A D R I S R L S | Kit di rilievo architettonico certificato di un immobile o di un terreno e metodo associato |
CN109934869A (zh) * | 2017-12-19 | 2019-06-25 | 欧姆龙株式会社 | 位置姿势推断装置、位置姿势推断方法及记录介质 |
CN109934869B (zh) * | 2017-12-19 | 2023-04-07 | 欧姆龙株式会社 | 位置姿势推断装置、位置姿势推断方法及记录介质 |
GB2576878B (en) * | 2018-09-03 | 2023-01-25 | Univ Of Northumbria At Newcastle | Object tracking |
WO2020049272A1 (en) * | 2018-09-03 | 2020-03-12 | University Of Northumbria At Newcastle | Object tracking |
CN111144478B (zh) * | 2019-12-25 | 2022-06-14 | 电子科技大学 | 一种穿帮镜头的自动检测方法 |
CN111144478A (zh) * | 2019-12-25 | 2020-05-12 | 电子科技大学 | 一种穿帮镜头的自动检测方法 |
CN114378825A (zh) * | 2022-01-21 | 2022-04-22 | 四川长虹智能制造技术有限公司 | 一种多相机视觉定位方法、系统及电子设备 |
CN114378825B (zh) * | 2022-01-21 | 2023-05-12 | 四川长虹智能制造技术有限公司 | 一种多相机视觉定位方法、系统及电子设备 |
Also Published As
Publication number | Publication date |
---|---|
EP2788717A1 (de) | 2014-10-15 |
WO2013083650A1 (en) | 2013-06-13 |
CN103959012B (zh) | 2016-09-21 |
EP2788717B8 (de) | 2018-07-25 |
CN103959012A (zh) | 2014-07-30 |
EP2788717B1 (de) | 2018-04-04 |
US9443308B2 (en) | 2016-09-13 |
US20140286536A1 (en) | 2014-09-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2788717B1 (de) | Positions- und ausrichtungsbestimmung in 6-dof | |
US9915521B2 (en) | Method for optically measuring three-dimensional coordinates and controlling a three-dimensional measuring device | |
US7242460B2 (en) | Method and apparatus for automatic registration and visualization of occluded targets using ladar data | |
EP1792282B1 (de) | Verfahren zur automatischen 3D-Bildgebung | |
EP2111530B1 (de) | Automatische stereomessung eines punkts von interesse in einer szene | |
US9322646B2 (en) | Adaptive mechanism control and scanner positioning for improved three-dimensional laser scanning | |
US7777761B2 (en) | Method and apparatus for specifying and displaying measurements within a 3D rangefinder data set | |
EP2068280B1 (de) | Korrektur von Bildverzerrungen | |
US7403268B2 (en) | Method and apparatus for determining the geometric correspondence between multiple 3D rangefinder data sets | |
EP3373251A1 (de) | Scan-kolorierung mit einer unkalibrierten kamera | |
JP4885584B2 (ja) | レンジファインダ校正方法及び装置 | |
EP1580523A1 (de) | Verfahren zur dreidimensionalen formmessung und einrichtungdafür | |
JP2012533222A (ja) | 画像ベースの表面トラッキング | |
WO2015058017A1 (en) | Balancing colors in a scanned three-dimensional image | |
WO2016040271A1 (en) | Method for optically measuring three-dimensional coordinates and controlling a three-dimensional measuring device | |
EP3989169A1 (de) | Hybride photogrammetrie | |
EP4332621A1 (de) | Laserscanner mit stereokamerasicht für verbesserte selektive merkmalabtastung | |
EP4224430A1 (de) | Scan-farbwiederherstellung | |
JP2020197493A (ja) | 画像を用いた三次元計測装置、画像を用いた三次元計測方法および画像を用いた三次元計測用プログラム | |
Gonçalves et al. | Sensor‐Based Depth Capturing | |
NO347027B1 (en) | Method and system for detecting a line above ground from a helicopter | |
JP2020197407A (ja) | 三次元計測装置、三次元計測方法および三次元計測用プログラム | |
George et al. | Optical System For Robot Guidance | |
AU2005279700A1 (en) | A method for automated 3D imaging |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
AX | Request for extension of the european patent |
Extension state: BA ME |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN |
|
18D | Application deemed to be withdrawn |
Effective date: 20131213 |