AU2001273360A1 - Mehthod and device for measuring three-dimensional shapes - Google Patents

Mehthod and device for measuring three-dimensional shapes

Info

Publication number
AU2001273360A1
AU2001273360A1 AU2001273360A AU7336001A AU2001273360A1 AU 2001273360 A1 AU2001273360 A1 AU 2001273360A1 AU 2001273360 A AU2001273360 A AU 2001273360A AU 7336001 A AU7336001 A AU 7336001A AU 2001273360 A1 AU2001273360 A1 AU 2001273360A1
Authority
AU
Australia
Prior art keywords
isa
light
intensity
capturing
reflector
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
AU2001273360A
Inventor
David S. Agabra
Basel F. Bahhour
Peter L. Diaconescu
David D. Drobnis
Jill E. Haynie
Mark S. Knighton
J. Douglas Logan
William D. Mckinley
Kent E. Sidney
Amit Tandon
Kevin H. Vuong
John Z. Zheng
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NextEngine Inc
Original Assignee
NextEngine Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US09/660,809 external-priority patent/US6639684B1/en
Application filed by NextEngine Inc filed Critical NextEngine Inc
Publication of AU2001273360A1 publication Critical patent/AU2001273360A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/24Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
    • G01B11/25Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
    • G01B11/2518Projection by scanning of the object
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/521Depth or shape recovery from laser ranging, e.g. using interferometry; from the projection of structured light

Description

DIGITIZER USING INTENSITY GRADIENT TO IMAGE FEATURES
OF THREE-DIMENSIONAL OBJECTS WITH ELEMENTARY
DEPTH DETECTION IN 3D IMAGING
BACKGROUND
Field of the Invention
[0001] The invention relates to an image capture device. More specifically, the invention relates to a low-cost three-dimensional digitizer. Background
[0002] There are four broad categories of non-contact three-dimensional digitizers. The first category is referred to as silhouette digitizers because the imaging device repeatedly takes the silhouette of the object as the object is rotated before the imaging device or the imaging device is rotated about the object. This type of digitizer is relatively ineffective at dealing with concavities in a three-dimensional object because the silhouette is unchanged by the concavity.
[0003] The second category is timing digitizers. Timing digitizers use a signal source such as a radar source. By determining the amount of time required for the signal to bounce back from the different points on the target object, surface features of the object are revealed. However, such digitizing methods are extremely expensive to implement.
[0004] A third category is projected pattern digitizers, in which a pattern of some sort is projected onto the object to be digitized and the dimensions of the object are determined from the interaction of the pattern with the object. Projected pattern digitizers fall into three main subcategories. The subcategories include contour digitizers which use spatial distortion from a projected pattern of contour lines to determine surface features of a three- dimensional object. A next subcategory is interference projected pattern digitizers, which use two sources and then based on the localized interference pattern of the two sources, determine the surface features of the three- dimensional object to be digitized. A third subcategory is referred to as color projected pattern digitizers because this category uses a projected color pattern and resulting color gradients to determine relevant information about the object to be digitized.
[0005] A final broad category is stereoscopic digitizers which employ multiple cameras to capture images of the object from different angles. From the picture, such systems perform feature identification. Then a correlation between the features in the different pictures is established to yield three- dimensional data.
BRIEF DESCRIPTION OF THE DRAWINGS [0006] The invention is illustrated by way of example and not by way of limitation in the figures of the accompanying drawings in which like references indicate similar elements. It should be noted that references to "an" or "one" embodiment in this disclosure are not necessarily to the same embodiment, and such references mean at least one.
[0007] Figure 1 shows a schematic diagram of a intensity gradient cone of one embodiment of the invention.
[0008] Figure 2 is a block diagram of a system of one embodiment of the invention.
[0009] Figure 3 is a block diagram of the control subsystem of one embodiment of the invention.
[0010] Figure 4 is a perspective view of a digitizer of one embodiment of the invention.
[0011] Figure 5 shows a perspective view of a digitizer of one embodiment of the invention with a portion of the housing removed. [0012] Figure 6 is a rear perspective view of one embodiment of the digitizer with a portion of the housing and base removed. [0013] Figure 7 is a bottom perspective view of a digitizer of one embodiment of the invention using an alternative optics arrangement. [0014] Figure 8 shows a top perspective view of the embodiment of the digitizer of Figure 7. [0015] Figure 9a shows an additional alternative embodiment of a digitizer of one embodiment of the invention.
[0016] Figure 9b is a sectional perspective view of a portion of one embodiment of the invention.
[0017] Figure 10 is a perspective view of a magnetic drive unit of one embodiment of the invention.
[0018] Figures lla-c show perspective views of components of a magnetic drive and reflector assembly of one embodiment of the invention.
[0019] Figure lib shows a rear perspective view of the reflector and arm assembly.
[0020] Figure lie shows a perspective view of the fulcrum mounting post
390.
[0021] Figure 12 is a perspective view of a projection subassembly of one embodiment of the invention.
[0022] Figure 13a is a sectional perspective view of an inclinometer disposed within the camera case.
[0023] Figure 13b is a perspective view of the inclinometer positioned relative to the image sensing array.
[0024] Figure 14a is a schematic diagram of a system of one embodiment of the invention at first mode of operation.
[0025] Figure 14b shows a schematic diagram of a macro lens solution to imaging small objects in one embodiment of the invention.
[0026] Figure 14c is a schematic diagram of an alternative macro lens solution.
[0027] Figure 15 is a perspective view of an imaging assembly of one embodiment of the invention.
[0028] Figure 16a is a diagram of a lens/aperture assembly of one embodiment of the invention.
[0029] Figure 16b is an exploded view of the assembly of Figure 15. DETAILED DESCRIPTION
[0030] In one embodiment, the system operates on the principle that depth data for a three-dimensional object may be calculated from an intensity difference resulting from an intensity gradient projected on the object. Existing image sensing arrays (ISAs) such as linear charge coupled device (CCD) sensors can detect illumination intensity to a high degree of accuracy. Based on this principle, if a light source is placed in fixed relation to the ISA such that the projected light forms an angle with the focal line of the ISA, and a gradient slide, for example, going from dark to light, from left to right, is interposed between the light source and the object, features of the object closer to the ISA are illuminated by greater intensity light than those features further away. Thus, the ISA captures a stripe of the object in which different intensities represent different depths of the object in that focal zone. This general principle works well for uniformly colored objects imaged in an otherwise dark environment, but different coloring and ambient light conditions may cause misinterpretations of the intensity data. However, if the ISA images the same stripe of the object under ambient conditions (e.g., when the light source is not illuminating the object within the focal zone) and images again when the object is illuminated by a uniform light (e.g., with no gradient (flat gradient)), these possible misinterpretations can be avoided. [0031] Particularly, the ratio VG1 - VA/ VG2- VA yields a differential that can be mapped to depth of the object. In the differential, VG1 is the value from the ISA at a point resulting from the gradient exposure, VA is the value from the ambient exposure at that point, and VG2 is the value at the point from a second gradient exposure such as the uniform light (flat gradient) or a second gradient created as described further below. The differential is computed for each point in the focal zone. Moreover, this differential also normalizes the effect of color variations and ambient light conditions. Notably, the differential is also substantially independent of intensity of the light source. Unfortunately, as a practical matter, changing slides and/or turning the light source on and off rapidly enough to permit digitization of many possible target objects is both expensive and problematic.
[0032] However, by taking advantage of the fact that the ISA integrates over time, the same effect may be created mechanically using a shutter which causes 0% to 100% of the light to illuminate the target object within the focal zone during the cycle. Moreover, by overdriving the shutter, the white light condition and ambient condition, can be created. Specifically, if the imaging time of the CCD is 5 milliseconds, in an initial 5 milliseconds the shutter does not impinge on the light source, thereby allowing the imaging sensing array to image the fully illuminated object. The next 5 milliseconds, the shutter passes from 0 to 100% blockage of the light, thereby creating the intensity gradient within the focal zone. During the next 5 milliseconds, the shutter continues to drive so that the light is entirely blocked and the ambient condition image is obtained. The processing of each of these images (including the creation of the differential) may be offloaded to an attached host as discussed in greater detail below.
[0033] An intensity gradient may alternatively be created by sweeping the light through the focal zone. For example, by sweeping a light stripe from left to right through the focal zone, the ambient light image may be captured before the light enters the zone. A first gradient is captured from the first entry of the light into the zone until the light is entirely within the zone. A second gradient is captured as a light translates out of the zone to the right. The second gradient is the opposite of the first gradient and is not flat as in the fully illuminated case. An analogous set of images may be captured as the light sweeps back from left to right. One advantage of sweeping the light is that two gradients are generated as the light moves from right to left and two gradients are generated as the light moves from left to right. Thus, the sweeping can be performed at half speed without a reduction in imaging performance. [0034] The differential may take the same form as discussed above. Alternatively, the differential may be computed as X1/(X1 + X2), where Xt = NG1 - NA and X2 = NG2 -NA. To reduce noise sensitivity, the larger magnitude gradient should be selected for the numerator of the ratio. Color intensity is given by X1 + X2.
[0035] Figure 1 shows a schematic diagram of a intensity gradient cone of one embodiment of the invention. This example has been simplified to include only a single capture of one gradient as might be applicable to uniformly shaped and colored objects in a dark environment. This example, of course, could be expanded to cover non-uniform cases as discussed above using intensity differentials. A light source is located a distance L along a line normal to a line of sight of a linear ISA. The light source projects a gradient having a minimum intensity IA and maximum intensity IB. The angle ΘA corresponds to the angle between the minimum intensity edge of the projected gradient and the line normal to the ISA line of sight. ΘB corresponds to the angle defined by the maximum intensity edge of the projected gradient and the normal line, θ corresponds to the angle defined by a line from the gradient origin to the point on the target for which distance is to be determined and the normal line. Between IA and IB , intensity is proportioned to the angle θ and is given by I(θ) = (IB- IA) [(θ - ΘA)/ (ΘB- ΘA)] + IA.
[0036] Thus, angle as a function of intensity is given by Θ(I) = (ΘB- ΘA) * [(I - IA)/ (IB - IA)] + ΘA. The distance D is the given by D = L * tan θ. Therefore, distance as a function of intensity measured at the ISA is given by D = L * tan
{(ΘB- ΘA) * [(I - IA)/ (IB- IA)] + ΘA}. In one embodiment, the light source is fixed relative to the ISA and the minimum and maximum intensities are known or can be calibrated at run time. This allows intensity derived during capture to be mapped easily to distance with limited computational complexity. [0037] In one embodiment, depth measurements are determined on a pixel by pixel basis. Because intensity captured by the single pixel is independent of intensity captured by other pixels and intensity maps directly to depth, it is possible to achieve one depth measurement per pixel. Moreover, it would be possible to obtain a depth measurement from an image sensor having a single sensing element. In an alternative embodiment, adjacent groups of pixels may be treated as an element with the aggregate captured intensity used to determine the depth measurement without regard to an energy distribution among the pixels in the group. This yields one measurement per elementary group of pixels.
[0038] As indicated, it is possible to calculate the depth data directly from the intensity information. However, the speed and processing power required are reduced when a lookup table (LUT) based on a prior calibration is employed to derive depth data based on the differentials. This also allows nonidealities of the physical equipment to be accounted for in the selected LUT entity, based on the prior calibration. Accordingly, the embodiment of the invention maintains a LUT and indexes into the LUT based on the differential. [0039] Figure 2 is a block diagram of a system of one embodiment of the invention. A digitizer 70 is coupled to a host node 50. This coupling may be by a bus 60 such as the Universal Serial Bus (USB), IEEE 1394 bus, or any other suitable data transfer system. It is also within the scope and contemplation of the invention for the digitizer to communicate with the host mode via a wireless interconnection. Host node 50 may be a personal computer, a work station, an internet appliance, or any other device that provides sufficient intelligence and processing power to render images from the data obtained by the ISA. The digitizer 70 captures image data and may forward it to the host node 50 for rendering. In this way, the processing on the digitizer 70 may be limited, permitting lower cost construction.
[0040] The digitizer 70 includes a projector to project a stripe of white light through a projection window 74 onto a remote object such as a person 82 on a turntable 80 remote from the digitizer. The digitizer also contains an image sensing array (ISA) aligned with an image capture window 76 which captures the image of the object 82 within a focal zone. In one embodiment, the ISA is a linear charge coupled device (CCD) or complementary metal oxide semiconductor (CMOS) sensor, and the focal zone is a line on the target object. In some embodiments, the digitizer includes a base 72 about which the upper unit, including the projector and the ISA, can rotate in either direction. This permits the focal line to be swept back and forth across a target object through an arc. This sweeping reduces the loss of detail in the captured image that results from shadowing on the object from the perspective of an immobile focal line. The digitizer 70 also includes a wireless interface to communicate with a turntable 80 via a wireless link 84.
[0041] Turntable 80 may be the type described in co-pending application entitled Wireless Turntable, Serial No. 09/660,810, assigned to the assignee of the instant application. Via wireless link 84, the digitizer sends commands to the turntable 80 and receives from the turntable indications of the angular position of the turntable surface relative to a home position. When the digitizer is activated, it searches for the turntable 80 by sending a signal to which the turntable 80 is required to respond. If the turntable responds, the digitizer looks for a predetermined pattern that is expected to be present on the turntable surface. For example, the pattern may be concentric circles on the turntable surface. In such case, based on the image captured, the digitizer can both find the turntable and determine its distance from the digitizer. Then after the response is received, the digitizer sends a "go home" signal to the turntable. In some embodiments, the digitizer sends acceleration and rotation profiles to the turntable to control its rotation. Each profile may be retained in firmware on the digitizer or downloaded from host node 50. [0042] Generally speaking, the projection portion of the digitizer 70 is retained in fixed relation to the imaging portion. The projection portion produces a light stripe as noted previously on the object 82. By either sweeping the light stripe back and forth through the focal line or by mechanically blocking the stripe at a known rate, the intensity gradient can be created. In one embodiment, the blocking is from 0% to 100% during a cycle. Because the ISA integrates the illumination over time, the outline of a three-dimensional surface is reflected in the data captured by the ISA. This is because protruding features will remain illuminated longer. Accordingly, more photons are captured by the ISA corresponding to those features. After repeating this process one stripe at a time as the object is rotated by turntable 80 or through the course of sweeping the entire digitizer back and forth as it rotates about the base, cost effective three-dimensional imaging is effected. The digitizer may also be used to capture high resolution scans of two dimensional objects by sweeping back and forth across the object. This feature is particularly desirable in the context of digitizing works of art.
[0043] Figure 3 is a block diagram of the control subsystem of one embodiment of the invention. The processor 304 is coupled to a random access memory (RAM) 312 and an erasable programmable read only memory (EPROM) 308. The EPROM 308 contains firmware necessary for booting the processor and may, for example, include rotation profiles and the command set for the wireless turntable. The wireless interface 302 is used by the processor 304 to signal the wireless turntable. The processor 304 is coupled to the ISA 300 which sends image data to the processor for storage in the RAM 312 or to be forwarded to the host over USB or other suitable connection. The processor 304 also controls the drive motor 306 to affect sweeping rotation of the digitizer. The processor 304 also controls projection subsystem 314, particularly whether the light source is on or off, as well as in certain embodiments, the activation and deactivation of the shuttering or sweeping of the light beam. The same general control structure is employed in the various mechanically varying embodiments of the invention described below.
[0044] Figure 4 is a perspective view of a digitizer of one embodiment of the invention. A housing 10 is coupled to a base 12. In some embodiments, the coupling between housing 10 and base 12 may be a rotatable coupling, such that the housing 10 projection and imaging units within may be rotated axially about base 12. Housing 10 defines a projection window 14 and an image capture window 16. In one embodiment, housing 10 is molded out of ABS. Housings made of other plastics or metal are all are within the scope and contemplation of the invention. The material used for the windows may vary from one embodiment to the next, depending on the optics employed within the housing. [0045] Figure 5 shows a perspective view of a digitizer of one embodiment of the invention with a portion of the housing removed. A mounting plate 18 which forms the bottom of the housing serves as the mounting surface for most of the internal components. A motor 42 is also mounted to the mounting surface to drive rotation of the assembly about the base 12. The requisite gear assembly may be arranged to reside in the housing and /or the base. A camera case 20 retains a lens in optical communication with an ISA. The camera case 20 prevents ambient light from distorting the image captured by the ISA. Also mounted in a fixed position relative to where the camera case 20 is a light source 22. Adjacent to the camera case 20 is a circuit board, including a processor and a memory that provide the brains and storage, respectively, for the digitizer.
[0046] A wireless interface is also provided and may signal the turntable (not shown) through the imaging windows. The wireless interface may for example be an infrared interface or a radio transceiver, either of which may employ well understood protocols for sending and receiving information from the turntable. In one embodiment, light source 22 may be a 300 watt halogen tube. A first elliptical reflector 24 is mounted on one side of the light source 22, while a second elliptical reflector 26 is mounted on the other side of the light source 22. In this manner, the first elliptical reflector 24 focuses the light from light source 22 back on the second elliptical reflector 26, which in turn, focuses a light to a focal point some distance from the light source. In one embodiment, first elliptical reflector 24 is semi-circular. Mounted at that focal point is a light homogenizer 28, which in one embodiment of the invention, may be polished float glass. The float glass basically functions as a light pipe that uniformly distributes the intensity such that a substantially uniform intensity light stripe exits the terminal side of the float glass. [0047] Coupled to the float glass is a light folding mirror 30, which is used to fold the light projecting out of the homogenizer 28 on itself. By folding the light, a smaller size lens may be employed subsequently to focus the light on a target object. In one embodiment, the folding mirror 30 is omitted and a larger lens is used. A lens 34 is mounted to focus light from the light source through the projection window to a location remote from the digitizer. An additional mirror or mirrors 36 may be employed to ensure optical communication between the light source 22 and the lens 34. The lens 34 and other optical component may be manufactured from various suitable materials known in the art.
[0048] A shutter blade 32 is mounted, such that when driven, it will impinge on the light exiting the homogenizer, such that it varies between blocking 0% and 100% of that light from reaching the lens 34. The shutter blade is mounted to a spider, such as might be found in a conventional stereo speaker, which is driven by an electromagnet 40. In this manner, the shutter can be driven to pass through an entire cycle of 0% to 100% blockage, in a hundredth of a second. Moreover, the spider mechanism has been found to be quite smooth, resulting in minimal mechanical vibration, which might otherwise have deleterious effects on the imaging of the system. In an alternative embodiment, the shutter may be mounted to a pair of leaf springs and driven by a coil.
[0049] Figure 6 is a rear perspective view of one embodiment of the digitizer with a portion of the housing and base removed. The housing and its projection and imaging subsystems are mounted on a central post 48 extending from a floor of base to engage the mounting platform 18 of the housing. A bronze bushing may be used around the central post to aid in achieving smooth rotation. By using a hollow axial post 48, the overall height of the unit is reduced as in such an embodiment the light source may be mounted to extend it down within the post 48.
[0050] Transformer 54 resides within the base and is used to power the various systems of the digitizer. An optical interrupter for 52 is used to identify where the upper unit is relative to the base as it rotates about the axis 48. To effect this, one or more blades are molded to interrupt the sensor as the upper unit rotates. A USB port 44 is provided to permit data to be sent back and forth to a host node. Other type of ports could be used instead or in addition to USB. An AC power port 46 is similarly provided within the base to provide the power to transformer 54.
[0051] Figure 7 is a bottom perspective view of a digitizer of one embodiment of the invention using an alternative optics arrangement. A portion of the camera case is removed to reveal the image sensing array 100, which in one embodiment of the invention, may be a 5340 pixel linear image sensor, such as the one available from Toshiba America, Inc. of New York, New York, available under the part number TCD2558D. The light source 122 is mounted within a parabolic reflector 124, which focuses the light from light source 122 onto a curved reflector 126. In one embodiment, the light source 122 is a 150 watt single end halogen bulb. The curved reflector 126 spreads the light into a light stripe that is then reflected through the projection window onto the target object at a predetermined distance from the digitizer. The curved reflector 126 may be moved back and forth to sweep the light stripe through the focal zone on the target object. The reflectors may be manufactured from plastic or polished metal. In one embodiment, stamped aluminum is used.
[0052] Figure 8 shows a top perspective view of the embodiment of the digitizer of Figure 6. The curved reflector 126 can still be seen beneath the light source (not shown). Motor 142 drives a gear assembly 152, which engages drive gear 154 to permit rotation of the upper unit, including the projection system (light source and reflectors) and the image sensing array 100 about the base 112. By rotating about the base, the digitizer is provided with an additional degree of freedom, which facilitates scanning in some situations, as described in more detail below.
[0053] Figure 9a shows an additional alternative embodiment of a digitizer of one embodiment of the invention. Similar to the embodiment discussed with reference to Figure 6 and Figure 7, this embodiment uses parabolic reflector 224 in conjunction with light source 222 to produce a spotlight. However, rather than using a curved reflector (which performs a light spreading function), a flat reflector 226 is employed to reflect the spot of light to the projection window 214. Projection window 214 is manufactured from a pane of glass or plastic to have a plurality of concave or convex ridges. Each such ridge acts as a lens spreading the light thereby changing the spot of light to a light stripe which has relatively good intensity uniformity from top to bottom. A magnetic drive unit 250 is used to move reflector 226 back and forth to cause the spot of light to move back and forth across the projection window 214 and, therefore, the light stripe to move back and forth across in the focal line of the ISA. It is also within the scope and contemplation of the invention to effect the sweeping by moving the light source and parabolic reflector while maintaining the other reflector stationary.
[0054] In one embodiment, an IR filtering or hot mirror (not shown) is interposed between the light source 222 and the reflector 226. A fan may be introduced between the IR filter and the light source 222 to cool the light source 222. However, in such an embodiment, partitions may be desirable such that the housing, in conjunction with the partitions, the IR filter, and the parabolic reflector 224 form a fan enclosure which prevents turbulence created by the fan from disturbing the regular movement of reflector 226. The motor 242 is provided to drive gear assembly 252, which in turn drives major gear 254, thereby causing the upper portion of the digitizer including the projection unit and the imaging unit to rotate about the base 212. This permits the digitizer to sweep back and forth while scanning an object. Thus, an object some distance from the digitizer may be scanned, for example, 90° of the object at a time, requiring only four rotations by the turntable. In one embodiment, the gear ratio is 512. Similar motor and gear assemblies may be used in each of the various above described embodiments. It is, of course, possible for the digitizer to image an object continually rotated by the turntable. The sweeping the focal line (even in a continually rotating environment) also permits features to be more accurately identified than would be possible with an immovable focal zone. Particularly, sweeping the focal line reduces inaccuracies due to shadowing. [0055] The electronics board 258 is substantially the same as in the other embodiments, as is the imaging subsystem. Wireless signaling interface 260, which in this embodiment is a infrared signaling interface, signals the turntable through the imaging window. An activation switch 262 is supplied on the electronics board 258 to permit the system to be activated. [0056] Figure 9b is a sectional perspective view of a portion of one embodiment of the invention. Gear assembly 252 is mounted on gear box 286. Gear box 286 is elastically coupled to the lower unit by bias spring 282. Bias spring 282 biases gear assembly 252 into engagement with major gear 254. Additionally, bias spring 282 biases the shaft 248 to lean in a consistent direction. This is desirable, because if the shaft were free to float from side to side, even given quite tight tolerances, that minor variation at the digitizer may represent a significant deviation eight feet away in the focal zone. Thus, absent some biasing mechanism to ensure consistency in the shaft, risk of lost data is increased. Anti-vibration spring 280 is compressed between washer 290 that moves with the shaft and bushing 288 that moves with the upper unit. As a result, anti-vibration spring 280 increases rigidity of the upper unit and prevents vibration and wobble from side to side. Use of these bias springs permits a less expensive bearing with greater tolerances to be used. Power cables 284 are fed through the hollow shaft 248 to provide power to the light source (not shown).
[0057] Figure 10 is a perspective view of a magnetic drive unit of one embodiment of the invention. As previously noted, the reflector 226 is coupled to the magnetic drive unit 250. Specifically, it is coupled to an arm 276 which is able to pivot in a horizontal plane. The distance of pivot is controlled to some degree by a pair of spring steel bands 280, which provide a restraining force against rotation from a central position. A first magnet 270 having a first polarity is positioned on one side of the arm 276. A second magnet having the opposite polarity is positioned on another side of arm 276. A coil 274 runs around the arm and between the first and second magnets 270 and 272. When the coil is pulsed with current, the magnets 270 and 272 intermittently apply torque to the arm 276. The result is that arm 276 moves back and forth in an arc within a horizontal plane. Movement of the arm 276 is translated to movement of the reflector 226, and consequently, sweeping back and forth with the light beam reflected thereby. This system employs certain resonance principles to improve energy efficiency such that the power cost of sweeping the reflector is quite low.
[0058] Figures lla-c show perspective views of components of a magnetic drive and reflector assembly of one embodiment of the invention. The reflector magnetic drive assembly as shown in Figure 10a is similar in many respects to that shown in Figure 9. However, rather than mounting the magnetic drive portion behind the reflector, it is mounted in front and below the reflector. Thus, first magnet 370 having a first polarity is positioned on one side of arm 376, and a second magnet having the opposite polarity is positioned on another side of arm 376. A coil runs around the arm between first and second magnets 370 and 372. When the coil is pulsed with current, the magnets 370 and 372 intermittently apply a torque to the arm 376. The result is that arm 376 moves back and forth in an arc within a horizontal plane. [0059] Movement of the arm 376 is translated into movement of the reflector 326, and consequently, sweeping back and forth of the light being reflected thereby. The reflector 326 is mounted on a fulcrum post 390. No spring steel bands are provided. Rather, a pair of springs 386 are coupled at the base of the mirror to provide opposing restraining forces to cause the mirror to move back and forth in a resonant manner. The spring constants of springs 386 dictate the resonance frequency, which translates to the sweep rate. This mounting results in smooth consistent low friction movement of the reflector during operation. In one embodiment, the springs 386 are selected to yield a resonance frequency of approximately 50 Hz. A much smaller spring 384 is coupled to fulcrum post 390 and mounting clip 382. The purpose of spring 384 is merely to hold the reflector into the circular fulcrum member discussed below. By having the springs 386 which dictate the resonance frequency of the apparatus positioned at the bottom of the reflector to be moved, the torque created by the magnetic attraction and repulsion of the arm 376 is applied significantly more proximately to those springs 386 than were they at the top of the reflector, thereby improving system efficiency. Additionally, a light weight reflector can be used without risking deformation of the reflector during operation.
[0060] Figure lib shows a rear perspective view of the reflector and arm assembly. Supporting members 377 that provide structural rigidity to the reflector 326 are minimized to reduce weight of the overall structure that must be moved. A circular fulcrum engaging member 394 is molded on mounting clip 382. Upper spring engaging post 398 is also coupled to mounting clip 382. An extension of the arm 376 forms lower fulcrum engaging end 392. Lower spring mounting posts 396 are also evident.
[0061] Figure lie shows a perspective view of the fulcrum mounting post 390. A fulcrum mounting post defines an upper circular fulcrum 391 and a lower channel fulcrum 393. This fulcrum arrangement prevents lateral shifting of the reflector 326 and ensures a clean pivot side to side. The fulcrum defines a true pivot point with minimal frictional engagement. Mounting the magnetic drive below and in front of the reflector to be moved, permits a more compact finished system.
[0062] Figure 12 is a perspective view of a projection subassembly of one embodiment of the invention. A parabolic reflector 524 is coupled to a drive shaft of motor 550. Also coupled along the drive shaft of the motor are plurality of sensor blades which interrupt an optical sensor 554 to indicate positioning of the parabolic reflector 524 during rotation. The motor shaft is positioned to be aligned with the light source 522 such that a linear extension of the motor shaft would intercept the light source 522. The parabolic reflector 524 as mounted on the shaft slightly off center. However, the tilt of the parabolic reflector 524 as mounted ensures that a focal line of the reflector intersects the light source 522 throughout the rotation. The rim 526 of reflector 524 is counterbalanced to provide for smooth rotation of the parabolic reflector 524. Fan blades 556 may be coupled to the rim 526 to assist in the evacuation of heat generated by the light source 522.
[0063] The parabolic reflector reflects a light spot from the light source 522 onto the projection spreading window 514 which is discussed in connection with a previously described embodiment has the effect of spreading the light spot into a vertical light stripe. Because the parabolic reflector is mounted off center as the motor rotates the reflector, the light spot translates through a substantially circular path on the projection window 514. As a result of the light spreading, the effect in the focal zone, some distance from the projection window, is a light bar sweeping back and forth. In this manner, two gradients may be generated and the three dimensional features calculated as described above in connection with other embodiments that sweep the light stripe through the focal zone.
[0064] Figure 13a is a sectional perspective view of an inclinometer disposed within the camera case. It has been found that a tilt of as little as .08° will change the imaging characteristics of the ISA where a desired resolution is .1 inches. Inasmuch as table surfaces may often have slopes greater than this, it is desirable to be able to detect the tilts with an accuracy of at least .08° within the digitizer. Once detected, the slope can be factored out in the rendering of the imaged object on the host node. The purpose of the inclinometer is to permit a determination of the amount the digitizer is tilted when resting on a surface. Thus, where the digitizer is placed on an uneven table surface, the inclinometer is used to detect the tilt to permit the subsequently rendered image to be compensated for the tilt and resulting distortion in the image captured. Since gravity provides a force having a true direction independent of the surface tilt, the tilt can be measured off a gravitational orientation unit such as a pendulum, a plum line, a liquid level, etc. [0065] A reflector 306 is disposed on a pendulum 304. The pendulum assembly rests in fulcrum mount 302 which is coupled to the camera case 220. The mount 302, reflector 306, and pendulum 304 collectively are referred to herein as the inclinometer. The inclinometer is mounted within the camera case 220, such that regardless of the tilt, no blockage of lens 320 occurs. Since the pendulum 304 will hang true vertical, regardless of the tilt of the table, by appropriately shining a light on the reflector 306 disposed on the pendulum 304, the light is reflected to a point on the ISA (not shown), and where the light strikes indicates the tilt in the direction the turntable is facing. [0066] Figure 13b is a perspective view of the inclinometer positioned relative to the image sensing array. The LED 308 is disposed on circuit board 258 to shine on the reflector of the inclinometer. This light is then reflected to the ISA 300. Because the fulcrum mount 302 has very low friction, the settling time of the pendulum is very high. Very low friction is desirable because it is desired that very small changes in tilt result in movement of the pendulum 304. Unfortunately, due to this long settling time, a single measurement of the reflected light may be at significant variance from the actual tilt by virtue of the swinging of the pendulum. Various ways exist to account for this in determining tilt. One is to take the maximum and minimum as the pendulum swings through its arc and average those. A second way would be to integrate over, for example, 20 seconds. Either method allows the inclinometer in conjunction with the ISA to determine to a high degree of accuracy the tilt to which the digitizer is subjected.
[0067] While the pendulum only determines a tilt in a single direction, because the upper unit of the digitizer can rotate, it can rotate by, for example, 90° and determine the tilt in the second direction, thereby determining the tilt in a second direction, and accordingly determining the combined tilt in an X and Y direction for the surface on which it is placed.
[0068] Figure 14a is a schematic diagram of a system of one embodiment of the invention at first mode of operation. In normal mode of operation as shown in Figure 12a, ISA 400 is a focal line that can image an object between a minimum distance and a maximum distance perpendicularly from a lens 420. The minimum and maximum distance at which the ISA can focus dictate the maximum dimension of an object that can be imaged. It also dictates the width that the light provided by light source 422 must be when it reaches the focal zone. In a typical embodiment, the minimum distance might be four feet, and the maximum distance might be eight feet. This yields the maximum cross-dimension for the object of four feet. However, when imaging small objects, for example, the size of a penny, imaging at a distance of four feet is unlikely to yield an acceptable imaging result.
[0069] Figure 14b shows a schematic diagram of a macro lens solution to imaging small objects in one embodiment of the invention. By interposing an additional lens 424 and a wedge prism 426 along the focal line of ISA 400, the focal line is bent to intersect the projected light at a point closer to the digitizer. Thus, with the additional magnification resulting from additional lens 424 and the closer focal zone caused by the wedge prism 426, significantly smaller objects can be imaged.
[0070] Figure 14c is a schematic diagram of an alternative macro lens solution. In this embodiment, instead of a wedge prism, a pair of 45° reflectors 428 are used to move the focal line to intersect the light beam closer to the digitizer.
[0071] Figure 15 is a perspective view of an imaging assembly of one embodiment of the invention. Camera case 520 is coupled to mounting plate 518 of an upper unit and circuit board 258 on which the ISA (not shown) is mounted. A macro lens 530 is movably coupled to mounting plate 518, such that in the first position, it is interposed (as shown) in the focal line of the ISA, and in the second position, it does not impinge on the focal line of the ISA. It is envisioned that the digitizer may be switched back and forth in and out of macro mode with a toggle switch, slider, or some other mechanism which causes the macro lens to move from the first position to the second position, and vice versa.
[0072] Figure 16a is a diagram of a lens/aperture assembly of one embodiment of the invention. Camera case 620 is mounted to enclose the ISA of any of the various embodiments of the invention. A camera portal 660 provides an optical path from the ISA to the outside world. In some embodiments of the invention the camera portal 660 may include a lens barrel. Lens/aperture assembly 640 holds a plurality of lens/aperture combinations 642, 644, 646. The lens /aperture assembly 640 includes a toothed wheel mounted on a shaft 656 and biased by bias spring 654 for stability. Stepper motor 650 drives the toothed wheel via drive gear 652. The lens/aperture combinations, in one embodiment, may each be distinct lens barrels. In another embodiment the lens /aperture combination may merely be a lens and aperture to add on to an existing lens barrel for the ISA. In the second case, one of the locations on the wheel such as 642 may have no lens and provide a large enough aperture so that it does not impinge on the existing lens barrel. While three lens/aperture combinations are shown, more or fewer may be provided. [0073] Additionally, it is within the scope and contemplation of the invention to have apertures and lenses on distinct wheels so that each aperture can be applied with each lens to yield a larger number of possible lens/aperture combinations. One embodiment has three possible combinations, one for distant three-dimensional imaging, one for distant two- dimensional imaging and one for close-up two- and three-dimensional imaging. In one embodiment, the selection of the lens /aperture combination may be based on input from a user at a host. Once the user identifies the conditions, e.g., desired focal distance, the correct lens/aperture assembly is positioned by the system automatically. In another embodiment, the digitizer itself identifies the correct lens /aperture combination in the course of its acquisition of the orientation fixture. For example, if the digitizer sweeps looking for the turntable using the distance three-dimensional lens and does not find the turntable, it may then transition to the close-up three-dimensional lens/apparatus combination and sweep again. If the turntable is then found, the close-up combination is selected. In another example, the digitizer may sweep looking for the turntable and then select a correct combination for the turntable at the distance it is found. It should be understood that this is merely illustrative and other methods of lens/aperture combination selection are within the scope and contemplation of the invention. [0074] Figure 16b is an exploded view of the assembly of Figure 14a. Camera portal 660 is shown along with positioning posts 670. Positioning posts 670 engage recesses 672 in the back surface of lens/aperture assembly 640. In this manner proper alignment of the lens /aperture combination over the camera portal 660 is assured. When a transition between lens/aperture combination is desired, the stepper motor 650 drives the wheel to approximately align the desired lens. The bias spring (not shown) biases the recesses 672 over the posts 670 such that the desired consistent alignment is achieved. In this manner because the alignment will be consistent from one use of the lens/aperture combination to the next, an initial calibration will compensate for any deviation caused by manufacturing tolerances. [0075] A majority of the discussion above has been related to scanning an object rotated by a turntable some distance from the scanner, or in the alternative, a two-dimensional scan (of an object that is not rotated). However, the same digitizer configuration can be used to, for example, image a room from the inside creating a panoramic three-dimensional view. For example, by setting the digitizer near the center of the room and allowing it to rotate somewhat more than 360° while scanning, it will image its surroundings. It is then a software matter for the host to reconstruct the room. [0076] In another mode of operation, the system may be used in a modified stereographic image techniques. The system uses the turntable under the digitizer control to present two aspects of the object to the digitizers. By capturing two dimensional images of both aspects using the ISA described above, two very high resolution pictures are created. These pictures may then be used to construct a three-dimensional image following known stereoscopic techniques. In some cases, it may be desirable to use multiple image capture techniques to ensure the best possible resulting image. Thus, for example, the digitizer may capture the target object using the earlier described intensity gradient based imaging and then also capture the image in a stereoscopic mode. By comparing and/or averaging the resulting images, certain anomalies resulting from either technique alone may be eliminated. [0077] It is desirable that the upper unit not be permitted to rotate indefinitely in one direction, as such rotation could cause damage to the connecting cables and create additional stresses in the system degrading the system's longevity. However, a hard stop is not feasible, because that would prevent the rotation of greater than 360° which is required to ensure a good matchup of a 360° panoramic image. One solution to this is to provide a stop which shifts in, for example, a 30° arc but has hard stops on the extremes of that arc. In this manner, the digitizer can rotate clockwise until the shifting stop reaches its hard stop at the far edge of the 30° arc. Then scanning can begin in a counter-clockwise direction and continue until the sliding stop has been pushed back across its 30° arc to the opposite side hard stop. In this manner, the digitizer can scan a 390° arc. Larger and smaller arc stops are within the scope and contemplation of the invention.
[0078] In the foregoing specification, the invention has been described with reference to specific embodiments thereof. In some cases, certain subassemblies are only described in detail with one such embodiment. Nevertheless, it is recognized and intended that such subassemblies may be used in other embodiments of the invention. It will also be evident that various modifications and changes can be made to the specific embodiments described without departing from the broader spirit and scope of the invention as set forth in the appended claims. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims (46)

CLAIMSWhat is claimed is:
1. A method comprising: projecting a spatially-distributed gradient of light intensity onto an object; capturing a measurement for a light intensity at a point on the object, which relates to location of the point along the gradient; deriving dimensional data for the point based on the measurement; and assembling the dimensional data along with dimensional data similarly captured for a set of other points on the object into a three- dimensional model of at least a portion of the object.
2. The method of Claim 1 wherein projecting comprises: varying illumination within a focal zone over time and location to create the gradient.
3. The method of Claim 1 wherein projecting comprises: forming a stripe of light having a width, the stripe to illuminate the object; and moving the stripe of light across a focal zone at least until a portion of the stripe has traversed the entire focal zone.
4. The method of Claim 1 wherein projecting comprises: illuminating the object with a gradient fixed in time for a capture period.
5. The method of Claim 2 wherein capturing comprises: integrating the illumination intensity within the focal zone during one capture period.
6. The method of Claim 1 wherein deriving comprises: creating a ratio of measurements for the point corresponding to intensity at the point under a plurality of lighting conditions.
7. The method of Claim 1 wherein the capturing is performed by a linear image sensing array and a focal zone is a single line of points on the object.
8. An apparatus comprising: a projection unit to project a spatially distributed gradient of light intensity onto an object; and a capture unit that measures light intensity at a plurality of points on the object, the light intensity at each point indicative of a location of that point along the gradient.
9. The apparatus of Claim 8 wherein the projection unit comprises: a curved reflector mounted radially asymmetrically on a rotatable shaft; and a light source disposed to be in a focal line of the curved reflector while rotated.
10. The apparatus of Claim 9 further comprising: a light spreader interposed between the curved reflector and a focal zone.
11. The apparatus of Claim 8 wherein the projection unit comprises: a reflector; a light source to illuminate at least a portion of the reflector; and a drive unit coupled to the reflector to drive the reflector to sweep incident light from the light source through a focal zone.
12. The apparatus of Claim 11 wherein the projection unit further comprises: a light spreader interposed between the reflector and the focal zone.
13. The apparatus of Claim 11 wherein the reflector is curved to cause light from the light source to form a light stripe.
14. The apparatus of Claim 8 wherein the projection unit comprises: a light source; and a variable shutter to prevent a first percentage to a second percentage of light from the light source from illuminating a focal zone during a capture period.
15. The apparatus of Claim 8 wherein the projection unit and the capture unit are retained in fixed relation to each other.
16. A method comprising: rotating an element in an optical path relative to an object to sweep a focal zone of an image sensing array (ISA) over a portion of the object; and capturing data in the ISA which can be used to determine three-dimensional data for the object within the portion over which the focal zone sweeps.
17. The method of Claim 16 further comprising: repeating the rotating and capturing for a plurality of aspects of the object.
18. The method of Claim 16 further comprising: assembling from the data captured a three-dimensional model of at least a portion of the object.
19. The method of claim 16 wherein the ISA is a linear image sensor.
20. A method comprising: moving an element in an optical path relative to an object to sweep a focal zone of a linear image sensor over a portion of the object; capturing in the linear image sensor successive lines of data as the focal zone sweeps, which can be used to determine three-dimensional data for the object within the portion of the object over which the focal zone sweeps.
21. The method of Claim 20 further comprising: repeating the moving and capturing for a plurality of aspects of the object.
22. The method of Claim 20 further comprising: assembling from the data captured a three-dimensional model of at least a portion of the object.
23. A method comprising: capturing a first two-dimensional image of an object; moving the object relative to a capturing device to expose a different aspect of the object to the capturing device; capturing a second two-dimensional image of the object; and deriving a first three-dimensional representation of the object from the first and second two-dimensional images.
24. The method of Claim 23 when the capturing device is a linear image sensing array.
25. The method of Claim 23 further comprising: capturing with the capturing device an intensity gradient based three-dimensional representation of the object.
26. The method of Claim 23 further comprising: creating a three-dimensional model by an alternative method; and comparing the first three-dimensional representation with the alternative created model to yield an improved accuracy image.
27. An apparatus comprising: an image sensing array (ISA) to capture image data in a first mode and tilt data in a second mode; and a gravitational orientation unit (GOU) responsive to a relative orientation of gravity to alter light falling on the ISA in relation to the relative orientation of gravity.
28. The apparatus of Claim 27 wherein the GOU comprises: a pendulum having a reflective element mounted thereon.
29. The apparatus of Claim 28 further comprising: a light emitting diode (LED) mounted to cast light on the reflective element when the LED is on.
30. An apparatus comprising: an image sensor array (ISA) to capture three-dimensional data about an object; a lens/aperture assembly having a plurality of lens/aperture combinations; and a controller to automatically select a suitable lens /aperture combination based on at least a distance of the object from the ISA.
31. The apparatus of claim 30 wherein each lens/aperture combination is a lens barrel.
32. An apparatus comprising: a shaft; a housing; a bearing mounted to rotate relative to the shaft; an image sensing array (ISA) coupled to the housing; and a spring to bias the shaft relative to the housing, such that successive captures by the image sensing array are consistently aligned with respect to one another.
33. The apparatus of claim 32 comprising a bias spring along each bearing axis having a tolerance above a threshold.
34. A method comprising: capturing an intensity at a location on a surface in a single pixel of an image sensing array (ISA); and converting the intensity into a measurement of distance to the location relative to a reference point independently of data from other pixels of the ISA.
35. The method of claim 34 wherein the ISA is a linear image sensor.
36. The method of claim 35 wherein the linear image sensor is one of a linear charge coupled device (CCD) and a photo diode array.
37. The method of claim 34 further comprising: comparing a plurality of captures of the intensity at the location under different conditions to compensate for non-homogenous environments or surface.
38. The method of claim 34 further comprising: comparing a plurality of captures of the intensity at the location at different points in time to compensate for non-homogeneous environments or surfaces.
39. A method comprising: capturing an intensity at a location on a surface in an elementary group of pixels on an image sensing array (ISA) without regard to intensity distribution within the group; and converting the intensity into a measurement of distance to the location independently of data from other pixels on the ISA.
40. The method of claim 39 wherein the ISA is a linear image sensor.
41. The method of claim 40 wherein the linear image sensor is one of a linear charge coupled device (CCD) and a photo diode array.
42. The method of claim 39 further comprising: comparing a plurality of captures of the intensity at the location under different conditions to compensate for non-homogenous environments or surfaces.
43. The method of claim 39 further comprising: comparing a plurality of captures of the intensity at the location at different points in time to compensate for non-homogeneous environments or surfaces.
44. A method comprising: capturing a spectral energy distribution returned from a location on a surface in a single pixel of an ISA; and converting the spectral energy distribution into a measurement of distance to the location relative to a reference point independently of data from other pixels of the ISA.
45. A method comprising: altering one of a spatial and optical relationship between an image sensing array (ISA) and a surface; observing a variation of an electrical signal at a single pixel on the ISA responsive to the alteration; and converting the variation to a measure of distance to a location on the surface relative to a reference point, independently of data from other pixels of the ISA.
46. A method comprising: altering one of a spatial and optical relationship between an image sensing array (ISA) and a surface; observing a variation of an electrical signal at an elementary group of pixels on the ISA without regard to variations in electrical signals within the group responsive to the alteration; and converting the variation to a measure of distance to a location on the surface relative to a reference point, independently of data from other pixels of the ISA.
AU2001273360A 2000-09-13 2001-07-11 Mehthod and device for measuring three-dimensional shapes Abandoned AU2001273360A1 (en)

Applications Claiming Priority (5)

Application Number Priority Date Filing Date Title
US09/660,809 2000-09-13
US09/660,809 US6639684B1 (en) 2000-09-13 2000-09-13 Digitizer using intensity gradient to image features of three-dimensional objects
US09/839,755 US6856407B2 (en) 2000-09-13 2001-04-19 Method for depth detection in 3D imaging providing a depth measurement for each unitary group of pixels
US09/839,755 2001-04-19
PCT/US2001/021838 WO2002023918A2 (en) 2000-09-13 2001-07-11 Mehthod and device for measuring three-dimensional shapes

Publications (1)

Publication Number Publication Date
AU2001273360A1 true AU2001273360A1 (en) 2002-03-26

Family

ID=27098192

Family Applications (1)

Application Number Title Priority Date Filing Date
AU2001273360A Abandoned AU2001273360A1 (en) 2000-09-13 2001-07-11 Mehthod and device for measuring three-dimensional shapes

Country Status (8)

Country Link
US (2) US6856407B2 (en)
EP (1) EP1317651A2 (en)
JP (1) JP2004509396A (en)
CN (1) CN1474932A (en)
AU (1) AU2001273360A1 (en)
CA (1) CA2422263A1 (en)
TW (1) TW531688B (en)
WO (1) WO2002023918A2 (en)

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7337093B2 (en) * 2001-09-07 2008-02-26 Purdue Research Foundation Systems and methods for collaborative shape and design
AU2002332918A1 (en) * 2001-09-07 2003-03-24 Abhishek Kumar Agrawal Systems and methods for collaborative shape design
US20040184653A1 (en) * 2003-03-20 2004-09-23 Baer Richard L. Optical inspection system, illumination apparatus and method for use in imaging specular objects based on illumination gradients
EP1782228A1 (en) * 2004-08-03 2007-05-09 Silverbrook Research Pty. Ltd Walk-up printing
CN100394141C (en) * 2004-12-28 2008-06-11 陈胜勇 Method and equipment for realizes structured light in high performance based on uniqueness in field
EP2521587B1 (en) * 2010-01-08 2020-04-08 Wake Forest University Health Sciences Delivery system
US8742309B2 (en) 2011-01-28 2014-06-03 Aptina Imaging Corporation Imagers with depth sensing capabilities
AU2012225644B2 (en) * 2011-03-07 2017-05-04 Wake Forest University Health Sciences Delivery system
US10015471B2 (en) 2011-08-12 2018-07-03 Semiconductor Components Industries, Llc Asymmetric angular response pixels for single sensor stereo
US8786730B2 (en) 2011-08-18 2014-07-22 Microsoft Corporation Image exposure using exclusion regions
US9554115B2 (en) * 2012-02-27 2017-01-24 Semiconductor Components Industries, Llc Imaging pixels with depth sensing capabilities
US8855404B2 (en) 2012-08-27 2014-10-07 The Boeing Company Methods and systems for inspecting a workpiece
TWI498526B (en) * 2013-06-05 2015-09-01 Nat Univ Chung Cheng Environment depth measurement method and its image acquisition device
WO2015015718A1 (en) * 2013-07-31 2015-02-05 パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ Sensor assembly
US9291447B2 (en) 2014-07-09 2016-03-22 Mitutoyo Corporation Method for controlling motion of a coordinate measuring machine
US9578311B2 (en) 2014-10-22 2017-02-21 Microsoft Technology Licensing, Llc Time of flight depth camera
US10742961B2 (en) * 2015-09-02 2020-08-11 Industrial Technology Research Institute Depth sensing apparatus with self-calibration and self-calibration method thereof
TWI620926B (en) 2016-11-04 2018-04-11 財團法人工業技術研究院 Workpiece surface detection method and system using the same
US10929994B2 (en) * 2016-12-07 2021-02-23 Electronics And Telecommunications Research Institute Image processing device configured to generate depth map and method of operating the same
US10122997B1 (en) 2017-05-03 2018-11-06 Lowe's Companies, Inc. Automated matrix photo framing using range camera input
US10885622B2 (en) 2018-06-29 2021-01-05 Photogauge, Inc. System and method for using images from a commodity camera for object scanning, reverse engineering, metrology, assembly, and analysis

Family Cites Families (67)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US3636250A (en) 1964-02-26 1972-01-18 Andrew V Haeff Apparatus for scanning and reproducing a three-dimensional representation of an object
US4089608A (en) 1976-10-21 1978-05-16 Hoadley Howard W Non-contact digital contour generator
US4490617A (en) * 1979-11-26 1984-12-25 European Electronic Systems Limited Optical width measuring system using two cameras
US4590608A (en) 1980-05-30 1986-05-20 The United States Of America As Represented By The Secretary Of The Army Topographic feature extraction using sensor array system
US4443705A (en) * 1982-10-01 1984-04-17 Robotic Vision Systems, Inc. Method for locating points on a three-dimensional surface using light intensity variations
US4564295A (en) 1983-03-07 1986-01-14 New York Institute Of Technology Apparatus and method for projection moire topography
US4641972A (en) 1984-09-14 1987-02-10 New York Institute Of Technology Method and apparatus for surface profilometry
US4657394A (en) 1984-09-14 1987-04-14 New York Institute Of Technology Apparatus and method for obtaining three dimensional surface contours
US4724525A (en) 1984-12-12 1988-02-09 Moore Special Tool Co., Inc. Real-time data collection apparatus for use in multi-axis measuring machine
US4705401A (en) 1985-08-12 1987-11-10 Cyberware Laboratory Inc. Rapid three-dimensional surface digitizer
US4737032A (en) 1985-08-26 1988-04-12 Cyberware Laboratory, Inc. Surface mensuration sensor
JPH0615968B2 (en) 1986-08-11 1994-03-02 伍良 松本 Three-dimensional shape measuring device
US4846577A (en) 1987-04-30 1989-07-11 Lbp Partnership Optical means for making measurements of surface contours
GB8716369D0 (en) 1987-07-10 1987-08-19 Travis A R L Three-dimensional display device
US5315512A (en) 1989-09-01 1994-05-24 Montefiore Medical Center Apparatus and method for generating image representations of a body utilizing an ultrasonic imaging subsystem and a three-dimensional digitizer subsystem
DE3941144C2 (en) 1989-12-13 1994-01-13 Zeiss Carl Fa Coordinate measuring device for the contactless measurement of an object
US5067817A (en) 1990-02-08 1991-11-26 Bauer Associates, Inc. Method and device for noncontacting self-referencing measurement of surface curvature and profile
DE4007500A1 (en) 1990-03-09 1991-09-12 Zeiss Carl Fa METHOD AND DEVICE FOR CONTACTLESS MEASUREMENT OF OBJECT SURFACES
JP2892430B2 (en) 1990-03-28 1999-05-17 株式会社日立製作所 Method and apparatus for displaying physical quantity
EP0488987B1 (en) 1990-11-26 1996-01-31 Michael Dr. Truppe Method for representing moving bodies
US5131844A (en) 1991-04-08 1992-07-21 Foster-Miller, Inc. Contact digitizer, particularly for dental applications
US5231470A (en) 1991-09-06 1993-07-27 Koch Stephen K Scanning system for three-dimensional object digitizing
US5377011A (en) 1991-09-06 1994-12-27 Koch; Stephen K. Scanning system for three-dimensional object digitizing
DE4134546A1 (en) 1991-09-26 1993-04-08 Steinbichler Hans METHOD AND DEVICE FOR DETERMINING THE ABSOLUTE COORDINATES OF AN OBJECT
US5175601A (en) 1991-10-15 1992-12-29 Electro-Optical Information Systems High-speed 3-D surface measurement surface inspection and reverse-CAD system
DE4304276A1 (en) 1992-02-17 1993-08-19 Galram Technology Ind Ltd Forming high resolution image of planar or three=dimensional object - combining sharp image data provided by detector matrix for successive scanning of object via optical imaging system.
US5216817A (en) 1992-03-18 1993-06-08 Colgate-Palmolive Company Digitizer measuring system
US5636025A (en) 1992-04-23 1997-06-03 Medar, Inc. System for optically measuring the surface contour of a part using more fringe techniques
US5432622A (en) 1992-05-29 1995-07-11 Johnston; Gregory E. High-resolution scanning apparatus
US5307292A (en) 1992-06-24 1994-04-26 Christopher A. Brown Method of quantifying the topographic structure of a surface
AT399647B (en) 1992-07-31 1995-06-26 Truppe Michael ARRANGEMENT FOR DISPLAYING THE INTERIOR OF BODIES
JP2651093B2 (en) * 1992-10-27 1997-09-10 松下電工株式会社 Shape detection method and device
US5337149A (en) 1992-11-12 1994-08-09 Kozah Ghassan F Computerized three dimensional data acquisition apparatus and method
US5414647A (en) 1992-11-23 1995-05-09 Ford Motor Company Non-contact method and system for building CAD models by integrating high density data scans
US5611147A (en) 1993-02-23 1997-03-18 Faro Technologies, Inc. Three dimensional coordinate measuring apparatus
DE4313860A1 (en) * 1993-04-28 1994-11-03 Ralf Lampalzer Optical sensor for shape recognition of three-dimensional objects
JP3057960B2 (en) 1993-06-22 2000-07-04 トヨタ自動車株式会社 Evaluation equipment for three-dimensional workpieces
US5999641A (en) 1993-11-18 1999-12-07 The Duck Corporation System for manipulating digitized image objects in three dimensions
US5661667A (en) 1994-03-14 1997-08-26 Virtek Vision Corp. 3D imaging using a laser projector
US5471303A (en) 1994-04-29 1995-11-28 Wyko Corporation Combination of white-light scanning and phase-shifting interferometry for surface profile measurements
GB9413870D0 (en) 1994-07-09 1994-08-31 Vision 1 Int Ltd Digitally-networked active-vision camera
US5531520A (en) 1994-09-01 1996-07-02 Massachusetts Institute Of Technology System and method of registration of three-dimensional data sets including anatomical body data
US5747822A (en) 1994-10-26 1998-05-05 Georgia Tech Research Corporation Method and apparatus for optically digitizing a three-dimensional object
US5621529A (en) * 1995-04-05 1997-04-15 Intelligent Automation Systems, Inc. Apparatus and method for projecting laser pattern with reduced speckle noise
US5617645A (en) 1995-05-02 1997-04-08 William R. W. Wick Non-contact precision measurement system
IL114278A (en) 1995-06-22 2010-06-16 Microsoft Internat Holdings B Camera and method
US6057909A (en) 1995-06-22 2000-05-02 3Dv Systems Ltd. Optical ranging camera
JPH0969978A (en) 1995-08-30 1997-03-11 Sanyo Electric Co Ltd Image pickup device
JP4031841B2 (en) 1995-09-05 2008-01-09 株式会社コダックデジタルプロダクトセンター Digital camera
US5689446A (en) 1995-11-03 1997-11-18 Amfit, Inc. Foot contour digitizer
US5799082A (en) 1995-11-07 1998-08-25 Trimble Navigation Limited Secure authentication of images
US5646733A (en) 1996-01-29 1997-07-08 Medar, Inc. Scanning phase measuring method and system for an object at a vision station
US6044170A (en) 1996-03-21 2000-03-28 Real-Time Geometry Corporation System and method for rapid shape digitizing and adaptive mesh generation
US5771310A (en) 1996-12-30 1998-06-23 Shriners Hospitals For Children Method and apparatus for recording three-dimensional topographies
US5988862A (en) 1996-04-24 1999-11-23 Cyra Technologies, Inc. Integrated system for quickly and accurately imaging and modeling three dimensional objects
US5870220A (en) 1996-07-12 1999-02-09 Real-Time Geometry Corporation Portable 3-D scanning system and method for rapid shape digitizing and adaptive mesh generation
AU722769B2 (en) 1996-08-23 2000-08-10 Her Majesty The Queen In Right Of Canada As Represented By The Department Of Agriculture And Agri-Food Canada Method and apparatus for using image analysis to determine meat and carcass characteristics
US5831621A (en) 1996-10-21 1998-11-03 The Trustees Of The University Of Pennyslvania Positional space solution to the next best view problem
US5864640A (en) 1996-10-25 1999-01-26 Wavework, Inc. Method and apparatus for optically scanning three dimensional objects using color information in trackable patches
US5805289A (en) 1997-07-07 1998-09-08 General Electric Company Portable measurement system using image and point measurement devices
IL121267A0 (en) 1997-07-09 1998-01-04 Yeda Res & Dev Method and device for determining the profile of an object
US5910845A (en) 1997-12-02 1999-06-08 Brown; Thomas Mattingly Peripheral viewing optical scanner for three dimensional surface measurement
US6141753A (en) 1998-02-10 2000-10-31 Fraunhofer Gesellschaft Secure distribution of digital representations
AU3991799A (en) * 1998-05-14 1999-11-29 Metacreations Corporation Structured-light, triangulation-based three-dimensional digitizer
US6040910A (en) * 1998-05-20 2000-03-21 The Penn State Research Foundation Optical phase-shift triangulation technique (PST) for non-contact surface profiling
CA2306515A1 (en) 2000-04-25 2001-10-25 Inspeck Inc. Internet stereo vision, 3d digitizing, and motion capture camera
EP1325624A2 (en) 2000-09-13 2003-07-09 Nextengine, Inc. Imaging system monitored or controlled to ensure fidelity of file captured

Also Published As

Publication number Publication date
US20020033955A1 (en) 2002-03-21
JP2004509396A (en) 2004-03-25
CN1474932A (en) 2004-02-11
US6856407B2 (en) 2005-02-15
WO2002023918A2 (en) 2002-03-21
WO2002023918A3 (en) 2003-02-06
US20040239949A1 (en) 2004-12-02
TW531688B (en) 2003-05-11
CA2422263A1 (en) 2002-03-21
EP1317651A2 (en) 2003-06-11

Similar Documents

Publication Publication Date Title
US6639684B1 (en) Digitizer using intensity gradient to image features of three-dimensional objects
AU2001273360A1 (en) Mehthod and device for measuring three-dimensional shapes
US7995834B1 (en) Multiple laser scanner
US7551771B2 (en) Methods, systems, and computer program products for acquiring three-dimensional range information
US9879975B2 (en) Method for optically measuring three-dimensional coordinates and calibration of a three-dimensional measuring device
CN100430752C (en) Position measuring device
US6832044B2 (en) Iris imaging apparatus
US6392744B1 (en) Range measurement system
US9170097B2 (en) Hybrid system
US20020191076A1 (en) Iris imaging apparatus
EP1331807A1 (en) Iris imaging device
JP2000188712A5 (en)
JP2006514739A5 (en)
CN108718406A (en) A kind of varifocal 3D depth cameras and its imaging method
US11350077B2 (en) Handheld three dimensional scanner with an autoaperture
TWI385361B (en) A test apparatus having a plurality of combinations of object distances and a test method
KR100232256B1 (en) Variable range measurment apparatus
JP4129789B2 (en) Three-dimensional distance imaging apparatus and three-dimensional distance measuring method
JP3504698B2 (en) Distance measuring device
JPH09287927A (en) Projecting-point moving type range finder
JP2002318994A (en) Optical information reader
Pipitone et al. Moving-correlation-code triangulation range imaging
JPH0652196B2 (en) Infrared thermal imager
JPH04324808A (en) Distance measuring device
JPH02211766A (en) Image scanner

Legal Events

Date Code Title Description
MK5 Application lapsed section 142(2)(e) - patent request and compl. specification not accepted