US20020145103A1 - System, method, and progam product for acquiring accurate object silhouettes for shape recovery - Google Patents

System, method, and progam product for acquiring accurate object silhouettes for shape recovery Download PDF

Info

Publication number
US20020145103A1
US20020145103A1 US09/825,685 US82568501A US2002145103A1 US 20020145103 A1 US20020145103 A1 US 20020145103A1 US 82568501 A US82568501 A US 82568501A US 2002145103 A1 US2002145103 A1 US 2002145103A1
Authority
US
United States
Prior art keywords
shadow
image
translucent panel
light source
digital images
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US09/825,685
Other versions
US6455835B1 (en
Inventor
Fausto Bernardini
Henning Biermann
Holly Rushmeier
Silvio Savarese
Gabriel Taubin
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
LinkedIn Corp
Original Assignee
International Business Machines Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by International Business Machines Corp filed Critical International Business Machines Corp
Priority to US09/825,685 priority Critical patent/US6455835B1/en
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TAUBIN, GABRIEL, RUSHMEIER, HOLLY E., SAVARESE, SILVIO, BIERMANN, HENNING, BERNARDI, FAUSTO
Assigned to INTERNATIONAL BUSINESS MACHINES CORPORATION reassignment INTERNATIONAL BUSINESS MACHINES CORPORATION CORRECTIVE ASSIGNMENT TO CORRECT THE NAME OF THE ASSIGNOR. FILED ON APRIL 4, 2001, RECORDED ON REEL 11709 FRAME 0671. Assignors: BERNARDINI, FAUSTO, BIERMANN, HENNING, SAVARESE, SILVIO, TAUBIN, GABRIEL, RUSHMEIER, HOLLY E.
Application granted granted Critical
Publication of US6455835B1 publication Critical patent/US6455835B1/en
Publication of US20020145103A1 publication Critical patent/US20020145103A1/en
Assigned to LINKEDIN CORPORATION reassignment LINKEDIN CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: INTERNATIONAL BUSINESS MACHINES CORPORATION
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/24Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
    • G01B11/2433Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures for measuring outlines by shadow casting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/564Depth or shape recovery from multiple images from contours

Definitions

  • the teachings of this invention relate generally to computer vision and computer graphics and, more specifically, the teachings of this invention relate to techniques for acquiring silhouettes from an image.
  • the teachings herein address the problem of acquiring a numerical description of the shape of an object. Given a numerical description of the object's shape it is possible, using well-known computer graphics algorithms, to generate images of the object from different points of view and under different lighting conditions.
  • One important application of such synthetic imagery is in e-commerce, where the seller of an object allows potential customers to inspect a virtual copy of an object interactively using a computer.
  • Numerical representations of objects can be used for other purposes, such as in CAD (computer-aided design) systems as a starting point for the design of new objects.
  • a class of popular methods for acquiring a numerical representation of an object's shape is known as shape from silhouette, also referred to by similar names such as shape from occluding contour or shape from boundaries.
  • Shape from silhouette algorithms use an image of an object captured by a camera, or any other imaging device. Using the known position of the camera, and the silhouette of the object in the image (i.e. the curve that marks the boundary in the image between the object and the background), an estimate of the numerical shape can be made. A very crude estimate of shape can be obtained from a single image. An improved estimate is obtained using a number of silhouettes from images of the object in different positions relative to the camera.
  • An alternative class of algorithms for extracting shape from silhouettes uses the variation of contour shape in successive images.
  • An example is described by Zheng (Jiang Yu Zheng, “Acquiring 3-D Models from Sequences of Contours”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 16, No. 2, February 1994, pp. 163-178.)
  • Zheng Jiang Yu Zheng, “Acquiring 3-D Models from Sequences of Contours”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 16, No. 2, February 1994, pp. 163-178.
  • many silhouette images are obtained as the object is rotated in front of the camera.
  • An estimate of 3D location of points on the object's surface is obtained from the location of silhouettes in the image relative to the projection of the axis of rotation, and the rate of change of these positions with respect to angular change.
  • a basic operation required by either class of the shape from silhouette algorithms is the accurate extraction of the boundary between the object and the background.
  • This is an example of the classic image segmentation problem from the field of image processing.
  • Systems for extracting shape attempt to simplify the segmentation by designing a suitable backdrop.
  • An example of such a design is illustrated in Jones and Oakely (M. Jones and J. P. Oakley, “Efficient representation of object shape for silhouette intersection”, IEEE Proc.-Vis. Image Signal Process, Vol. 142, No. 6, December 1995, pp. 359-364.)
  • the backdrop for the object is painted a uniform color ( in the case of Jones and Oakely “Chromakey Blue”).
  • the silhouette is defined as the boundary of the image regions that are the uniform background color.
  • An alternative approach uses a large flat diffuse light source in place of the colored backdrop.
  • the silhouette is defined as the boundary of the bright image regions, with the object itself generally appearing dark.
  • Shape from silhouettes is a popular technique because it can be implemented inexpensively.
  • the major cost of the system resides in the camera and in a mechanism to control the position of the object, such as a turntable.
  • the implementation with volume carving is particularly attractive for applications because the method guarantees a closed surface.
  • An alternative and related method for capturing object shape is “shape from shadows”, as described in U.S. Pat. Nos: 4,792,696 and 4,873,651. These methods are similar to shape from silhouettes, since a sharp shadow is the silhouette projected from a point light source.
  • the camera is placed on the same side of the object as the direction of light incident on the object, and images are taken of the shadows cast by the object.
  • the surface is a height field. That is, the object sits on a reference plane with locations on the plane specified by (x,y) Cartesian coordinates.
  • the shape of the object is given by a third coordinate z that is descriptive of the height of the object surface above the reference plane. With this assumption, the shape of the object surface is inferred from where shadows begin and end, and from knowledge of the light source direction.
  • U.S. Pat. No.: 4,604,807 employs a shadow that is observed using a camera on the opposite side of the object from the light source.
  • the shadow is formed by pressing a relatively flat object, e.g., a person's foot, onto a translucent panel. The shadow is observed from the opposite side to obtain a numerical description of the two dimensional area of the foot, and is not used to estimate the three dimensional shape of the foot.
  • a number of other problems are encountered with the prior art techniques for finding object silhouettes.
  • Both of the backdrop approaches allow only one silhouette to be obtained for each position of the object.
  • a device with one degree of freedom to provide accurate positioning such as a turntable
  • one position of the object on the turntable may not be adequate to obtain a view of the entire object surface.
  • the object is placed once, a series of images is obtained for one rotation of the device.
  • the object is placed in a different position relative to the turntable, and another series is obtained. This process may need to be repeated many times, and the geometries recovered by each rotation must be registered to one another by an additional geometric processing step.
  • the system described by Leibe et al. requires the object to be scanned to sit on a fixed translucent surface. Although the shape of some objects can be estimated from a sparse set of views spanning the full space of directions around the object, the system described by Leibe et al. is limited to shadows that can be cast from light sources above the translucent surface.
  • the goal of the Leibe et al. system is to produce crude shape representations only, and the design does not permit the calibrated repositioning of an object, nor does it include a way to obtain additional information, such as shape from photometric data, to improve the estimate of shape and to include concavities.
  • the system includes a side camera above the translucent surface, but obtaining silhouettes from this camera presents all of the problems of traditional silhouette extraction, and cannot, for example, be used for shiny objects.
  • apparatus for obtaining the silhouette of an object in a form suitable for use by a shape from silhouette algorithm for obtaining a numerical description of the object's three dimensional shape. Also disclosed are methods for processing the output of the apparatus into a numerical description of the object that is suitable for interactive display on a computer graphics system.
  • At least one light source preferably a point light source
  • An imaging device referred to for convenience as a camera
  • the silhouette or shadow contour is obtained from the image of the shadow as the region of the shadow that is substantially darker than the region outside of the shadow. This is true for any opaque object regardless of its surface finish or shape.
  • the quantity of light reflected by the object in the direction of the translucent panel is orders of magnitude smaller than light that impinges on the panel directly from the point source, thereby enhancing the contrast between the object's shadow and the illumination from the light source.
  • a further benefit obtained by the use of the point light source is that the object need not be in contact with the translucent panel to obtain a shadow having sharp edges.
  • the full object silhouette is obtained since nothing (including the object itself) is in the path between the camera and the translucent panel.
  • the full silhouette obtained can be processed by any suitable shape from silhouette algorithm, and thus the to be imaged are not limited in topological type.
  • a color image of the object can optionally be obtained simultaneously with the shadow image by using another camera, such as a color camera, that is placed on the same side of the object as the light source.
  • another camera such as a color camera
  • multiple silhouettes can be captured for one object position, reducing the number of rotations needed on a turntable system, and reducing the post-processing needed to register geometries obtained from multiple different positions.
  • a system and method for obtaining a three dimensional image of an object.
  • the method includes the steps of (a) shining light from at least one light source on to the object from a first direction to create a first shadow cast by the object on a first surface of a translucent panel, where the object is disposed between a light source and the first surface of the translucent panel and has a first pose; (b) obtaining a first digital image of the first shadow from a second, opposite surface of the translucent panel; (c) changing the pose of the object and obtaining additional digital images of additional shadows cast by the object for different object poses; and (d) processing the first and the additional digital images to create a three dimensional image of the object.
  • the step of processing preferably employs a space carving process.
  • the step of processing operates to identify a boundary of the image of the shadow in each of the first and additional digital images, where the boundary is identified in a given one of the digital images by applying a pixel thresholding process to determine whether a given pixel is located within the image of the shadow or outside of the image of the shadow.
  • the step of processing further defines a virtual volume as a list of volume elements, projects individual ones of the volume elements onto the plane of the image of the shadow, and retains only those volume elements in the list that lie within the image of the shadow or on the identified boundary.
  • the step of processing then further applies an isosurface extraction algorithm to the list of surviving volume elements to obtain a numerical description of the shape of the surface of the object.
  • the step of shining light on to the object can also be done from a second, or third, or fourth, etc., direction to create an additional shadow or shadows cast by the object on the first surface of the translucent panel.
  • the resulting shadow image(s) are processed in the same manner as the first shadow.
  • a plurality of light sources each having a different color can be used, as can array of light sources that are operated in sequence.
  • a single light source may be translated with respect to the object to shine light on the object from a plurality of different directions.
  • the method may include additional steps of obtaining a digital image of the object for each object pose; processing the digital images of the object to derive surface normals and color maps; and applying the surface normals and color maps to the surface of the three dimensional image of the object.
  • FIG. 1 depicts a presently preferred embodiment of a system for obtaining the shape of an object by observing silhouettes of the object..
  • FIGS. 2 a and 2 b are diagrams depicting ideal properties of the light source and the translucent panel shown in FIG. 1, wherein FIG. 2 a illustrates an ideal light scattering distribution for the translucent panel, and FIG. 2 b shows an ideal light emission distribution for the light source.
  • FIG. 3 shows an exemplary shadow image produced by the acquisition system of FIG. 1.
  • FIG. 4 is a logic flow diagram of the processing of the shadow image of FIG. 3.
  • FIG. 5 shows an image of a contour found in the shadow image after processing.
  • FIG. 6 is a logic flow diagram of the processing of the images with contours into a shape approximation in the form of a set of volume elements, also referred to as boxes.
  • FIG. 7 is a logic flow diagram of the processing of the set of boxes computed in accordance with FIG. 6 into a surface.
  • FIG. 8 is a block diagram of a second embodiment of the image acquisition system in accordance with these teachings, wherein three color light sources are used in lieu of the single light source of FIG. 1.
  • FIG. 9 is a block diagram of a third embodiment of the image acquisition system, wherein an array of light sources replaces the single light source of FIG. 1.
  • FIG. 10 is a logic flow diagram of the image processing associated with the output from the third embodiment of this invention shown in FIG. 9, wherein detail and color are added to the surface.
  • FIG. 1 A presently preferred first embodiment of image acquisition apparatus is depicted in FIG. 1.
  • a point light source 140 is placed in front of an object 120 that is to be imaged, thereby casting a shadow 125 on a translucent panel 100 .
  • the light source has a diameter of about 0.025 H, and is located a distance about 5 H from the object 120 .
  • the light source 140 has a nearly uniform intensity output 300 in the direction of the object 120 , as is diagramed in FIG. 2 b.
  • the translucent panel 100 is preferably a thin sheet of partially light transmissive material, for example a less than 1 mm thick sheet of diffusely transmitting material.
  • the panel 100 is thin to eliminate significant scattering in the plane of the panel 100 , to thereby avoid blurring of the image of the object's shadow 125 , and has a forward scattering distribution 210 that is nearly uniform for light 200 incident on the panel 100 .
  • the translucent panel 100 is preferably non-colored or color neutral.
  • a sheet of white writing paper with no water marking may be used, with the sheet of paper being sandwiched between thin (3 mm or less) plates of transparent glass for support.
  • Other types of translucent panels may also be used, such as a sheet of certain polymer materials, frosted glass, and other materials that are only partially transmissive to impinging light.
  • the translucent panel 100 is located a distance of about 2H from the object 120 , and has dimensions of at least about 1.5H by 1.5H.
  • the object 120 to be measured is placed on a device that has a calibrated position.
  • this device is embodied as a turntable 130 which is controlled by a computer 150 .
  • a camera 110 (a black and white, or a color camera) is placed behind the translucent panel 100 , and is preferably also controlled by the computer 150 (although manual control of the turntable and/or camera could be used as well.)
  • the camera 110 has a 32 degree field of view (wider angles are preferably avoided to eliminate potential distortion effects in the camera optics), and is located a distance of about 5H from the second, rear surface of the translucent panel 100 .
  • the positions of the camera 110 , translucent panel 100 and the light source 140 are calibrated with respect to a coordinate system defined on the turntable 130 (or other positioning device) in its initial position. Any well-known calibration or measurement techniques for obtaining camera parameters and measuring object locations may be used. Assuming that the positions are suitably calibrated, the object 120 need not be located at the center of the turntable 130 , the light source 140 need not lie on the optical axis of the camera 110 , and the optical axis of the camera 110 need not be perpendicular to the plane of the translucent plate 100 .
  • the light from source 140 is incident on the front of the object 120 (i.e. light source 140 is in front of the object 120 , or the direction of light from the source 140 , if behind the object 120 , is redirected to be incident from the front of the object 120 by the use of a mirror or mirrors), (b) the object 120 is in front of the translucent panel 100 , and the panel 100 is in front of the camera 110 .
  • the object 120 For each rotation increment of the turntable 130 the object 120 , and hence its shadow 125 , assumes a different pose with respect to the image plane of the camera 112 .
  • the rotation increment of the turntable 130 may be a function of the surface complexity of the object 120 , as the more complex is the surface shape the more shadow images will be required to capture the surface shape. That is, the rotation increment of the turntable 130 may be larger when the object 120 is a coffee cup as compared to the rotation increment when the object 120 is a decorative vase.
  • the rotation increment of the turntable 130 may be about 30 degrees.
  • An image is taken by the camera 110 with respect to each pose of the object 120 .
  • the images that are acquired by the system such as the exemplary object shadow image 350 shown in FIG. 3, are processed using the method shown in FIG. 4.
  • each pixel is identified as being inside or outside the shadow in process 410 .
  • Any suitable pixel thresholding analysis may be used in process 410 , such as the well-known k-means algorithm for unsupervised identification of clusters of values.
  • the boundary of the shadow 125 is then found in process 420 with, preferably, sub-pixel accuracy using any image edge detector, such as the well-known Sobel edge detector.
  • the exemplary shadow image 450 in FIG. 5 shows the results of processing image 350 with the method shown in FIG. 4.
  • any suitable method may be employed for obtaining an estimated shape from silhouettes may be used to estimate the object shape from the derived object contours, such as the contour shown in the image 450 .
  • the preferred embodiment shown in FIG. 6 uses a volume carving approach.
  • a virtual array of volume elements (such as, but not limited to, boxes) of dimension h ⁇ k ⁇ l are defined, where h,k and l are 0.01 H or less, in the coordinate system defined on the turntable 130 , such that the extent of the array encompasses the full object 120 .
  • Initially all vertices on all volume elements are assigned a signed-distance value (i.e., negative for inside the object 120 , positive for outside the object 120 ) of ⁇ 0.01H.
  • the volume elements in list 520 are projected along lines emanating from the light source 140 position and ending on the plane of the translucent panel 100 using processes 530 .
  • a test 540 is performed to determine if the volume element (box) is projected into the shadow region. If the result of test 540 is no, another test 550 is performed to see if the box is projected on the boundary of the shadow region. If the result of test 550 is yes, a process 555 computes a new signed-distance that is assigned to each vertex of the volume element equal to the distance of the projection of the vertex to the shadow boundary. If the result of the test in process 550 is no, the volume element is marked “out” in step 560 , given a signed distance value of 0.01H, and is eliminated from list of volume elements for the processing of subsequent images.
  • FIG. 7 The further processing of the list of boxes (or volume elements) 600 is shown in FIG. 7.
  • the numerical description of the object shape 620 is extracted by using any well-known isosurface algorithm 610 to find the surface that passes through the volume at signed-distance values of zero.
  • FIG. 8 shows a second embodiment of the image acquisition system, wherein components that are also found in FIG. 1 are numbered accordingly.
  • a plurality of radiation sources in this case three sources 142 , 144 and 146 ), each with a narrow, but not necessarily visible, spectral distribution are used in place of the single point light source 140 .
  • point lights with red 142 , green 144 and blue 146 filters are used.
  • the sources 142 , 144 and 146 are arranged in this embodiment in a triangular shape, with each light source being placed at a vertex of the triangle, and separated from adjacent sources by about 2H.
  • the size of the light sources is again 0.025H, and the plane of the triangularly-disposed light sources is located about 5H from the object 120 .
  • the camera 112 that is used is capable of sensing radiation in each of the spectral bands. For point light sources with visible red 142 , green 144 and blue 146 filters a commodity digital camera 112 can be used. Each time the turntable 130 is moved a color image is obtained, with three separate shadows for the red, green and blue sources.
  • the N images are processed as before (i.e., as in FIGS. 4, 6 and 7 ), with a total of 3N images being processed, and with each of the color images being separated into three grey-scale images.
  • FIG. 9 shows a third embodiment of the system, wherein components that are also found in FIG. 1 are numbered accordingly.
  • the light sources 161 , 162 , 163 , 164 , 165 , 166 , 167 , 168 , 169 are mounted on a frame 180 , with a distance 2H between adjacent light sources, and the plane of the array 160 of light sources is located about 5H from the object 120 .
  • a color camera 170 is placed in front of the object 120 adjacent to light source 169 (i.e., at about the center of the array 160 .
  • both cameras acquire an image.
  • a series of M shadow images and M color images are thus obtained for each position of the turntable 130 .
  • the shadow images are processed as before (i.e., as in FIGS. 4, 6 and 7 ).
  • This embodiment thus uses photometric stereo techniques to obtain additional shape information, and assumes the use of the visible spectrum of light.
  • the array of M light sources 160 can be replaced by a single point source and a mechanism to translate the point light source to different known positions.
  • person or a machine may move a single light with a tracking system and record the light position each time an image is acquired.
  • a single light source can also be made to impinge on the object 120 from many different directions by reflecting against a mirror that is controlled to move into a series of known positions.
  • step 630 The processing of the M color images obtained by the camera 170 is shown in FIG. 10.
  • the M color images are processed by means of a photometric stereo technique, preferably one described in Rushmeier et al. “Acquiring Input for Rendering at Appropriate Levels of Detail: Digitizing a Pieta”, Proceedings of the 9th Eurographics Rendering Workshop, Vienna, Austria, June 1998, and in Rushmeier and Bernardini, “Computing Consistent Normals and Colors from Photometric Data”, Proceedings of 3DIM ‘99, Ottawa, Canada, October, 1999, incorporated by reference herein, to produce detailed maps of color and surface normals for the object 120 .
  • step 640 the color and surface normal maps are projected on to the estimated shape of the object 120 and combined into a single non-redundant map of normals and colors by the methods described in Bernardini et al., “High-Quality Texture Synthesis from Multiple Scans”, IBM Research Division Report, RC21656, Feb. 1, 2000, incorporated by reference herein.
  • the result is a model 650 that contains of a numerical description of shape and a map of detailed colors and normals.
  • the resultant model 650 is suitable for display using software available on most commodity personal computers.

Abstract

Disclosed are methods and apparatus for obtaining the shape of an object by observing silhouettes of the object. At least one point light source is placed in front of the object, thereby casting a shadow of the object on a translucent panel that is placed behind the object. A camera, or other imaging device, captures an image of the shadow from behind the translucent panel. The object's full silhouette is obtained from the image of the shadow as the region of the shadow is substantially darker than the region outside of the shadow. The full silhouette thus obtained may be processed by any suitable shape from silhouette algorithm, and thus objects are not limited in topological type. A color image of the object can optionally be obtained simultaneously with the shadow image using a camera placed on the same side of the object as the light source. Multiple silhouettes can be captured for one object position, reducing the number of rotations needed on a turntable system, and reducing the post-processing needed to register geometries obtained from multiple different positions.

Description

    FIELD OF THE INVENTION
  • The teachings of this invention relate generally to computer vision and computer graphics and, more specifically, the teachings of this invention relate to techniques for acquiring silhouettes from an image. [0001]
  • BACKGROUND OF THE INVENTION
  • A number of different techniques have been developed to compute shapes from silhouettes or contours in the field of computer imaging. [0002]
  • The teachings herein address the problem of acquiring a numerical description of the shape of an object. Given a numerical description of the object's shape it is possible, using well-known computer graphics algorithms, to generate images of the object from different points of view and under different lighting conditions. One important application of such synthetic imagery is in e-commerce, where the seller of an object allows potential customers to inspect a virtual copy of an object interactively using a computer. Numerical representations of objects can be used for other purposes, such as in CAD (computer-aided design) systems as a starting point for the design of new objects. [0003]
  • A class of popular methods for acquiring a numerical representation of an object's shape is known as shape from silhouette, also referred to by similar names such as shape from occluding contour or shape from boundaries. Shape from silhouette algorithms use an image of an object captured by a camera, or any other imaging device. Using the known position of the camera, and the silhouette of the object in the image (i.e. the curve that marks the boundary in the image between the object and the background), an estimate of the numerical shape can be made. A very crude estimate of shape can be obtained from a single image. An improved estimate is obtained using a number of silhouettes from images of the object in different positions relative to the camera. [0004]
  • Many algorithms have been devised to compute a numerical description of the three dimensional shape of an object from silhouettes. One class of algorithms is known as volumetric or space carving, as originally described by Martin and Aggrawal (Worthy N. Martin and J. K Agrawal, “Volumetric Descriptions of Objects from Multiple Views”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. PAMI-5, No. 2, March 1983, pp. 150-158.) In this technique a volume of small boxes is numerically defined that completely encloses the object. For each image the boxes are projected onto an image plane. If the projection of a box falls outside of the object silhouette, it is marked as “outside” and is eliminated from a current estimate of the object shape. As each silhouette image is considered more of the boxes are eliminated, or “carved away” from the initial volume. The boxes remaining after all of the silhouette images have been examined is the estimate of the object's shape. A smooth representation of the surface of the object can then be obtained by any well-known isosurface algorithm. [0005]
  • An alternative class of algorithms for extracting shape from silhouettes uses the variation of contour shape in successive images. An example is described by Zheng (Jiang Yu Zheng, “Acquiring 3-D Models from Sequences of Contours”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 16, No. 2, February 1994, pp. 163-178.) In this method, many silhouette images are obtained as the object is rotated in front of the camera. An estimate of 3D location of points on the object's surface is obtained from the location of silhouettes in the image relative to the projection of the axis of rotation, and the rate of change of these positions with respect to angular change. [0006]
  • There are fundamental limitations on the accuracy of the shape that can be recovered by shape from silhouettes, as discussed by Laurentini (Aldo Laurentini, “How Far 3D Shapes Can Be Understood from 2D Silhouettes”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 17, No. 2, February 1995, pp. 188-195.). For example, object concavities will not appear in silhouettes, and so will not be captured. To provide the illusion of concavities, and to add color to the model, capture systems generally acquire color images of the object from known camera positions. These color images can be related to the captured geometry by the well-known computer graphics technique known as projective texture mapping. Geometries (generally in the form of triangular meshes) with texture maps can be displayed with hardware and software available on typical personal computers. [0007]
  • A basic operation required by either class of the shape from silhouette algorithms is the accurate extraction of the boundary between the object and the background. This is an example of the classic image segmentation problem from the field of image processing. Systems for extracting shape attempt to simplify the segmentation by designing a suitable backdrop. An example of such a design is illustrated in Jones and Oakely (M. Jones and J. P. Oakley, “Efficient representation of object shape for silhouette intersection”, IEEE Proc.-Vis. Image Signal Process, Vol. 142, No. 6, December 1995, pp. 359-364.) The backdrop for the object is painted a uniform color ( in the case of Jones and Oakely “Chromakey Blue”). The silhouette is defined as the boundary of the image regions that are the uniform background color. [0008]
  • An alternative approach uses a large flat diffuse light source in place of the colored backdrop. The silhouette is defined as the boundary of the bright image regions, with the object itself generally appearing dark. [0009]
  • Shape from silhouettes, particularly with the addition of color textures, is a popular technique because it can be implemented inexpensively. The major cost of the system resides in the camera and in a mechanism to control the position of the object, such as a turntable. The implementation with volume carving is particularly attractive for applications because the method guarantees a closed surface. [0010]
  • An alternative and related method for capturing object shape is “shape from shadows”, as described in U.S. Pat. Nos: 4,792,696 and 4,873,651. These methods are similar to shape from silhouettes, since a sharp shadow is the silhouette projected from a point light source. In both of these patents the camera is placed on the same side of the object as the direction of light incident on the object, and images are taken of the shadows cast by the object. In both of these patents it is assumed that the surface is a height field. That is, the object sits on a reference plane with locations on the plane specified by (x,y) Cartesian coordinates. The shape of the object is given by a third coordinate z that is descriptive of the height of the object surface above the reference plane. With this assumption, the shape of the object surface is inferred from where shadows begin and end, and from knowledge of the light source direction. [0011]
  • U.S. Pat. No.: 4,604,807 employs a shadow that is observed using a camera on the opposite side of the object from the light source. In this patent the shadow is formed by pressing a relatively flat object, e.g., a person's foot, onto a translucent panel. The shadow is observed from the opposite side to obtain a numerical description of the two dimensional area of the foot, and is not used to estimate the three dimensional shape of the foot. [0012]
  • In an article by Leibe et al. (B. Leibe, T. Starner, W. Ribarsky, Z. Wartell, D. Krum, J. Weeks, B. Singletary and L. Godges, “Toward Spontaneous Interaction with the Perceptive Workbench”, IEEE Computer Graphics and Applications, November/December 2000, pp. 54-65.) a system is described that observes shadows cast by objects on a translucent table with a camera located underneath the table. The system can produce only a crude estimate of shape, because the object cannot be repositioned in a calibrated manner. [0013]
  • All of the prior art techniques known to the inventors assume that an accurate silhouette can be extracted from the image. However, if an accurate silhouette cannot be extracted, then the shape of the object will be inaccurate. [0014]
  • The segmentation approach fails if the object is shiny, transparent, or is same color as the background. Segmentation can also fail even with the use of a large diffused light source. [0015]
  • A number of other problems are encountered with the prior art techniques for finding object silhouettes. First consider the approach of using a background of known color. The silhouette is detected where the backdrop color ends in the image. This method fails for glossy objects that reflect some of the background color in the direction of the camera, and for objects which transmit light. This method also fails when camera characteristics cause “bleeding” of color from one region of the image to another. The method can also fail if inter-reflections on the object cast color from the background onto the object. The method also fails if the object happens to be the same color as the backdrop. [0016]
  • Some methods attempt to avoid these problems by taking an image of the backdrop alone and then an image of the object in front of the background, and then taking the difference between the two images. However, this approach fails for very shiny objects. It also fails when any shadow is cast by the object onto the backdrop. [0017]
  • The approach of using a large diffuse light source seeks to avoid the problem of the object possibly being the same color as the background. However, this technique also fails for shiny surfaces, light transmitting surfaces, and for surfaces in which self-interreflections transmit light from the backdrop onto the object. This approach also prevents the simultaneous acquisition of color images to be used as texture maps, since the bright background causes most of the object to appear very dark in the image. Having to acquire the color images separately extends the length of time required to obtain the numerical description of the object. [0018]
  • Both of the backdrop approaches allow only one silhouette to be obtained for each position of the object. For simple systems employing a device with one degree of freedom to provide accurate positioning, such as a turntable, one position of the object on the turntable may not be adequate to obtain a view of the entire object surface. The object is placed once, a series of images is obtained for one rotation of the device. The object is placed in a different position relative to the turntable, and another series is obtained. This process may need to be repeated many times, and the geometries recovered by each rotation must be registered to one another by an additional geometric processing step. [0019]
  • The methods that employ shadows have been in part motivated by the problem of segmentation from the backdrop when shiny objects are being scanned. However, for the shadow methods, with the camera in the same direction as the direction of incident light, the problem remains of separating the image of the object and the image of its shadow. Such segmentation is difficult for objects with a dark or partially dark surface, and is impossible for black objects. The shadow methods are also limited by the height field assumption for 3-D shape recovery. Objects with even moderately complex topologies, e.g., a coffee mug with a handle, cannot be measured with such techniques without substantial error. [0020]
  • The method described in U.S. Pat. No.: 4,604,807 employs optics and geometry that require that the object being measured rest against the translucent panel, and that the object shape is almost flat. The apparatus can only measure 2-D areas, and cannot be used to capture silhouettes of objects of arbitrary shape for 3-D shape recovery. [0021]
  • The system described by Leibe et al. requires the object to be scanned to sit on a fixed translucent surface. Although the shape of some objects can be estimated from a sparse set of views spanning the full space of directions around the object, the system described by Leibe et al. is limited to shadows that can be cast from light sources above the translucent surface. The goal of the Leibe et al. system is to produce crude shape representations only, and the design does not permit the calibrated repositioning of an object, nor does it include a way to obtain additional information, such as shape from photometric data, to improve the estimate of shape and to include concavities. The system includes a side camera above the translucent surface, but obtaining silhouettes from this camera presents all of the problems of traditional silhouette extraction, and cannot, for example, be used for shiny objects. [0022]
  • OBJECTS AND ADVANTAGES OF THE INVENTION
  • It is a first object and advantage of this invention to provide an improved system and method to obtain 3-D shapes from one or more images. [0023]
  • It is a further object and advantage of this invention to provide a system and method for deriving the surface shape of an object from shadow images of the object obtained from behind a translucent panel that is interposed between an image capture device, referred to for convenience as a camera, and the object, where the object is interposed between the front of the translucent panel and one or more point light sources. [0024]
  • SUMMARY OF THE INVENTION
  • The foregoing and other problems are overcome and the foregoing objects and advantages are realized by methods and apparatus in accordance with embodiments of this invention. [0025]
  • Disclosed herein are embodiments of apparatus for obtaining the silhouette of an object in a form suitable for use by a shape from silhouette algorithm for obtaining a numerical description of the object's three dimensional shape. Also disclosed are methods for processing the output of the apparatus into a numerical description of the object that is suitable for interactive display on a computer graphics system. [0026]
  • More particularly, disclosed herein are methods and apparatus for obtaining the shape of an object by observing silhouettes of the object. At least one light source, preferably a point light source, is placed in front of the object, thereby casting a shadow of the object on a translucent panel that is placed behind the object. An imaging device, referred to for convenience as a camera, captures an image of the shadow from behind the translucent panel. The silhouette or shadow contour is obtained from the image of the shadow as the region of the shadow that is substantially darker than the region outside of the shadow. This is true for any opaque object regardless of its surface finish or shape. By using a point light source, rather than a large diffuse light source, the quantity of light reflected by the object in the direction of the translucent panel is orders of magnitude smaller than light that impinges on the panel directly from the point source, thereby enhancing the contrast between the object's shadow and the illumination from the light source. A further benefit obtained by the use of the point light source is that the object need not be in contact with the translucent panel to obtain a shadow having sharp edges. The full object silhouette is obtained since nothing (including the object itself) is in the path between the camera and the translucent panel. The full silhouette obtained can be processed by any suitable shape from silhouette algorithm, and thus the to be imaged are not limited in topological type. Unlike systems with large diffuse lights as backgrounds, which make the object appear black, a color image of the object can optionally be obtained simultaneously with the shadow image by using another camera, such as a color camera, that is placed on the same side of the object as the light source. Unlike conventional silhouette systems, multiple silhouettes can be captured for one object position, reducing the number of rotations needed on a turntable system, and reducing the post-processing needed to register geometries obtained from multiple different positions. [0027]
  • In accordance with the teachings herein, a system and method is disclosed for obtaining a three dimensional image of an object. The method includes the steps of (a) shining light from at least one light source on to the object from a first direction to create a first shadow cast by the object on a first surface of a translucent panel, where the object is disposed between a light source and the first surface of the translucent panel and has a first pose; (b) obtaining a first digital image of the first shadow from a second, opposite surface of the translucent panel; (c) changing the pose of the object and obtaining additional digital images of additional shadows cast by the object for different object poses; and (d) processing the first and the additional digital images to create a three dimensional image of the object. The step of processing preferably employs a space carving process. The step of processing operates to identify a boundary of the image of the shadow in each of the first and additional digital images, where the boundary is identified in a given one of the digital images by applying a pixel thresholding process to determine whether a given pixel is located within the image of the shadow or outside of the image of the shadow. The step of processing further defines a virtual volume as a list of volume elements, projects individual ones of the volume elements onto the plane of the image of the shadow, and retains only those volume elements in the list that lie within the image of the shadow or on the identified boundary. The step of processing then further applies an isosurface extraction algorithm to the list of surviving volume elements to obtain a numerical description of the shape of the surface of the object. [0028]
  • The step of shining light on to the object can also be done from a second, or third, or fourth, etc., direction to create an additional shadow or shadows cast by the object on the first surface of the translucent panel. The resulting shadow image(s) are processed in the same manner as the first shadow. A plurality of light sources each having a different color can be used, as can array of light sources that are operated in sequence. A single light source may be translated with respect to the object to shine light on the object from a plurality of different directions. [0029]
  • Further in accordance with these teachings the method may include additional steps of obtaining a digital image of the object for each object pose; processing the digital images of the object to derive surface normals and color maps; and applying the surface normals and color maps to the surface of the three dimensional image of the object.[0030]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The above set forth and other features of the invention are made more apparent in the ensuing Detailed Description of the Invention when read in conjunction with the attached Drawings, wherein: [0031]
  • FIG. 1 depicts a presently preferred embodiment of a system for obtaining the shape of an object by observing silhouettes of the object.. [0032]
  • FIGS. 2[0033] a and 2 b are diagrams depicting ideal properties of the light source and the translucent panel shown in FIG. 1, wherein FIG. 2a illustrates an ideal light scattering distribution for the translucent panel, and FIG. 2b shows an ideal light emission distribution for the light source.
  • FIG. 3 shows an exemplary shadow image produced by the acquisition system of FIG. 1. [0034]
  • FIG. 4 is a logic flow diagram of the processing of the shadow image of FIG. 3. [0035]
  • FIG. 5 shows an image of a contour found in the shadow image after processing. [0036]
  • FIG. 6 is a logic flow diagram of the processing of the images with contours into a shape approximation in the form of a set of volume elements, also referred to as boxes. [0037]
  • FIG. 7 is a logic flow diagram of the processing of the set of boxes computed in accordance with FIG. 6 into a surface. [0038]
  • FIG. 8 is a block diagram of a second embodiment of the image acquisition system in accordance with these teachings, wherein three color light sources are used in lieu of the single light source of FIG. 1. [0039]
  • FIG. 9 is a block diagram of a third embodiment of the image acquisition system, wherein an array of light sources replaces the single light source of FIG. 1. [0040]
  • FIG. 10 is a logic flow diagram of the image processing associated with the output from the third embodiment of this invention shown in FIG. 9, wherein detail and color are added to the surface.[0041]
  • DETAILED DESCRIPTION OF THE INVENTION
  • A presently preferred first embodiment of image acquisition apparatus is depicted in FIG. 1. A point [0042] light source 140 is placed in front of an object 120 that is to be imaged, thereby casting a shadow 125 on a translucent panel 100. In the preferred embodiment, to measure an object 120 of maximum linear dimension H, the light source has a diameter of about 0.025 H, and is located a distance about 5 H from the object 120. The light source 140 has a nearly uniform intensity output 300 in the direction of the object 120, as is diagramed in FIG. 2b. Referring also to FIG. 2a, the translucent panel 100 is preferably a thin sheet of partially light transmissive material, for example a less than 1 mm thick sheet of diffusely transmitting material. The panel 100 is thin to eliminate significant scattering in the plane of the panel 100, to thereby avoid blurring of the image of the object's shadow 125, and has a forward scattering distribution 210 that is nearly uniform for light 200 incident on the panel 100. The translucent panel 100 is preferably non-colored or color neutral. A sheet of white writing paper with no water marking may be used, with the sheet of paper being sandwiched between thin (3 mm or less) plates of transparent glass for support. Other types of translucent panels may also be used, such as a sheet of certain polymer materials, frosted glass, and other materials that are only partially transmissive to impinging light. The translucent panel 100 is located a distance of about 2H from the object 120, and has dimensions of at least about 1.5H by 1.5H.
  • The [0043] object 120 to be measured is placed on a device that has a calibrated position. In FIG. 1 this device is embodied as a turntable 130 which is controlled by a computer 150. A camera 110 (a black and white, or a color camera) is placed behind the translucent panel 100, and is preferably also controlled by the computer 150 (although manual control of the turntable and/or camera could be used as well.) In the preferred embodiment the camera 110 has a 32 degree field of view (wider angles are preferably avoided to eliminate potential distortion effects in the camera optics), and is located a distance of about 5H from the second, rear surface of the translucent panel 100.
  • The positions of the [0044] camera 110, translucent panel 100 and the light source 140 are calibrated with respect to a coordinate system defined on the turntable 130 (or other positioning device) in its initial position. Any well-known calibration or measurement techniques for obtaining camera parameters and measuring object locations may be used. Assuming that the positions are suitably calibrated, the object 120 need not be located at the center of the turntable 130, the light source 140 need not lie on the optical axis of the camera 110, and the optical axis of the camera 110 need not be perpendicular to the plane of the translucent plate 100.
  • What is important to the operation of the imaging system is that: (a) the light from [0045] source 140 is incident on the front of the object 120 (i.e. light source 140 is in front of the object 120, or the direction of light from the source 140, if behind the object 120, is redirected to be incident from the front of the object 120 by the use of a mirror or mirrors), (b) the object 120 is in front of the translucent panel 100, and the panel 100 is in front of the camera 110.
  • For each rotation increment of the [0046] turntable 130 the object 120, and hence its shadow 125, assumes a different pose with respect to the image plane of the camera 112. The rotation increment of the turntable 130, and hence the number of poses attained by the object 120, may be a function of the surface complexity of the object 120, as the more complex is the surface shape the more shadow images will be required to capture the surface shape. That is, the rotation increment of the turntable 130 may be larger when the object 120 is a coffee cup as compared to the rotation increment when the object 120 is a decorative vase.
  • As an example, if the [0047] object 120 is a coffee mug with a handle, the rotation increment of the turntable 130 may be about 30 degrees.
  • An image is taken by the [0048] camera 110 with respect to each pose of the object 120. The images that are acquired by the system, such as the exemplary object shadow image 350 shown in FIG. 3, are processed using the method shown in FIG. 4. In a loop 400 for each shadow image, each pixel is identified as being inside or outside the shadow in process 410. Any suitable pixel thresholding analysis may be used in process 410, such as the well-known k-means algorithm for unsupervised identification of clusters of values. The boundary of the shadow 125 is then found in process 420 with, preferably, sub-pixel accuracy using any image edge detector, such as the well-known Sobel edge detector. The exemplary shadow image 450 in FIG. 5 shows the results of processing image 350 with the method shown in FIG. 4.
  • Any suitable method may be employed for obtaining an estimated shape from silhouettes may be used to estimate the object shape from the derived object contours, such as the contour shown in the image [0049] 450. The preferred embodiment shown in FIG. 6 uses a volume carving approach. In step 500 a virtual array of volume elements (such as, but not limited to, boxes) of dimension h×k×l are defined, where h,k and l are 0.01 H or less, in the coordinate system defined on the turntable 130, such that the extent of the array encompasses the full object 120. Initially all vertices on all volume elements are assigned a signed-distance value (i.e., negative for inside the object 120, positive for outside the object 120) of −0.01H. This indicates initially that all vertices are inside the object 120. For the loop 510 over each image acquired, the volume elements in list 520 are projected along lines emanating from the light source 140 position and ending on the plane of the translucent panel 100 using processes 530. A test 540 is performed to determine if the volume element (box) is projected into the shadow region. If the result of test 540 is no, another test 550 is performed to see if the box is projected on the boundary of the shadow region. If the result of test 550 is yes, a process 555 computes a new signed-distance that is assigned to each vertex of the volume element equal to the distance of the projection of the vertex to the shadow boundary. If the result of the test in process 550 is no, the volume element is marked “out” in step 560, given a signed distance value of 0.01H, and is eliminated from list of volume elements for the processing of subsequent images.
  • The further processing of the list of boxes (or volume elements) [0050] 600 is shown in FIG. 7. The numerical description of the object shape 620 is extracted by using any well-known isosurface algorithm 610 to find the surface that passes through the volume at signed-distance values of zero.
  • FIG. 8 shows a second embodiment of the image acquisition system, wherein components that are also found in FIG. 1 are numbered accordingly. A plurality of radiation sources (in this case three [0051] sources 142, 144 and 146), each with a narrow, but not necessarily visible, spectral distribution are used in place of the single point light source 140. In the preferred system, point lights with red 142, green 144 and blue 146 filters are used. The sources 142, 144 and 146 are arranged in this embodiment in a triangular shape, with each light source being placed at a vertex of the triangle, and separated from adjacent sources by about 2H. The size of the light sources is again 0.025H, and the plane of the triangularly-disposed light sources is located about 5H from the object 120. The camera 112 that is used is capable of sensing radiation in each of the spectral bands. For point light sources with visible red 142, green 144 and blue 146 filters a commodity digital camera 112 can be used. Each time the turntable 130 is moved a color image is obtained, with three separate shadows for the red, green and blue sources. The N images are processed as before (i.e., as in FIGS. 4, 6 and 7), with a total of 3N images being processed, and with each of the color images being separated into three grey-scale images.
  • FIG. 9 shows a third embodiment of the system, wherein components that are also found in FIG. 1 are numbered accordingly. In this embodiment an [0052] array 160 of M light sources (in this case M=9) is used in place of the single light source 140. The light sources 161,162,163,164,165,166,167, 168, 169 are mounted on a frame 180, with a distance 2H between adjacent light sources, and the plane of the array 160 of light sources is located about 5H from the object 120. A color camera 170 is placed in front of the object 120 adjacent to light source 169 (i.e., at about the center of the array 160. As each light source (161, 162, 163, 164, 165, 166, 167, 168, 169) is illuminated in turn, both cameras (170 and 110) acquire an image. A series of M shadow images and M color images are thus obtained for each position of the turntable 130. The shadow images are processed as before (i.e., as in FIGS. 4, 6 and 7). This embodiment thus uses photometric stereo techniques to obtain additional shape information, and assumes the use of the visible spectrum of light.
  • Alternatively, the array of M [0053] light sources 160 can be replaced by a single point source and a mechanism to translate the point light source to different known positions. For example, person or a machine may move a single light with a tracking system and record the light position each time an image is acquired. A single light source can also be made to impinge on the object 120 from many different directions by reflecting against a mirror that is controlled to move into a series of known positions.
  • The processing of the M color images obtained by the [0054] camera 170 is shown in FIG. 10. Using the numerical surface description 620 obtained in FIG. 7, in step 630 the M color images are processed by means of a photometric stereo technique, preferably one described in Rushmeier et al. “Acquiring Input for Rendering at Appropriate Levels of Detail: Digitizing a Pieta”, Proceedings of the 9th Eurographics Rendering Workshop, Vienna, Austria, June 1998, and in Rushmeier and Bernardini, “Computing Consistent Normals and Colors from Photometric Data”, Proceedings of 3DIM ‘99, Ottawa, Canada, October, 1999, incorporated by reference herein, to produce detailed maps of color and surface normals for the object 120. In step 640 the color and surface normal maps are projected on to the estimated shape of the object 120 and combined into a single non-redundant map of normals and colors by the methods described in Bernardini et al., “High-Quality Texture Synthesis from Multiple Scans”, IBM Research Division Report, RC21656, Feb. 1, 2000, incorporated by reference herein. The result is a model 650 that contains of a numerical description of shape and a map of detailed colors and normals. The resultant model 650 is suitable for display using software available on most commodity personal computers.
  • Other techniques for deriving surface color and normals maps could be employed as well. [0055]
  • Note should be made that the [0056] color camera 170, and the associated processing shown in FIG. 10, could be incorporated as well into the system embodiments shown in FIGS. 1 and 8.
  • While the invention has been particularly shown and described with respect to preferred embodiments thereof, it will be understood by those skilled in the art that changes in form and details may be made therein without departing from the scope and spirit of the invention. [0057]

Claims (28)

What is claimed is:
1. A method of obtaining a three dimensional image of an object, comprising the steps of:
shining light from at least one light source on to the object from a first direction to create a first shadow cast by the object on a first surface of a translucent panel, where the object is disposed between a light source and the first surface of the translucent panel and has a first pose;
obtaining a first digital image of the first shadow from a second, opposite surface of the translucent panel;
changing the pose of the object and obtaining additional digital images of additional shadows cast by the object for different object poses; and
processing the first and the additional digital images to create a three dimensional image of the object.
2. A method as in claim 1, wherein the step of processing employs a space carving process.
3. A method as in claim 1, wherein the step of processing identifies a boundary of the image of the shadow in each of the first and additional digital images.
4. A method as in claim 3, wherein the boundary is identified in a given one of the digital images by applying a pixel thresholding process to determine whether a given pixel is located within the image of the shadow or outside of the image of the shadow.
5. A method as in claim 3, wherein the step of processing defines a virtual volume as a list of volume elements, projects individual ones of the volume elements onto the plane of the image of the shadow, and retains only those volume elements in the list that lie within the image of the shadow or on the identified boundary.
6. A method as in claim 5, wherein the step of processing further comprises a step of applying an isosurface extraction algorithm to the list of surviving volume elements to obtain a numerical description of the shape of the surface of the object.
7. A method as in claim 1, and further comprising steps of shining light on to the object from a second direction to create an additional shadow cast by the object on the first surface of the translucent panel, and processing an image of the additional shadow in the same manner as the first shadow.
8. A method as in claim 1, wherein the step of shining light operates a plurality of light sources each having a different color.
9. A method as in claim 1, wherein the step of shining operates individual ones of a plurality of light sources in turn to shine light on the object from a plurality of different directions.
10. A method as in claim 1, wherein the step of shining translates a single light source with respect to the object to shine light on the object from a plurality of different directions.
11. A method as in claim 1, and further comprising steps of:
for each object pose, obtaining a digital image of the object;
processing the digital images of the object to derive surface normals and color maps; and
applying the surface normals and color maps to the surface of the three dimensional image of the object.
12. A method as in claim 1, wherein the object has a maximum linear dimension of H and is located about 2H from the first surface of the translucent panel, wherein the light source is located about 5H from the object, and wherein the image is obtained with a camera located about 5H from the second surface of the translucent panel.
13. A method as in claim 12, wherein the translucent panel has dimensions of at least about 1.5 H.
14. A method as in claim 12, wherein the light source has a size of about 0.025 H.
15. A system for obtaining a three dimensional image of an object, comprising:
a translucent panel;
a mechanism for changing the pose of the object;
at least one light source for illuminating the object from a first direction to create a first shadow cast by the object on a first surface of said translucent panel, where the object is disposed between said at least one light source and said first surface of said translucent panel;
a camera for obtaining a first digital image of the first shadow from a second, opposite surface of said translucent panel; and
a controller, responsive to said first digital image and to additional digital images of additional shadows cast by the object for different object poses, for processing the first and the additional digital images to create a three dimensional image of the object.
16. A system as in claim 15, wherein said controller employs a space carving process.
17. A system as in claim 15, wherein said controller processes the first and the additional digital images to identify a boundary of the image of the shadow in each of the first and additional digital images.
18. A system as in claim 17, wherein the boundary is identified in a given one of the digital images by applying a pixel thresholding process to determine whether a given pixel is located within the image of the shadow or outside of the image of the shadow.
19. A system as in claim 17, wherein said controller processes the first and the additional digital images to define a virtual volume as a list of volume elements, to project individual ones of the volume elements onto the plane of the image of the shadow, and to retain only those volume elements in the list that lie within the image of the shadow or on the identified boundary.
20. A system as in claim 19, wherein said controller further processes the first and the additional digital images by applying an isosurface extraction algorithm to the list of surviving volume elements to obtain a numerical description of the shape of the surface of the object.
21. A system as in claim 15, wherein there are a plurality of light sources for illuminating the object from a plurality of directions for creating additional shadows cast by the object on said first surface of said translucent panel, and wherein said controller processes images of the additional shadows in the same manner as the first shadow.
22. A system as in claim 21, wherein said plurality of light sources each have a different color.
23. A system as in claim 21, wherein individual ones of said plurality of light sources are operated in turn to illuminate the object from a plurality of different directions.
24. A system as in claim 15, and further comprising means for translating said light source with respect to the object to illuminate the object from a plurality of different directions
25. A system as in claim 15, and further comprising a second camera, said controller being responsive to each object pose for obtaining a digital image of the object with the second camera; for processing the digital images of the object to derive surface normals and color maps; and for applying the surface normals and color maps to the surface of the three dimensional image of the object.
26. A system as in claim 15, wherein the object has a maximum linear dimension of H and is located about 2H from said first surface of said translucent panel, wherein said light source is located about 5H from the object, and wherein said camera is located about 5H from said second surface of said translucent panel.
27. A system as in claim 26, wherein said translucent panel has dimensions of at least about 1.5 H.
28. A system as in claim 26, wherein said light source has a size of about 0.025 H.
US09/825,685 2001-04-04 2001-04-04 System, method, and program product for acquiring accurate object silhouettes for shape recovery Expired - Lifetime US6455835B1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US09/825,685 US6455835B1 (en) 2001-04-04 2001-04-04 System, method, and program product for acquiring accurate object silhouettes for shape recovery

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US09/825,685 US6455835B1 (en) 2001-04-04 2001-04-04 System, method, and program product for acquiring accurate object silhouettes for shape recovery

Publications (2)

Publication Number Publication Date
US6455835B1 US6455835B1 (en) 2002-09-24
US20020145103A1 true US20020145103A1 (en) 2002-10-10

Family

ID=25244672

Family Applications (1)

Application Number Title Priority Date Filing Date
US09/825,685 Expired - Lifetime US6455835B1 (en) 2001-04-04 2001-04-04 System, method, and program product for acquiring accurate object silhouettes for shape recovery

Country Status (1)

Country Link
US (1) US6455835B1 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1413851A1 (en) 2002-10-25 2004-04-28 General Electric Company Space carving for measurement of high-curvature areas on a mechanical part
US20050099506A1 (en) * 2003-11-10 2005-05-12 Microsoft Corporation System and method for obtaining accurate image content
US20050135706A1 (en) * 2003-12-19 2005-06-23 General Electric Company System and method for shadow imaging
US20050212820A1 (en) * 2004-03-26 2005-09-29 Ross Video Limited Method, system, and device for automatic determination of nominal backing color and a range thereof
US20060001760A1 (en) * 2004-06-23 2006-01-05 Canon Technology Europe Ltd. Apparatus and method for object shape detection
US20070237908A1 (en) * 2006-02-28 2007-10-11 Alexander Garnett Methods of making artistic sculpture and sculptures so made
WO2009156027A1 (en) * 2008-06-25 2009-12-30 Khs Ag Method and device for detecting the edge profile of bottles or similar containers
US20100220184A1 (en) * 2005-06-20 2010-09-02 Robert Meredith Appleyard Imaging and safety system and method for an industrial machine
FR2951563A1 (en) * 2009-10-16 2011-04-22 Logiparts Sa Object's geometrical parameter e.g. volume, analyzing method, involves capturing object illuminated at backlight to obtain image with silhouette of object, and exploiting silhouette to obtain geometrical parameters of object
WO2011131770A1 (en) * 2010-04-21 2011-10-27 Accenture Global Services Limited Method and system for analyzing geometric parameters of an object
DE102012005966A1 (en) * 2012-03-23 2013-09-26 S.E.A. - Science & Engineering Applications Datentechnik GmbH Device for generating representation of three dimensional body in form of vehicle, has illumination device for generating light strip, particularly parallel light, where electronic sensor is arranged behind or in front of shadow wall
EP2660556A1 (en) * 2012-05-02 2013-11-06 proventavia GmbH Method for recording outlines of objects
JP2015118023A (en) * 2013-12-19 2015-06-25 富士通株式会社 Normal vector extraction device, normal vector extraction method, and normal vector extraction program
US9600927B1 (en) * 2012-10-21 2017-03-21 Google Inc. Systems and methods for capturing aspects of objects using images and shadowing
JP2018189623A (en) * 2017-05-08 2018-11-29 国立大学法人横浜国立大学 Three-dimensional shape estimation system, three-dimensional shape estimation device, method for estimating three-dimensional shape, and program

Families Citing this family (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6903738B2 (en) * 2002-06-17 2005-06-07 Mitsubishi Electric Research Laboratories, Inc. Image-based 3D modeling rendering system
US6831641B2 (en) * 2002-06-17 2004-12-14 Mitsubishi Electric Research Labs, Inc. Modeling and rendering of surface reflectance fields of 3D objects
GB2398469B (en) * 2003-02-12 2005-10-26 Canon Europa Nv Image processing apparatus
US7084386B2 (en) * 2003-05-02 2006-08-01 International Business Machines Corporation System and method for light source calibration
GB0319677D0 (en) * 2003-08-21 2003-09-24 Canon Res Ct Europ Ltd Photographic apparatus for use in creating three-dimensional models
US20050110801A1 (en) * 2003-11-20 2005-05-26 I-Jong Lin Methods and systems for processing displayed images
DE102004015785B4 (en) * 2004-03-25 2012-06-06 Sikora Ag Method for determining the dimension of a cross-section of a flat cable or a sector conductor
US8186280B2 (en) * 2005-02-16 2012-05-29 Dean Stoops Load bearing robotic turntable
US7555163B2 (en) * 2004-12-16 2009-06-30 Sony Corporation Systems and methods for representing signed distance functions
US7623250B2 (en) * 2005-02-04 2009-11-24 Stryker Leibinger Gmbh & Co. Kg. Enhanced shape characterization device and method
US7478601B2 (en) * 2005-02-16 2009-01-20 Dean Stoops Robotic turntable
US8094964B2 (en) * 2007-10-26 2012-01-10 Adobe Systems Incorporated Methods and systems for estimating illumination source characteristics from a single image
JP2010176380A (en) 2009-01-29 2010-08-12 Sony Corp Information processing device and method, program, and recording medium
US8643717B2 (en) * 2009-03-04 2014-02-04 Hand Held Products, Inc. System and method for measuring irregular objects with a single camera
DE102009039657A1 (en) * 2009-09-02 2011-03-10 Msg Maschinenbau Gmbh Apparatus and method for measuring the shape of an object
US9410801B2 (en) * 2011-03-17 2016-08-09 Cadscan Limited Scanner
US9575013B2 (en) * 2011-05-17 2017-02-21 Gii Acquisition, Llc Non-contact method and system for inspecting a manufactured part at an inspection station having a measurement axis
KR20130050685A (en) * 2011-11-08 2013-05-16 한국전자통신연구원 Apparatus and method for measuring a 3-dimensional object shape
US9111349B2 (en) * 2011-12-16 2015-08-18 Microsoft Technology Licensing, Llc Object identification using 3-D curve matching
US8436853B1 (en) * 2012-07-20 2013-05-07 Google Inc. Methods and systems for acquiring and ranking image sets
KR20170034727A (en) * 2015-09-21 2017-03-29 삼성전자주식회사 Shadow information storing method and apparatus, 3d rendering method and apparatus
US20170188010A1 (en) * 2015-12-29 2017-06-29 Canon Kabushiki Kaisha Reconstruction of local curvature and surface shape for specular objects
CN105866736B (en) * 2016-04-05 2018-07-06 华中科技大学 A kind of indoor orientation method based on light pattern
US20180330018A1 (en) * 2017-05-12 2018-11-15 The Boeing Company Methods and systems for part geometry extraction
KR102367640B1 (en) * 2017-07-14 2022-02-24 캐파시티 인크. Systems and methods for the creation and display of interactive 3D representations of real objects
CN109460766B (en) * 2018-09-29 2022-03-15 济南企财通软件有限公司 Device and method for extracting ticket image from ticket paper
US10909713B2 (en) * 2018-10-25 2021-02-02 Datalogic Usa, Inc. System and method for item location, delineation, and measurement
EP3745081B1 (en) * 2019-05-28 2023-03-22 Tecan Trading Ag Position detector and method for 3d positioning
CN111210473A (en) * 2020-01-21 2020-05-29 上海悦易网络信息技术有限公司 Mobile phone contour positioning method and equipment
WO2024054784A1 (en) 2022-09-08 2024-03-14 Digimarc Corporation Image analysis methods and arrangements

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4604807A (en) 1984-12-21 1986-08-12 Interco, Incorporated Electronic foot measuring apparatus and method
US4873651A (en) 1987-04-21 1989-10-10 Case Western Reserve University Method and apparatus for reconstructing three-dimensional surfaces from two-dimensional images
US4792696A (en) 1987-06-05 1988-12-20 Trustees Of Columbia University In The City Of New York Method and an apparatus for determining surface shape utilizing object self-shadowing
US6289107B1 (en) * 1996-05-23 2001-09-11 Nike, Inc. Apparatus and method of measuring human extremities using peripheral illumination techniques

Cited By (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040079907A1 (en) * 2002-10-25 2004-04-29 General Electric Company Space carving for measurement of high-curvature areas on a mechanical part
US6969836B2 (en) * 2002-10-25 2005-11-29 General Electric Company Space carving for measurement of high-curvature areas on a mechanical part
EP1413851A1 (en) 2002-10-25 2004-04-28 General Electric Company Space carving for measurement of high-curvature areas on a mechanical part
US20050099506A1 (en) * 2003-11-10 2005-05-12 Microsoft Corporation System and method for obtaining accurate image content
US7348993B2 (en) 2003-11-10 2008-03-25 Microsoft Corporation System and method for obtaining accurate image content
US20050135706A1 (en) * 2003-12-19 2005-06-23 General Electric Company System and method for shadow imaging
US7508455B2 (en) * 2004-03-26 2009-03-24 Ross Video/Live Production Technology Method, system, and device for automatic determination of nominal backing color and a range thereof
US20050212820A1 (en) * 2004-03-26 2005-09-29 Ross Video Limited Method, system, and device for automatic determination of nominal backing color and a range thereof
US20060001760A1 (en) * 2004-06-23 2006-01-05 Canon Technology Europe Ltd. Apparatus and method for object shape detection
US8692877B2 (en) * 2005-06-20 2014-04-08 Lazer Safe Pty Ltd Imaging and safety system and method for an industrial machine
US9557007B2 (en) 2005-06-20 2017-01-31 Lazer Safe Pty Ltd Imaging and safety system and method for an industrial machine
US20100220184A1 (en) * 2005-06-20 2010-09-02 Robert Meredith Appleyard Imaging and safety system and method for an industrial machine
US7983787B2 (en) * 2006-02-28 2011-07-19 Alexander Garnett Methods of making artistic sculpture and sculptures so made
US20070237908A1 (en) * 2006-02-28 2007-10-11 Alexander Garnett Methods of making artistic sculpture and sculptures so made
WO2009156027A1 (en) * 2008-06-25 2009-12-30 Khs Ag Method and device for detecting the edge profile of bottles or similar containers
US9909861B2 (en) 2008-06-25 2018-03-06 Khs Gmbh Method and device for detecting the edge profile of bottles or similar containers
FR2951564A1 (en) * 2009-10-16 2011-04-22 Acceria METHOD AND INSTALLATION FOR ANALYZING GEOMETRIC PARAMETERS OF AN OBJECT
FR2951563A1 (en) * 2009-10-16 2011-04-22 Logiparts Sa Object's geometrical parameter e.g. volume, analyzing method, involves capturing object illuminated at backlight to obtain image with silhouette of object, and exploiting silhouette to obtain geometrical parameters of object
US9273954B2 (en) 2009-10-16 2016-03-01 Accenture Global Services Limited Method and system for analyzing geometric parameters of an object
AU2011244225B2 (en) * 2010-04-21 2015-04-23 Accenture Global Services Limited Method and system for analyzing geometric parameters of an object
WO2011131770A1 (en) * 2010-04-21 2011-10-27 Accenture Global Services Limited Method and system for analyzing geometric parameters of an object
DE102012005966A1 (en) * 2012-03-23 2013-09-26 S.E.A. - Science & Engineering Applications Datentechnik GmbH Device for generating representation of three dimensional body in form of vehicle, has illumination device for generating light strip, particularly parallel light, where electronic sensor is arranged behind or in front of shadow wall
DE102012005966B4 (en) 2012-03-23 2022-06-09 S.E.A. - Science & Engineering Applications Datentechnik GmbH Device and method for generating a two-dimensional representation of a three-dimensional body and lighting device for this
EP2660556A1 (en) * 2012-05-02 2013-11-06 proventavia GmbH Method for recording outlines of objects
US9600927B1 (en) * 2012-10-21 2017-03-21 Google Inc. Systems and methods for capturing aspects of objects using images and shadowing
JP2015118023A (en) * 2013-12-19 2015-06-25 富士通株式会社 Normal vector extraction device, normal vector extraction method, and normal vector extraction program
JP2018189623A (en) * 2017-05-08 2018-11-29 国立大学法人横浜国立大学 Three-dimensional shape estimation system, three-dimensional shape estimation device, method for estimating three-dimensional shape, and program

Also Published As

Publication number Publication date
US6455835B1 (en) 2002-09-24

Similar Documents

Publication Publication Date Title
US6455835B1 (en) System, method, and program product for acquiring accurate object silhouettes for shape recovery
US6792140B2 (en) Image-based 3D digitizer
US6831641B2 (en) Modeling and rendering of surface reflectance fields of 3D objects
US6791542B2 (en) Modeling 3D objects with opacity hulls
Sadlo et al. A practical structured light acquisition system for point-based geometry and texture
US6858826B2 (en) Method and apparatus for scanning three-dimensional objects
Tarini et al. 3D acquisition of mirroring objects using striped patterns
Hernández et al. Non-rigid photometric stereo with colored lights
Hertzmann et al. Example-based photometric stereo: Shape reconstruction with general, varying brdfs
US7098435B2 (en) Method and apparatus for scanning three-dimensional objects
Fechteler et al. Adaptive colour classification for structured light systems
EP2024707B1 (en) Scanner system and method for scanning
US20030231175A1 (en) Image-based 3D modeling rendering system
Brostow et al. Video normals from colored lights
US20020061130A1 (en) Image processing apparatus
Savarese et al. 3d reconstruction by shadow carving: Theory and practical evaluation
EP3261332A1 (en) System and method for image processing and generating a body model
US20110292179A1 (en) Imaging system and method
EP3382645B1 (en) Method for generation of a 3d model based on structure from motion and photometric stereo of 2d sparse images
US20200258300A1 (en) Method and apparatus for generating a 3d reconstruction of an object
Wang et al. Relief texture from specularities
JP4335589B2 (en) How to model a 3D object
Wang et al. Capturing and rendering geometry details for BTF-mapped surfaces
Palka et al. 3D object digitization devices in manufacturing engineering applications and services
Savarese Shape reconstruction from shadows and reflections

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:BERNARDI, FAUSTO;BIERMANN, HENNING;RUSHMEIER, HOLLY E.;AND OTHERS;REEL/FRAME:011709/0671;SIGNING DATES FROM 20010327 TO 20010403

AS Assignment

Owner name: INTERNATIONAL BUSINESS MACHINES CORPORATION, NEW Y

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE NAME OF THE ASSIGNOR. FILED ON APRIL 4, 2001, RECORDED ON REEL 11709 FRAME 0671;ASSIGNORS:BERNARDINI, FAUSTO;BIERMANN, HENNING;RUSHMEIER, HOLLY E.;AND OTHERS;REEL/FRAME:012064/0822;SIGNING DATES FROM 20010327 TO 20010403

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

REMI Maintenance fee reminder mailed
FPAY Fee payment

Year of fee payment: 12

SULP Surcharge for late payment

Year of fee payment: 11

AS Assignment

Owner name: LINKEDIN CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:INTERNATIONAL BUSINESS MACHINES CORPORATION;REEL/FRAME:035201/0479

Effective date: 20140331