WO2005122064A1 - A technique for estimating the pose of surface shapes using tripod operators - Google Patents

A technique for estimating the pose of surface shapes using tripod operators Download PDF

Info

Publication number
WO2005122064A1
WO2005122064A1 PCT/US2004/015580 US2004015580W WO2005122064A1 WO 2005122064 A1 WO2005122064 A1 WO 2005122064A1 US 2004015580 W US2004015580 W US 2004015580W WO 2005122064 A1 WO2005122064 A1 WO 2005122064A1
Authority
WO
WIPO (PCT)
Prior art keywords
pose
feature vector
unknown
tripod
image
Prior art date
Application number
PCT/US2004/015580
Other languages
French (fr)
Inventor
Frank Pipitone
Original Assignee
The United States Of America, As Represented By The Secretary Of The Navy
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by The United States Of America, As Represented By The Secretary Of The Navy filed Critical The United States Of America, As Represented By The Secretary Of The Navy
Priority to PCT/US2004/015580 priority Critical patent/WO2005122064A1/en
Publication of WO2005122064A1 publication Critical patent/WO2005122064A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • G06T7/74Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/60Type of objects
    • G06V20/64Three-dimensional objects

Definitions

  • This invention pertains generally to an improved technique for recognizing and locating objects and more specifically a technique for estimating the pose of an object from a range image containing the object.
  • range image being a two-dimensional array of numbers which gives the depth of a scene along many directions from a central point in the instrument. Instead of measuring the brightness of many points in a scene, as in a television camera, these instruments measure where each point is in a three-dimensional space.
  • range images and the more conventional intensity images from digital cameras have been used in the computer vision research community to dete ⁇ nine the pose of observed objects.
  • object as used herein, means a particular surface shape.
  • Pose means a complete description of an object's position and orientation.
  • Tripod operators are a versatile class of feature extraction operators for surfaces. They are useful for recognition and/or localization (pose estimation) based on range or tactile data.
  • a TO consists of three points in 3-space fixed at the vertices of a triangle and a procedure for making several "depth" measurements in the coordinate frame of the triangle, which is placed on the surface like a surveyor's tripod. TOs can be embedded in a vision system in many ways and applied to almost any surface shape.
  • a TO consists of three points in space fixed at the vertices of a triangle of fixed edge lengths and a procedure for making several depth measurements in the coordinate frame of the triangle, which is placed on the surface like a surveyor's tripod. These measurements take the form of arc-lengths along "probe curves" at which the surface is intersected.
  • Figure la through lc shows three examples of TO's.
  • Figure la shows a very simple TO with one line probe fixed symmetrically with respect to the rigid triangle ABC.
  • the single scalar feature is the distance from the plane of ABC at which the probe intersects the surface. This resembles a mechanical optician's tool called a spherometer.
  • a spherometer is the simplest possible tripod operator.
  • the probe is a sliding rod which contacts the lens at a depth approximately proportional to the curvature of the lens, and hence to its optical strength.
  • the number d of scalar features is called the order of the operator.
  • Figures lb and lc show TO's that can be viewed as a set of equilateral triangles hinged together so that all d+3 points can be made to contact a surface.
  • the angles of the d binges are the features.
  • This type, called linkable TO's is preferred because of their symmetry and uniform sensitivity to noise.
  • Many variations of these TO's could be constructed.
  • the N sampled surface points yield only N - 3 independent scalar features, and the order d is N -3.
  • These features contain all the surface shape information in the 3N components of the points since they suffice to reconstruct the relative positions of the N points. They contain no other information. For example, they have complete six DOF invariance under rigid motions, the group R 3 x SO(3). Thus, they depend upon where the tripod lies on the surface, but upon nothing else.
  • a key property is that for any dimensionality d of feature vector only a 3 (or fewer)-dimensional manifold of feature space points can be generated from a given surface, since the tripod can be moved in only 3 DOF on a surface.
  • This set is a kind of invariant signature. For brevity, this is called the signature of the object or surface, with respect to a particular type TO. It can be stored in an array of bins in feature space, e.g., of dimension 3 or 4, for later efficient access of near neighbors to TO features measured at recognition time. These bins can optionally contain precomputed probability densities, analytic expressions for distances to nearby signature manifolds, and partial or complete descriptions of the relative poses of tripods and models, all to serve various purposes in a recognition system.
  • Point A can be chosen as any point on the image surface. Interpolation is to be done locally as needed, e.g., using piecewise triangular facets. Point B can be found by moving along a line at orientation a in image coordinates, pixel indices, until the 3D distance
  • AB [ e.
  • an operator with a 3-fold symmetry produces signatures unchanged by cyclically permuting each triple of corresponding features.
  • the three 3-cycles (1, 2, 3), (4, 5, 6) and (7, 8, 9) show this property, for features ⁇ , through ⁇ 9 , respectively.
  • This allows a 3-fold storage reduction, e.g., by permuting the features so that ⁇ , is the largest.
  • the signature can be modified by a procedure that allows recognition of the "other side" of any surface already recognizable. This is called inversion of a signature.
  • This operator in the invention's preferred form, the linkable TO is essentially a 3D simulated structure consisting of several triangles hinged together. It is applied to a computer-represented surface, such as range image, by moving it until all of its points lie on that surface. Then the "hinge angles" provide information about the shape of the surface.
  • tripod operators there are two major steps. The first is training the system on a new object so that it will later be able for the system to recognize (or estimate the pose of) that object when seen again in some range image. The second step is the actual recognition (or, pose estimation).
  • the objective of this invention is to provide an improved technique for estimating the pose of surface shapes in six degrees of freedom from a range image containing an object possessing such a surface shape.
  • This and other objectives are accomplished by using a software procedure, with associated hardware, for estimating the pose of an object from a range image containing the object.
  • a range image is a two dimensional array of numbers which represent the distances from a reference point in the range imaging instrument to observed surface points in a scene. All six parameters of the pose of an object are estimated; three translational and three angular parameters.
  • TOs tripod operators
  • non-pose-distinctive placement removal a new technique known as "non-pose-distinctive placement removal” and with other new ideas, including the "pillbox method” for approximating signatures with compact piecewise-analytic descriptions.
  • TOs are geometric procedures which obtain small sparse sets of points from range images in a regimented way. They are useful for surface shape recognition and pose estimation.
  • the new technique is composed of two steps. The first is training the system on a new object so that it will be later able to estimate the pose of that object when seen again in some range image. The second is the actual pose estimation, where a TO is placed at a random location on a new range image containing the object of interest.
  • the nearest neighbor in the TO feature space signature from the framing data is computed. If the distance to the nearpoint is less than some appropriate threshold, then the surface is recognized and pose estimation proceeds by computing the six pose parameters of a central triangle of the new TO placement in the coordinate system of the range imaging instrument. Then the pose parameters associated with the nearpoint are retrieved. An estimate of the pose of the surface shape in the new image is recovered using those two pose six-vectors; the pose of the central TO triangle in the new image and the retrieved pose of the central TO triangle in the training image are composed together to determine an estimate where the object actually is with respect to the location of its original model used in training.
  • Figure la shows a simple TO with one line probe fixed symmetrically with respect to the rigid triangle ABC.
  • Figure 2a shows a 2D projection of TO signatures taken with the TO of Figure lb with superimposed signatures of six hyperbolic paraboloid patches, four elliptic patches, and 10 spheres, each signature being a nearly circular ring.
  • Figure 2b shows a 2D projection of a TO signature of a torus taken with the TO of Figure lb whose signature is a piece of cone in ⁇ , ⁇ 2 ⁇ 3 space.
  • Figure 2c shows a 2D projection of a TO signature of a 90° planar dihedral taken with the TO of Figure lb, viewed diagonally.
  • Figure 2d shows a 2D projection of a TO signature of a 90° planar dihedral taken with the TO of Figure lb, viewed along the ⁇ , axis.
  • Figure3a shows a pictorial reference for describing the geometry of pose estimation using tripod operators and shows pose estimation using training with a tripod operator shown on the object in the training position in a range image.
  • Figure 3b shows a pictorial reference for describing the geometry of pose estimation using tripod operators and shows pose estimation with a tripod operator in the test position.
  • Figure 4 shows a flow chart for generating the prestored or training data-
  • Figure 5 shows a flow chart for the ideal description of pose estimation method using tripod operators.
  • FIG 3a illustrates the object in the range image used in the training stage.
  • the TO is placed many times on this image and information is stored as described below.
  • FIG. 3b illustrates the object in the range image used to perform actual pose estimation of the object.
  • a TO is placed somewhere, as indicated m FIG. 3b by "TEST PLACEMENT OF TRIPOD OPERATOR".
  • our pose estimation method determines, using information from the training process, what place on the object in the training image has the same shape as the test placement encountered. This place is indicated in FIG 3a by "ESTIMATED MATCHING TRIPOD OPERATOR
  • An "object-pose” is defined to be the rigid motion that takes the framing image of the object into alignment with .the image for which the object-pose is desired. This requires six parameters, e.g., three translational parameters, and three rotational ones. It is well known to those skilled in the art of computer vision or robotics how to represent pose in various mathematical forms.
  • the 'TO-pose" of a TO is defined as the rigid motion that takes the points A, B, and C of the TO placement in a prespecified "standard position" into the corresponding points A, B, and C of the TO whose TO-pose is desired.
  • Relative pose of two three-dimensional point sets is defined to be in rigid motion that brings the first point set into alignment with the second point set.
  • Object- TO-pose is defined to consist of two parts: first, a description of the location of a predefined central reference point on an object in a coordinate system fixed with respect to the triangle ABC, and secondly, a description of the relative pose of the triangle ABC as it lies on the object and the triangle ABC as it lies on the same place on the object with the object in a prespecified fraining position.
  • a tripod placement is said to be pose-distinctive if all other placements on the same fraining object giving nearly the same feature vector have nearly the same pose in 6 degrees of freedom (DOF) with respect to the object.
  • DOF degrees of freedom
  • a trihedral comer is not pose distinctive because there are three TO placements 120 degrees apart that give the same feature vector. Irregular or freeform shapes tend to have lots of pose-distinctive placements on them; symmetrical things have few. More precisely, a tripod operator (TO) placement P is said to
  • epsilon-delta pose distinctive if and only if all other placements which have a feature vector within a distance e of that of P also have object-TO-poses within a distance 6 of the object-TOrpose of P, using appropriate metrics in the two spaces.
  • an L-. (Euclidean) metric is used to compute distances in feature space. For pose, the metric is slightly more complicated. The distance between two object-TO-poses is regarded as characterizing how far an object would have to be moved to take it from one object-TO-pose to a second.
  • the distance is defined to be the Lj norm (root sum- of-squares) of the 4- vector (X, Y, Z, A) where X, Y, Z, are the Cartesian coordinates of the translation of a central reference point p ref in the object from one pose to a second, and A is the product of a scalar constant called posemetric and the angle in radians of the net rotation between the two poses. It is well known to those skilled in the art of computer vision or robotics how to compute this net rotation using "axis-angle" rotation representations. Posemetric is roughly the average radius of the.object, so that A is a rough measure of the distance moved by surface points due to rotation.
  • TOs are placed at many ( e -g-, 50,000) places on the computer-represented surface shape, such as a range image or a computer assisted design model.
  • the system stores 416 the resulting feature vector (f,, f 2 , ..., fj) and the associated information about the relative pose of the TO and the represented surface shape, after first computing such information. This is done by computing in the coordinate system of the represented surface shape the rigid motion
  • a test 418 is made to determine if the feature space signature has been sampled densely enough to allow the desired pose estimation accuracy. This density is measured by the RMS Euclidean distance between each feature vector and its nearest neighbor. Other measures of average besides RMS can be used and other metrics than Euclidean can be used.
  • the stored data for the training placements are screened for usefulness in pose estimation 422, and some are rejected, using "non-pose-distinctive" placement removal.
  • Non-pose- distinctive placement removal is performed during the training process as follows: First, make numerous TO placements on die object of interest, and store them. Then make numerous new placements, each of which is tested against the stored placements in an effort to prove them non-pose-distinctive. Specifically, if any stored placement is near the nearer than e to the new one in feature space, but farther than ⁇ from it in pose space, then it is discarded. This testing continues until a specified number of consecutive new placements are tried without it resulting in the removal of any stored placement. A stored point set with most or all non-pose-distinctive placements removed is called a "pose-distinctive signature.” An alternative to computing and storing a pose-distinctive signature as discrete points 422 is to
  • the nearpoint is found; that prestored point whose sum-of-squares distance to the newly measured feature vector is smallest, as in the recognition problem (without pose estimation) described in Pipitone; TRIPOD OPERATORS FOR REALTIME RECOGNITION OF SURFACE SHAPES IN RANGE IMAGES; NASA Tech. 2004 Symp., Washington, DC, Nov. 1994, which is hereby incorporated in total herein. If this distance is less than some appropriate small threshold, then the surface is recognized and pose estimation can proceed. This is done by computing the TO-pose (six parameters) of the central triangle A, B, C of the new placement in the coordinate system of the range imaging instrument, just as in the training phase. Then the object-TO- pose parameters associated with the nearpoint above are retrieved.
  • a given placement either results in no pose estimate, because it is not close to any feature-space point in the signature, or it returns a pose estimate.
  • the TO placement either does not lie on the desired object or it lies on a "symmetrical" or otherwise pose-ambiguous region of the object.
  • PILLBOX ALGORITHM A second deficiency of the pose estimation method described in the naive method earlier is that the stored points are discrete. This means that the more resolution desired in the pose, the more points that must be stored in the computer memory. Referring again to Figure 4, this tradeoff can be mitigated by making piecewise smooth approximations of the data 424 in such a way that the measured point in feature space is mapped by six scalar functions into six corresponding pose parameters. This is called the "pillbox method.” Each pillbox defines a small compact volume of feature space within in which a linear 3-manifold approximations the set of stored feature space points in that region. Also, a linear mapping is established within the pillbox from feature space to pose space.
  • a union of a number of pillboxes then can represent an extended region of a surface.
  • Making a pillbox representation, 424 in Figure 4 has several advantages.
  • Each pillbox is a simple function representing a small piece of feature space allowing fast mapping from feature values to the six pose parameters, using array indexing to quickly find the pillboxes near a measured feature point. The accuracy is high compared to using a discrete point signature of moderate density.
  • Finally the amount of required storage is small, since each pillbox typically covers many discrete points and only a moderate number (typically a
  • pillboxes 19 few hundred) of pillboxes is required to cover a reasonably large fraction of the pose- distinctive part of a feature-space signature by a union of the pillboxes.
  • Some of the following description is somewhat computer-oriented, using specific variable and array names, because it is based on text used to describe the algorithm to a programmer in order for it to be implemented.
  • To implement the pillbox generation algorithm the following steps are executed. (1) Take a random point "centerplace" in placement space, yielding a feature vector f c and a pose vector "centerpose.” (2) Take additional placements at many points "near" "centerplace” in placement space.
  • pillbox is defined by (df.t,) + (df.t,) 2 +(df.tj) 2 ⁇ dischen and (df.n,) 2 + (df.n 2 ) 2 + (df.n 3 ) 2 + ... ⁇ d t
  • a single application of a TO on the test range image 502 is done by first selecting an arbitrary point on the range image to be Point A of the TO.
  • point on the range image it is meant that either an actual range pixel expressed as a Cartesian triple (x, y, z) or an interpolated point between adjacent pixels is selected; next, an arbitrary point on the range image at the prescribed distance
  • a point C on the range image is chosen such that it is at the prescribed distance I AC I from A and at the prescribed distance
  • A, B and C must occur consistently in a clockwise (or CCW) order, as viewed from the range image sensor.
  • the d feature values f, through f d are computed according to "definition of a TO".
  • the d prototype feature values of fi through f d and the d + 3 points A, B, C, D , D d are stored in computer memory.
  • the case of the failed TO placement must be discussed. If any of the d + 3 points cannot be obtained, for any reason, the placement is aborted 104 and a fresh attempt begins at a different arbitrary point A 502. Failure usually occurs because valid range data is unavailable at the required
  • a test 504 if the number of TO placements exceeds some predefined threshold 102, quit. A reasonable value for this threshold is one corresponding to the maximum amount of time the user is willing to wait for a result in an application.
  • a test is made of whether FDIST is less than a prescribed threshold D-ra ⁇ H 508. If not, the match fails and Step 502 is executed again. If so, the program proceeds to 112.
  • D- ⁇ HRESH is selected so that it is small enough to exclude most false positive matches of surface shape, and large enough so that the results will not be mostly false negatives.
  • the pose estimates from the results of Step 502 and from prestored data from either 422 or 424 of Figure 4, are computed 512.
  • Pose estimation is performed by first computing the pose (six parameters) of the central triangle ABC of the new TO placement in the
  • the stored pose parameters represent the TO pose in the coordinate system of the fraining image, with the TO lying on the same surface feature in both the fraining and test images. An estimate of the pose of the surface shape is then recovered in the new image from those two pose 6- vectors.
  • the pose of the central triangle in the two images is known, and the disparity between these two poses represents where the object (surface shape) actually is with respect to the location of its original model. If the newly measured feature vector and its near point among the prestored data do indeed correspond exactly to TO placements at the same location on the surface shape, then the object pose estimate will be exact.
  • the estimated object pose is then output to a utilizing device 513 which is either a device for humans to read, or a robotic device which will interact with the object at its
  • FIGS. 6a and 6b an implementation of the full algorithm described above was made. It was implemented in the programming language "C"; however, any scientific language may be used. It was successfully tested on an object we call a "beveled outside comer", shown in FIGS.6a and 6b.
  • FIG. 6b is an example of what is illustrated in FIG. 3a; a synthetic range image of the object in fraining position.
  • FIG. 6a is an example of what is illustrated in FIG. 3b; a synthetic range image of the object in test position.
  • This "beveled comer” surface shape (or “object”) is essentialy an outside comer with one edge smoothed into a quarter cylinder, much like a comer of a typical computer monitor.
  • the system After training the system by placing the TO many times on the image of the object in its training position, the system stored the results in a "pillbox representation". A new synthetic range image was then generated, with noise added, in a different pose in 6 DOF, called the "test position" (FIG. 6a). Then pose estimation was performed. This consists of automatically placing the TO at several positions on this image until the system found a pose-distinctive placement. ( Figure 6a) The system was then displayed the place to which it believed the placement
  • 25 estimation procedure computes where it believes the "test placement” lies on the object, shown as "estimated matching tripod operator location 326, on Figure 3a.
  • the system estimates six parameters describing the rigid motion taking the object from its fraining position 326, as shown in Figure 3a, to its current position 328 in the test image.
  • the translational component of the pose of the object 332 is related to a reference point 334,'
  • a point in placement space is (i, j, ⁇ ). This is the usual way of specifying where to place a TO on a range image, e.g., in the related TO work. (See, supra, Pipitone, 1994.)
  • a point in feature space is (f lt f 2 , ..., J, these are the "hinge angles" described above in the case of the i plementated "linkable” type of TO, in the system.
  • a point in object-TO-pose space is (coefAB, coefAC, coefcross, axangX, axangY, axangZ) in the test prototype.
  • the first three give the location of a specified reference point "refpt" on the object, typically chosen more or less near its "center", in the reference frame of the tripod. See the included “get_cent_coefs( )", which generates the 3 coefficients and describes the recovery of refpt from A, B, C and the three coefficients.
  • the command "ps” allows the user to change the pose status (recognition or pose estimation) and set the object reference point.
  • the 3 “axang” numbers describe the orientation of the tripod with respect to the "standard orientation” defined in “orient3pts( )" ( an excerpt of the source
  • the 3-vector "axang” is oriented along the direction of the net rotation and has a magnitude equal to that of the net rotation.
  • Various angle normalizations are used, as described in the code.
  • the 6 pose numbers for each TO placement are stored as global variables, as follows (along with corresponding feature vectors): float refcoefAB[50000], refcoefAC[50000], refcoefCROSS[50000]; short tri_orient[50000][3]; /* angle in .01 degree units. */

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)

Abstract

A software procedure, with associated hardware, for estimating the pose of an object from a range image containing the object. A range image is a two dimensional array of numbers which represent the distances from a reference point in the range imaging instrument to observed surface points in a scene. All six parameters of the pose of an object are estimated; three translational and three angular parameters. A new technique known as 'non-pose-distinctive placement removal' is combined with of tripod operators (TOs), a method for interpreting range images, and is comprised of two steps. The first is training the system on a new object so that it will later be able to estimate the pose of that object when seen again in some range image (502). The second is the actual pose estimation, where a TO is placed at a random location on a new range image containing the object of interest. Then the nearest neighbor in feature space, the nearpoint, is computed. If the distance to the nearpoint is less than some appropriate threshold (508), then the surface is recognized and pose estimation proceeds by computing the six pose parameters (512) of a central triangle of the new placement in the coordinate system of the range imaging instrument. Then the pose parameters associated with the nearpoint are retrieved. An estimate of the pose of the surface shape in the new image is recoverable using those two pose six-vectors; the pose of the central TO triangle in the new image and the retrieved pose of the central TO triangle in the training image are composed together to determine an estimate where the object actually is with respect to the location of its original model used in training.

Description

A TECHNIQUE FOR ESTIMATING THE POSE OF SURFACE SHAPES USING TRIPOD OPERATIONS
BACKGROUND OF THE INVENTION Field of the Invention This invention pertains generally to an improved technique for recognizing and locating objects and more specifically a technique for estimating the pose of an object from a range image containing the object.
Description of the Related Art In recent decades, a wide variety of instruments have been built to obtain range images; a range image being a two-dimensional array of numbers which gives the depth of a scene along many directions from a central point in the instrument. Instead of measuring the brightness of many points in a scene, as in a television camera, these instruments measure where each point is in a three-dimensional space. Both range images and the more conventional intensity images from digital cameras have been used in the computer vision research community to deteπnine the pose of observed objects. The term "object" , as used herein, means a particular surface shape. "Pose" means a complete description of an object's position and orientation. For a rigid object this requires six numbers, such as X, Y, Z, pitch, yaw and roll, or six equivalent coordinates. The previous methods for pose estimation all suffer from either a lack of generality or from time inefficiency. It is possible to do pose estimation using tripod operators (TO). See, Pipitone; TRJPOD OPERATORS FOR THE INTERPRETATION OF RANGE IMAGES; NRL Memorandum Report 6780, Feb 1991 for a crude and incomplete discussion. Tripod operators are a versatile class of feature extraction operators for surfaces. They are useful for recognition and/or localization (pose estimation) based on range or tactile data. They extract a few sparse point samples in a regimented way so that N surface points yield only N-3 independent scalar features containing all the pose-invariant surface shape information in these points and no other information. They provide a powerful index into sets of prestored surface representations. A TO consists of three points in 3-space fixed at the vertices of a triangle and a procedure for making several "depth" measurements in the coordinate frame of the triangle, which is placed on the surface like a surveyor's tripod. TOs can be embedded in a vision system in many ways and applied to almost any surface shape. As stated above, a TO consists of three points in space fixed at the vertices of a triangle of fixed edge lengths and a procedure for making several depth measurements in the coordinate frame of the triangle, which is placed on the surface like a surveyor's tripod. These measurements take the form of arc-lengths along "probe curves" at which the surface is intersected. Figure la through lc shows three examples of TO's. Figure la shows a very simple TO with one line probe fixed symmetrically with respect to the rigid triangle ABC. The single scalar feature is the distance from the plane of ABC at which the probe intersects the surface. This resembles a mechanical optician's tool called a spherometer. A spherometer is the simplest possible tripod operator. It is a mechanical device whose feet A, B and C are placed on a lens whose curvature is to be measured. The probe is a sliding rod which contacts the lens at a depth approximately proportional to the curvature of the lens, and hence to its optical strength. The number d of scalar features is called the order of the operator. A spherometer therefore has order d = 1. Particularly useful are tripod operators with higher order, e.g. d = 4 to 15. These can be used to rapidly recognize a very wide range of surface shapes, because the d probes intersect a surface at d particular values which are frequently unique to that place on that surface shape. Figures lb and lc show TO's that can be viewed as a set of equilateral triangles hinged together so that all d+3 points can be made to contact a surface. The angles of the d binges are the features. This type, called linkable TO's, is preferred because of their symmetry and uniform sensitivity to noise. The application if this TO to a planar surface yields φ=0 for all the hinges. Many variations of these TO's could be constructed. Feature noise is related to range noise n by the approximate expression ^=51 x n/e, where n^ is the feature error in degrees, and n is expressed in the same distance units as the edge length e. From an N-point TO, the N sampled surface points yield only N - 3 independent scalar features, and the order d is N -3. These features contain all the surface shape information in the 3N components of the points since they suffice to reconstruct the relative positions of the N points. They contain no other information. For example, they have complete six DOF invariance under rigid motions, the group R3 x SO(3). Thus, they depend upon where the tripod lies on the surface, but upon nothing else. A key property is that for any dimensionality d of feature vector only a 3 (or fewer)-dimensional manifold of feature space points can be generated from a given surface, since the tripod can be moved in only 3 DOF on a surface. This allows objects to be densely sampled with TOs at preprocessing time with a manageable number of operator applications, typically a few thousand, to obtain almost all of the possible feature vector values obtainable from any range image of the object. This set is a kind of invariant signature. For brevity, this is called the signature of the object or surface, with respect to a particular type TO. It can be stored in an array of bins in feature space, e.g., of dimension 3 or 4, for later efficient access of near neighbors to TO features measured at recognition time. These bins can optionally contain precomputed probability densities, analytic expressions for distances to nearby signature manifolds, and partial or complete descriptions of the relative poses of tripods and models, all to serve various purposes in a recognition system. Since in some applications of the tripod operator, the computation consists only of placement and a little indexing, the cost of placing the operator should be kept small. This can be done by efficiently implementing a procedure similar to the following. Consider placing the TO's of Figures lb or lc on a dense range map. Point A can be chosen as any point on the image surface. Interpolation is to be done locally as needed, e.g., using piecewise triangular facets. Point B can be found by moving along a line at orientation a in image coordinates, pixel indices, until the 3D distance | AB [ = e. This can be done in logarithmic time, essentially constant here, using binary search around the circle of radius .5/3 e oriented coaxially around the center of the segment AB to find a point C close to the surface. A similar circular search yields each remaining point. A key step in the circular search is the mapping, specific to a range scanner's geometry, from a point (x, y, z) to the two indices of the range pixel whose ray (x, y, z) lies on. This allows the front/behind decision required by the binary search. In the case of a sequential random access range scanner, it may be efficient to monotonically search elliptical paths in image coordinates until the two distances being enforced, e.g., | AC | and | BC | , are both correct. The ellipses here are the projections of the previously described circles onto image coordinates. Finally, in the case of a tactile TO, the computation is mechanical; the feature values are to be read from position transducers, e.g., from linear potentiometers by an analog-to-digital (A D) converter. The following are a few of the symmetry properties of TOs of the types of figures lb and lc. Surfaces with one symmetry, such as extrusions, surfaces of revolution, and helical projections produce only a 2-dimensional manifold in feature space ( for Figures lb and lc). Cylinders, having two symmetries, produce only a nearly circular 1 -dimensional curve, and spheres a single point. Scaling a TO by changing its edgelength does not affect the signature of surfaces swept by a line with one point fixed, e.g., cones, planar n-hedral vertices, and planar dihedral edges. Regardless of the surface, an operator with a 3-fold symmetry, e.g., those in Figures lb and lc, produces signatures unchanged by cyclically permuting each triple of corresponding features. In Figure lc, the three 3-cycles (1, 2, 3), (4, 5, 6) and (7, 8, 9) show this property, for features φ, through φ9, respectively. This allows a 3-fold storage reduction, e.g., by permuting the features so that φ, is the largest. If the TO, in addition, has handedness symmetry, the signature can be modified by a procedure that allows recognition of the "other side" of any surface already recognizable. This is called inversion of a signature. It is done by transposing certain pairs of corresponding features, e.g., (7,5), (1,2), (4,8), and (6,9) in Figure lc, and replacing each feature value with - φ. Also, the signature of the opposite-handed (reflected) version of a surface can be found by performing those transpositions without negating the features. The signatures of order 3 operators, Figure lb, were rendered as a rotating cloud of points on a computer; selected 2D snapshots are shown in Figures 2a through 2d. In a special case of "smooth" surface regions, the signature is nearly a circular ring coaxial with the diagonal axis. The offset and radius of the ring can be readily used to compute estimates of the principle curvatures and other differential geometric parameters. Surfaces with Cj or C2 discontinuities tend to produce signatures with similar numbers and kinds of discontinuities, e.g., Figures 2c and 2d, and have roughly commensurate complexities of description. Thus, this umbrella-shaped 2-manifold can be well approximated with a few polynomials, whereas the discrete signature might need 20,000 points for thorough saturation. These signatures are point sets in feature space which are the same for a given surface shape, regardless of where in space that shape is located when the signature is formed using the TO. As noted above, they are at most three-dimensional manifolds of points, which occurs in the case of irregular, or free-form surfaces. This means that even if a high order TO is used, such as d = 9, the signature still has the low dimensionality of three at most. This makes it practical to store complete signatures of objects without using large amounts of computer memory. This operator (in the invention's preferred form, the linkable TO) is essentially a 3D simulated structure consisting of several triangles hinged together. It is applied to a computer-represented surface, such as range image, by moving it until all of its points lie on that surface. Then the "hinge angles" provide information about the shape of the surface. In a simplified application of tripod operators, there are two major steps. The first is training the system on a new object so that it will later be able for the system to recognize (or estimate the pose of) that object when seen again in some range image. The second step is the actual recognition (or, pose estimation).
SUMMARY OF THE INVENTION The objective of this invention is to provide an improved technique for estimating the pose of surface shapes in six degrees of freedom from a range image containing an object possessing such a surface shape. This and other objectives are accomplished by using a software procedure, with associated hardware, for estimating the pose of an object from a range image containing the object. A range image is a two dimensional array of numbers which represent the distances from a reference point in the range imaging instrument to observed surface points in a scene. All six parameters of the pose of an object are estimated; three translational and three angular parameters. This technique involves combining the previously existing method called tripod operators (TOs) with a new technique known as "non-pose-distinctive placement removal" and with other new ideas, including the "pillbox method" for approximating signatures with compact piecewise-analytic descriptions. TOs are geometric procedures which obtain small sparse sets of points from range images in a regimented way. They are useful for surface shape recognition and pose estimation. The new technique is composed of two steps. The first is training the system on a new object so that it will be later able to estimate the pose of that object when seen again in some range image. The second is the actual pose estimation, where a TO is placed at a random location on a new range image containing the object of interest. Then the nearest neighbor in the TO feature space signature from the framing data is computed. If the distance to the nearpoint is less than some appropriate threshold, then the surface is recognized and pose estimation proceeds by computing the six pose parameters of a central triangle of the new TO placement in the coordinate system of the range imaging instrument. Then the pose parameters associated with the nearpoint are retrieved. An estimate of the pose of the surface shape in the new image is recovered using those two pose six-vectors; the pose of the central TO triangle in the new image and the retrieved pose of the central TO triangle in the training image are composed together to determine an estimate where the object actually is with respect to the location of its original model used in training.
BRIEF DESCRIPTION OF THE DRAWINGS Figure la shows a simple TO with one line probe fixed symmetrically with respect to the rigid triangle ABC. Figure lb shows an order 3 "linkable" TO that can be viewed as a set of equilateral triangles hinged together so that all d+3=6 points can be made to contact a surface. Figure lc shows an order 9 "linkable" TO that can be viewed as a set of equilateral triangles hinged together so that all d+3=12 points can be made to contact a surface. Figure 2a shows a 2D projection of TO signatures taken with the TO of Figure lb with superimposed signatures of six hyperbolic paraboloid patches, four elliptic patches, and 10 spheres, each signature being a nearly circular ring.. Figure 2b shows a 2D projection of a TO signature of a torus taken with the TO of Figure lb whose signature is a piece of cone in φ, φ2φ3 space. Figure 2c shows a 2D projection of a TO signature of a 90° planar dihedral taken with the TO of Figure lb, viewed diagonally. Figure 2d shows a 2D projection of a TO signature of a 90° planar dihedral taken with the TO of Figure lb, viewed along the φ, axis. Figure3a shows a pictorial reference for describing the geometry of pose estimation using tripod operators and shows pose estimation using training with a tripod operator shown on the object in the training position in a range image. Figure 3b shows a pictorial reference for describing the geometry of pose estimation using tripod operators and shows pose estimation with a tripod operator in the test position. Figure 4 shows a flow chart for generating the prestored or training data- Figure 5 shows a flow chart for the ideal description of pose estimation method using tripod operators. points A, B and C are found, the circular probes are applied by rotating each new triangle about each successive hinge (labelled 1 through nine), resulting in angular feature values which have noise variances that depend fairly uniformly on the range error at the intersection points. This simplifies the use of the feature measurements. Thus, our most preferred embodiment uses this triangular
structure, called "linkable" in the Background. We regard regularly placed parallel linear probes as a close second in our preference. Definitions of several types of pose to facilitate the following descriptions are noted for clarification. It will be noted in Figures 3a and 3b that a number of items are shown in relation to a coordinate system fixed with respect to the range scanner, i.e., the range imaging instrument which acquires the images of the object in "test position" and in Gaining position".
FIG 3a illustrates the object in the range image used in the training stage. The TO is placed many times on this image and information is stored as described below. FIG. 3b illustrates the object in the range image used to perform actual pose estimation of the object. A TO is placed somewhere, as indicated m FIG. 3b by "TEST PLACEMENT OF TRIPOD OPERATOR". Then our pose estimation method determines, using information from the training process, what place on the object in the training image has the same shape as the test placement encountered. This place is indicated in FIG 3a by "ESTIMATED MATCHING TRIPOD OPERATOR
13 LOCATION". We will now elaborate this process further. An "object-pose" is defined to be the rigid motion that takes the framing image of the object into alignment with .the image for which the object-pose is desired. This requires six parameters, e.g., three translational parameters, and three rotational ones. It is well known to those skilled in the art of computer vision or robotics how to represent pose in various mathematical forms. The 'TO-pose" of a TO is defined as the rigid motion that takes the points A, B, and C of the TO placement in a prespecified "standard position" into the corresponding points A, B, and C of the TO whose TO-pose is desired. "Relative pose" of two three-dimensional point sets (e.g., TOs or range images of objects) is defined to be in rigid motion that brings the first point set into alignment with the second point set. Object- TO-pose is defined to consist of two parts: first, a description of the location of a predefined central reference point on an object in a coordinate system fixed with respect to the triangle ABC, and secondly, a description of the relative pose of the triangle ABC as it lies on the object and the triangle ABC as it lies on the same place on the object with the object in a prespecified fraining position. A tripod placement is said to be pose-distinctive if all other placements on the same fraining object giving nearly the same feature vector have nearly the same pose in 6 degrees of freedom (DOF) with respect to the object. Thus a trihedral comer is not pose distinctive because there are three TO placements 120 degrees apart that give the same feature vector. Irregular or freeform shapes tend to have lots of pose-distinctive placements on them; symmetrical things have few. More precisely, a tripod operator (TO) placement P is said to
14 be epsilon-delta pose distinctive if and only if all other placements which have a feature vector within a distance e of that of P also have object-TO-poses within a distance 6 of the object-TOrpose of P, using appropriate metrics in the two spaces. Specifically, an L-. (Euclidean) metric is used to compute distances in feature space. For pose, the metric is slightly more complicated. The distance between two object-TO-poses is regarded as characterizing how far an object would have to be moved to take it from one object-TO-pose to a second. The distance is defined to be the Lj norm (root sum- of-squares) of the 4- vector (X, Y, Z, A) where X, Y, Z, are the Cartesian coordinates of the translation of a central reference point pref in the object from one pose to a second, and A is the product of a scalar constant called posemetric and the angle in radians of the net rotation between the two poses. It is well known to those skilled in the art of computer vision or robotics how to compute this net rotation using "axis-angle" rotation representations. Posemetric is roughly the average radius of the.object, so that A is a rough measure of the distance moved by surface points due to rotation. TRAINING PHASE During the training portion of the algorithm , as shown in Figure 4, TOs are placed at many (e-g-, 50,000) places on the computer-represented surface shape, such as a range image or a computer assisted design model. For each such placement 414, the system stores 416 the resulting feature vector (f,, f2, ..., fj) and the associated information about the relative pose of the TO and the represented surface shape, after first computing such information. This is done by computing in the coordinate system of the represented surface shape the rigid motion
15 which takes the triangle ABC of the TO from a prespecified "standard position" into its current position on the surface. It is well known by those skilled in the art how to accomphsh this function, using a variety of representations of the six parameters of rotation and translation. A test 418 is made to determine if the feature space signature has been sampled densely enough to allow the desired pose estimation accuracy. This density is measured by the RMS Euclidean distance between each feature vector and its nearest neighbor. Other measures of average besides RMS can be used and other metrics than Euclidean can be used. The stored data for the training placements are screened for usefulness in pose estimation 422, and some are rejected, using "non-pose-distinctive" placement removal. This corresponds to ambiguous shapes, particularly those possessing symmetries. Non-pose- distinctive placement removal is performed during the training process as follows: First, make numerous TO placements on die object of interest, and store them. Then make numerous new placements, each of which is tested against the stored placements in an effort to prove them non-pose-distinctive. Specifically, if any stored placement is near the nearer than e to the new one in feature space, but farther than δ from it in pose space, then it is discarded. This testing continues until a specified number of consecutive new placements are tried without it resulting in the removal of any stored placement. A stored point set with most or all non-pose-distinctive placements removed is called a "pose-distinctive signature." An alternative to computing and storing a pose-distinctive signature as discrete points 422 is to
16 utilize the pill box method 424, which produces a pose-distinctive piecewise analytic approximation of the signature. THE NAIVE METHOD In order to more clearly explain pose estimation utilizing a pose-distinctive signature, a simple version of the method and its shortcomings is described hereafter. After prestoring the feature vector and associated six pose parameters for many TO placements on a model surface, a TO is placed at a random location on a new range image containing the object of interest. The nearest neighbor in feature space is sought. The nearpoint is found; that prestored point whose sum-of-squares distance to the newly measured feature vector is smallest, as in the recognition problem (without pose estimation) described in Pipitone; TRIPOD OPERATORS FOR REALTIME RECOGNITION OF SURFACE SHAPES IN RANGE IMAGES; NASA Tech. 2004 Symp., Washington, DC, Nov. 1994, which is hereby incorporated in total herein. If this distance is less than some appropriate small threshold, then the surface is recognized and pose estimation can proceed. This is done by computing the TO-pose (six parameters) of the central triangle A, B, C of the new placement in the coordinate system of the range imaging instrument, just as in the training phase. Then the object-TO- pose parameters associated with the nearpoint above are retrieved. If the newly measured feature vector and its nearpoint among the prestored data correspond to TO placements at the same location on the surface shape, then a recovery of an estimate of the object-TO-pose of the surface shape in the new image from those two pose 6- vectors is accomplished. That is, the pose of the central triangle in the new image is known, and the
17 retrieved pose of the central TO triangle in the triangle image, and the disparity between these two poses represents where the object (surface shape) actually is with respect to the location of its original model. To clarify this further, consider this analogy: The TO is like a spider whose feet are like the points of the TO. The spider feels a specific surface shape, and knows exactly where it is on the surface. This knowledge is like "Object-TO-pose". We can see the spider from our viewpoint, which is like the position of the range scanner. The spider's position from that viewpoint is like "TO-pose". Clearly, we can combine these two pieces of information to determine where the surface shape is from our viewpoint. This is like "object-pose", completing our analogy. One deficiency of the naive foregoing technique is that when the feature vector "matches" the prestored data in that a very close nearpoint is found, the two corresponding placements of the TO may not lie on the same place on the surface. This is particularly common in symmetrical surfaces such as surfaces of revolution, linear extrusions, planes, spheres, cylinders, trihedral corners, etc. These surfaces shapes do not allow the unambiguous determination of all six pose parameters. When a pose-distinctive signature is used in pose estimation, it is used just as in the naive method above, except that there will now be few problems with pose ambiguity. Specifically, in performing pose estimation, the TO is placed at random locations on the range image within some region, perhaps the whole image, believed to contain the object of
18 interest. A given placement either results in no pose estimate, because it is not close to any feature-space point in the signature, or it returns a pose estimate. In the former case, the TO placement either does not lie on the desired object or it lies on a "symmetrical" or otherwise pose-ambiguous region of the object.
PILLBOX ALGORITHM A second deficiency of the pose estimation method described in the naive method earlier is that the stored points are discrete. This means that the more resolution desired in the pose, the more points that must be stored in the computer memory. Referring again to Figure 4, this tradeoff can be mitigated by making piecewise smooth approximations of the data 424 in such a way that the measured point in feature space is mapped by six scalar functions into six corresponding pose parameters. This is called the "pillbox method." Each pillbox defines a small compact volume of feature space within in which a linear 3-manifold approximations the set of stored feature space points in that region. Also, a linear mapping is established within the pillbox from feature space to pose space. A union of a number of pillboxes then can represent an extended region of a surface. Making a pillbox representation, 424 in Figure 4, has several advantages. Each pillbox is a simple function representing a small piece of feature space allowing fast mapping from feature values to the six pose parameters, using array indexing to quickly find the pillboxes near a measured feature point. The accuracy is high compared to using a discrete point signature of moderate density. Finally the amount of required storage is small, since each pillbox typically covers many discrete points and only a moderate number (typically a
19 few hundred) of pillboxes is required to cover a reasonably large fraction of the pose- distinctive part of a feature-space signature by a union of the pillboxes. Some of the following description is somewhat computer-oriented, using specific variable and array names, because it is based on text used to describe the algorithm to a programmer in order for it to be implemented. To implement the pillbox generation algorithm, the following steps are executed. (1) Take a random point "centerplace" in placement space, yielding a feature vector fc and a pose vector "centerpose." (2) Take additional placements at many points "near" "centerplace" in placement space. Then pick ranges of values for the TO placement parameters i, j and θ (defined in Pipitone, supra, 1994) and take points randomly or uniformly within the ranges. For each placement made, the six corresponding object-TO-pose parameters are computed and stored along with the corresponding feature vector ( f,, f2, ..., which is preferably of dimension 4≤d≤20. (3) Retain only those of (2) whose pose is within some specified distance "pose_spread" of "centerpose." Measure the pose distance as described in the definition described earlier. Choose a value of "pose_spread" large enough to cover as large a piece of the signature as possible without encountering a severe deviation from linearity, either for the feature point set approximation or for the feature space to pose mapping. (4) In order to fit a linear manifold in feature space to the placements from (3) compute the eigenvalues of the scatter matrix of the feature space points, a method well
20 known to those skilled in the arts of statistics and computer vision. What is desired is three "large" eigenvalues and d-3 "small ones> indicating a fully 3D manifold in feature space. Otherwise the "centerplace" will be aborted and the algorithm started somewhere else. An aborted try usually means that either the linear manifold is less than three dimensional due to surface symmetry, or it is more than three dimensional due to high curvature, in which case a linear fit would be poor. (5) Make a "pillbox" around the feature space linear patch of (4). The pillbox is a
"hypercylinder" of radius d. centered on fc." The 3 orthogonal radial directions, the eigenvectors of the 3 largest eigenvalues, are the tangents and the d-3 other orthogonal directions are normals. Call these the t„ tτ, tj, nl5 n2, n3, n4, ... respectively. The pillbox is defined by imposing the radius "d," (radial) and a thickness 2d, (tangential); Let df = (f - Q be (the feature vector) - (the center of the pillbox in fspace). This is a d - dimensional vector. Then the pillbox is defined by (df.t,) + (df.t,)2 +(df.tj)2 < d „ and (df.n,)2 + (df.n2)2 + (df.n3)2 + ... < d t
(6) Make a 6 x (d+1) matrix which maps a feature vector f into the 6 pose parameters, for f within the pillbox. Do this by separately treating each of the 6 pose parameters as a linear function off; e.g., Pi = a + c,f, + c^ + ... + c9 ^ for pose parameter number 1 of the six (for d=9). This is a standard linear regression problem in the field of statistics; minimizing the mean squared eπor of the linear approximation leads to an order d+1 linear system. (7) Test the pillbox by making random placements all over placement space. For each one falling inside the pillbox in feature space, check to see that its recovered pose is
21 within some tolerance t of the actual pose for that placement using the posemetric, and retain only those that are within some tolerance t of the actual pose. This step eliminates non-pose- distinctive points as well as points for which the linear mapping is not a sufficiently accurate approximation. The foregoing algorithm allows the system to place the TO on a test image (502, 504, 506, 508 and 512 of Figure 5) and rapidly give one of these two distinctive
responses: (1) This is a "pose-distinctive" placement on the desired object, and here are the six pose parameters. (2) This is not useful, i.e., either the TO is not on the desired object, or it is on a place on the object that is confusable with another object, or the placement is not "pose- distinctive."
POSE ESTIMATION PHASE The implementation of the actual pose estimation portion of the algorithm is described, as shown in Figure 5. A single application of a TO on the test range image 502 is done by first selecting an arbitrary point on the range image to be Point A of the TO. By "point on the range image", it is meant that either an actual range pixel expressed as a Cartesian triple (x, y, z) or an interpolated point between adjacent pixels is selected; next, an arbitrary point on the range image at the prescribed distance | AB | from point A is chosen as point B. Next, a point C on the range image is chosen such that it is at the prescribed distance I AC I from A and at the prescribed distance | BC | from B. Furthermore, A, B and C must occur consistently in a clockwise (or CCW) order, as viewed from the range image sensor.
22 Having determined A, B and C, the d feature values f, through fd are computed according to "definition of a TO". Thus, at the end of step one 502, the d prototype feature values of fi through fd and the d + 3 points A, B, C, D , Dd are stored in computer memory. The case of the failed TO placement must be discussed. If any of the d + 3 points cannot be obtained, for any reason, the placement is aborted 104 and a fresh attempt begins at a different arbitrary point A 502. Failure usually occurs because valid range data is unavailable at the required
location. A test 504; if the number of TO placements exceeds some predefined threshold 102, quit. A reasonable value for this threshold is one corresponding to the maximum amount of time the user is willing to wait for a result in an application. The distance between the feature vector ( f, through fd ) 506 obtained in 502 and the nearest point in the prestored signature, 402 of Figure 4, is computed, the FDIST. This distance maybe computed using a variety of metrics; the familiar L2 ( Euclidean) metric is preferred. A test is made of whether FDIST is less than a prescribed threshold D-ra^H 508. If not, the match fails and Step 502 is executed again. If so, the program proceeds to 112. D-ΓHRESH is selected so that it is small enough to exclude most false positive matches of surface shape, and large enough so that the results will not be mostly false negatives. The pose estimates from the results of Step 502 and from prestored data from either 422 or 424 of Figure 4, are computed 512. Pose estimation is performed by first computing the pose (six parameters) of the central triangle ABC of the new TO placement in the
23 coordinate system of the range imaging instrument, just as shown in 416, Figure 4, of the fraining phase. This is the 'TO-pose". Next the stored object-TO-pose parameters associated with the near point from Step 508, above are retreived using the data of either 422 or 424. In the discrete point case of training data from 422, indexing is done as described in Pipitone 1994, supra, except with pose estimation occurring, not just recognition. In the case of 424, instead of finding the nearest neighbor point in the prestored discrete feature point data, pillboxes are used for matching. This is accomplished by initially prestoring the set of pillboxes in feature space, indexed by an array, in 3 or 4 dimensions in feature space, whose cells point to the few pillboxes which overlap or nearly overlap that cell. Then during pose estimation, a pillbox is quickly recovered which contains the feature point obtained by the current TO placement, and is used to map from feature space to pose space. The stored pose parameters represent the TO pose in the coordinate system of the fraining image, with the TO lying on the same surface feature in both the fraining and test images. An estimate of the pose of the surface shape is then recovered in the new image from those two pose 6- vectors. That is, the pose of the central triangle in the two images is known, and the disparity between these two poses represents where the object (surface shape) actually is with respect to the location of its original model. If the newly measured feature vector and its near point among the prestored data do indeed correspond exactly to TO placements at the same location on the surface shape, then the object pose estimate will be exact. The estimated object pose is then output to a utilizing device 513 which is either a device for humans to read, or a robotic device which will interact with the object at its
24 known pose. EXAMPLE Referring again to FIGS. 3a and 3b, an implementation of the full algorithm described above was made. It was implemented in the programming language "C"; however, any scientific language may be used. It was successfully tested on an object we call a "beveled outside comer", shown in FIGS.6a and 6b. FIG. 6b is an example of what is illustrated in FIG. 3a; a synthetic range image of the object in fraining position. Similarly, FIG. 6a is an example of what is illustrated in FIG. 3b; a synthetic range image of the object in test position. This "beveled comer" surface shape (or "object") is essentialy an outside comer with one edge smoothed into a quarter cylinder, much like a comer of a typical computer monitor. After training the system by placing the TO many times on the image of the object in its training position, the system stored the results in a "pillbox representation". A new synthetic range image was then generated, with noise added, in a different pose in 6 DOF, called the "test position" (FIG. 6a). Then pose estimation was performed. This consists of automatically placing the TO at several positions on this image until the system found a pose-distinctive placement. (Figure 6a) The system was then displayed the place to which it believed the placement
corresponded. This is the TO placement shown in Figure 6b. The accuracy of the estimated placement of the 12-point TO is apparent to the eye. The system also computed corresponding numerical pose parameters. The total time for this pose estimation operation was less than 50 milliseconds using a Sun Microsystems SPARC 20 computer. The pose
25 estimation procedure computes where it believes the "test placement" lies on the object, shown as "estimated matching tripod operator location 326, on Figure 3a. In numerical terms, the system estimates six parameters describing the rigid motion taking the object from its fraining position 326, as shown in Figure 3a, to its current position 328 in the test image. The translational component of the pose of the object 332 is related to a reference point 334,'
called p^ which is selected to be near the center of the object in the fraining phase. This facilitates the construction of a meaningful metric for pose difference. DEFINITIONS (1) A point in placement space is (i, j, θ). This is the usual way of specifying where to place a TO on a range image, e.g., in the related TO work. (See, supra, Pipitone, 1994.) (2) A point in feature space is (flt f2, ..., J, these are the "hinge angles" described above in the case of the i plementated "linkable" type of TO, in the system. (3) A point in object-TO-pose space is (coefAB, coefAC, coefcross, axangX, axangY, axangZ) in the test prototype. The first three give the location of a specified reference point "refpt" on the object, typically chosen more or less near its "center", in the reference frame of the tripod. See the included "get_cent_coefs( )", which generates the 3 coefficients and describes the recovery of refpt from A, B, C and the three coefficients. The command "ps" allows the user to change the pose status (recognition or pose estimation) and set the object reference point. The 3 "axang" numbers describe the orientation of the tripod with respect to the "standard orientation" defined in "orient3pts( )" ( an excerpt of the source
26 code is shown in Appendix A). The 3-vector "axang" is oriented along the direction of the net rotation and has a magnitude equal to that of the net rotation. Various angle normalizations are used, as described in the code. The 6 pose numbers for each TO placement are stored as global variables, as follows (along with corresponding feature vectors): float refcoefAB[50000], refcoefAC[50000], refcoefCROSS[50000]; short tri_orient[50000][3]; /* angle in .01 degree units. */
Although the invention has been described in relation to an exemplary embodiment thereof, it will be understood by those skilled in the art that still other variations and
Figure imgf000026_0001
detracting from the scope and spirit of the invention as described in the claims.
27

Claims

WHAT IS CLAIMED:
1. A system for identifying an object comprising; a first range imaging scanner for digitally scanning a pluraUty of known objects to determine image; a digital storage device for storing the plurality of known digital image; a second range imaging devices for digitally scanning a plurality of unknown objects to produce a plurality of unknown digital images; a computer for comparing the known digital images from the first range imaging scanner with the unknown digital images of the second range imaging scanner utilizing tripod operators to obtain the identity of the unknown digital images; and a utilizing device receiving the identity of the unknown digital image.
2. A system for identifying an object comprising; a means for digitally scanning a plurality of known objects to determine an image; means for storing the plurality of known digital image; means for digitally scanning a pluraUty of unknown objects to produce a pluraUty of unknown digital images; means for comparing the known digital images with the unknown digital images utilizing tripod operators to obtain the identity of the unknown digital images; and means for displaying the identity of the unknown digital images.
28
3. A system, as in Claim 2, wherein the means for digitally scanning a pluraUty of known objects to determine an image is digital range imaging scanner.
4. A system, as in Claim 2, wherein the means for storing the pluraUty of known digital images is a digital storage device.
5. A system, as in Claim 2, wherein the means for digitally scanning a plurality of unknown objects to produce a pluraUty of unknown digital images is a range imaging scanner.
6. A system, as in Claim 2, wherein the means for comparing the known digital images with the unknown digital images utilizing tripod operators to obtain the identity of the unknown digital images is a computer.
7. A system, as in Claim 2, wherein the means for displaying the identity of the unknown digital images is a computer display device.
8. A method for identifying an object comprising; digitally scanning a plurality of known images by applying tripod operators to the images at random points to obtain six numbers of pose of the tripod operators with respect to the image range data coordinate system to obtain the known images feature vector; storing the digital scanned image;
29 digitally scanning a pluraUty of unknown digital images to obtain the unknown images feature vector; randomly applying tripod operator to the at random points to obtain six numbers of pose of the tripod operators with respect to the image range data coordinate system; comparing the six pose numbers of the unknown images to the stored digital data of the known images to identify the unknown image; and displaying the identity of the unknown digital images.
9. A computer system for computing an estimated pose of an object from a range image containing the object, comprising: a computer for receiving scanned digital representation of known objects and (a) applying a tripod operator at a random place on the digital representation of a surface shape whose pose is later desired to estimate in some range image; (b) store in computer memory the feature vector and the relative pose of the tripod operator and the surface shape; (c) calculate if the density of the feature vector is greater than some predetermined threshold; (d) if the density of the feature vector is not greater than some predetermined threshold, reapply a tripod operator , as in (a); (e) if the density of the feature vector is greater than some predetermined threshold, perform non-distinctive placement removal on the data in (b);
30 said computer after receiving scanned digital representation of an unknown object known as a test range image; (f) applies a tripod operator at a random place on the test range image; (g) march a feature vector from (f) to the stored data in (b) for the unknown object; (h) if the feature vector is not within a predetermined distance of the stored feature data, the computer reappUes a tripod operator, as in (f); (i)if the feature vector is within a predetermined distance of the stored feature data, the computer computes an estimate of six pose parameters of the unknown object; and a utilizing device to receive the six pose parameters.
10. A computer system for computing an estimated pose of an object from a range image containing the object, comprising: a computer for receiving scanned digital representation of known objects and (a) applying a tripod operator at a random place on the digital representation of a surface shape whose pose is later desired to estimate in some range image; (b) store in computer memory the feature vector and the relative pose of the tripod operator and the surface shape; (c) calculate if the density of the feature vector is greater than some predetermined threshold;
31 (d) if the density of the feature vector is not greater than some predeteπnined threshold, reapply a tripod operator , as in (a); (e) if the density of the feature vector is greater than some predetermined threshold, convert the stored data into a compact piecewise analytic description; said computer after receiving scanned digital representation of an unknown object known as a test range image; (f) applies a tripod operator at a random place on the test range image; (g) march a feature vector from (f) to the stored data in (b) for the unknown object; (h) if the feature vector is not within a predetermined distance of the stored feature data, the computer reappUes a tripod operator, as in (f); (i) if the feature vector is within a predetermined distance of the stored feature data, the computer computes an estimate of six pose parameters of the unknown object; and a utilizing device to receive the six pose parameters.
11. A process for computing an estimated pose of an object from a range image containing the object, comprising: receiving scanned digital representation of known objects and (a) applying a tripod operator at a random place on the digital representation of a surface shape whose pose is later desired to estimate in some range image;
32 (b) storing in computer memory the feature vector and the relative pose of the tripod operator and the surface shape; (c) calculating if the density of the feature vector is greater than some predetermined threshold; (d) reapplying a tripod operator , as in (a); if the density of the feature vector is not greater than some predetermined threshold, (e) performing non-distinctive placement removal on the data in (b); if the density of the feature vector is greater than some predetermined threshold, after receiving scanned digital representation of an unknown object known as a test range image; (f) applying a tripod operator at a random place on the test range image; (g) matching a feature vector from (f) to the stored data in (b) for the unknown object; (h) reapplying a tripod operator, as in (f); if the feature vector is not within a predetermined distance of the stored feature data, (i) computing an estimate of six pose parameters of the unknown object; if the feature vector is within a predetermined distance of the stored feature data, and applying the six pose parameters to a utilizing device.
12. A process for computing an estimated pose of an object from a range image containing the object, comprising:
33' receiving scanned digital representation of known objects and (a) applying a tripod operator at a random place on the digital representation of a surface shape whose pose is later desired to estimate in some range image; (b) storing in computer memory the feature vector and the relative pose of the tripod operator and the surface shape; (c) calculating if the density of the feature vector is greater than some predetermined threshold; (d) reapplying a tripod operator , as in (a), if the density of the feature vector is not greater than some predetermined threshold, (e) converting the stored data into a compact piecewise analytic description, if the density of the feature vector is greater than some predetermined threshold, after receiving scanned digital representation of an unknown object known as a test range image; (f) applying a tripod operator at a random place on the test range image; (g) matching a feature vector from (f) to the stored data in (b) for the unknown object;
(h) reapplying a tripod operator, as in (f), if the feature vector is not within a predetermined distance of the stored feature data, (i) computing an estimate of six pose parameters of the unknown object, if the feature vector is within a predetermined distance of the stored feature data; and applying the six pose parameters to a utilizing device.
34
PCT/US2004/015580 2004-06-03 2004-06-03 A technique for estimating the pose of surface shapes using tripod operators WO2005122064A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/US2004/015580 WO2005122064A1 (en) 2004-06-03 2004-06-03 A technique for estimating the pose of surface shapes using tripod operators

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/US2004/015580 WO2005122064A1 (en) 2004-06-03 2004-06-03 A technique for estimating the pose of surface shapes using tripod operators

Publications (1)

Publication Number Publication Date
WO2005122064A1 true WO2005122064A1 (en) 2005-12-22

Family

ID=35503286

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2004/015580 WO2005122064A1 (en) 2004-06-03 2004-06-03 A technique for estimating the pose of surface shapes using tripod operators

Country Status (1)

Country Link
WO (1) WO2005122064A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114004808A (en) * 2021-10-29 2022-02-01 上海应用技术大学 Loop detection method based on differential manifold

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6393143B1 (en) * 1999-12-08 2002-05-21 The United States Of America As Represented By The Secretary Of The Navy Technique for estimating the pose of surface shapes using tripod operators

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6393143B1 (en) * 1999-12-08 2002-05-21 The United States Of America As Represented By The Secretary Of The Navy Technique for estimating the pose of surface shapes using tripod operators

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114004808A (en) * 2021-10-29 2022-02-01 上海应用技术大学 Loop detection method based on differential manifold
CN114004808B (en) * 2021-10-29 2024-04-16 上海应用技术大学 Loop detection method based on differential manifold

Similar Documents

Publication Publication Date Title
AU2005200218B2 (en) A technique for estimating the pose of surface shapes using tripod operators
Yang et al. Go-ICP: A globally optimal solution to 3D ICP point-set registration
Thompson et al. Three-dimensional model matching from an unconstrained viewpoint
Barrow et al. Computational vision
Lukács et al. Faithful least-squares fitting of spheres, cylinders, cones and tori for reliable segmentation
EP1694821B1 (en) Probable reconstruction of surfaces in occluded regions by computed symmetry
EP2048599B1 (en) System and method for 3D object recognition
US20060017720A1 (en) System and method for 3D measurement and surface reconstruction
US20030091227A1 (en) 3-D reconstruction engine
Drost et al. Local hough transform for 3d primitive detection
Chen et al. Maximal matching of 3-D points for multiple-object motion estimation
Bülow et al. Scale-free registrations in 3D: 7 degrees of freedom with Fourier Mellin SOFT transforms
Joo et al. Globally optimal inlier set maximization for Atlanta world understanding
US20040234167A1 (en) Technique for estimating the pose of surface shapes using tripod operators
SANDOVAL et al. Robust sphere detection in unorganized 3D point clouds using an efficient Hough voting scheme based on sliding voxels
WO2005122064A1 (en) A technique for estimating the pose of surface shapes using tripod operators
Hlubik et al. Advanced point cloud estimation based on multiple view geometry
Banerjee et al. A low-cost portable 3d laser scanning system with aptness from acquisition to visualization
Krueger Model based object classification and localisation in multiocular images
Booij et al. Efficient probabilistic planar robot motion estimation given pairs of images
Abdelmajid Investigation and comparison of 3D laser scanning software packages
Karami et al. Camera Arrangement in Visual 3D Systems using Iso-disparity Model to Enhance Depth Estimation Accuracy
Klette et al. Surface reconstruction based on visual information
Messer et al. Image-Based Alignment of 3D Scans
Kissi Automatic Fracture Orientation Extraction from SfM Point Clouds

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

122 Ep: pct application non-entry in european phase