WO2003031005A2 - Procede et dispositif de detection et de correction d'une penetration entre des objets - Google Patents

Procede et dispositif de detection et de correction d'une penetration entre des objets Download PDF

Info

Publication number
WO2003031005A2
WO2003031005A2 PCT/US2002/032235 US0232235W WO03031005A2 WO 2003031005 A2 WO2003031005 A2 WO 2003031005A2 US 0232235 W US0232235 W US 0232235W WO 03031005 A2 WO03031005 A2 WO 03031005A2
Authority
WO
WIPO (PCT)
Prior art keywords
act
vector
representation
intersection
primitive
Prior art date
Application number
PCT/US2002/032235
Other languages
English (en)
Other versions
WO2003031005A3 (fr
Inventor
Sriram Krishnan
Sanjay Sarma
Original Assignee
Massachusetts Institute Of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Massachusetts Institute Of Technology filed Critical Massachusetts Institute Of Technology
Publication of WO2003031005A2 publication Critical patent/WO2003031005A2/fr
Publication of WO2003031005A3 publication Critical patent/WO2003031005A3/fr

Links

Classifications

    • A63F13/10
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/45Controlling the progress of the video game
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/016Input arrangements with force or tactile feedback as computer generated output to the user
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T17/00Three dimensional [3D] modelling, e.g. data description of 3D objects
    • G06T17/20Finite element generation, e.g. wire-frame surface description, tesselation
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1037Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals being specially adapted for converting control signals received from the game device into a haptic signal, e.g. using force feedback
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/60Methods for processing data by generating or executing the game program
    • A63F2300/64Methods for processing data by generating or executing the game program for computing dynamical parameters of game objects, e.g. motion determination or computation of frictional forces for a virtual car
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/60Methods for processing data by generating or executing the game program
    • A63F2300/64Methods for processing data by generating or executing the game program for computing dynamical parameters of game objects, e.g. motion determination or computation of frictional forces for a virtual car
    • A63F2300/643Methods for processing data by generating or executing the game program for computing dynamical parameters of game objects, e.g. motion determination or computation of frictional forces for a virtual car by determining the impact between objects, e.g. collision detection
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/60Methods for processing data by generating or executing the game program
    • A63F2300/66Methods for processing data by generating or executing the game program for rendering three dimensional images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2210/00Indexing scheme for image generation or computer graphics
    • G06T2210/21Collision detection, intersection

Definitions

  • This invention relates to the computation of the spatial relationship between two virtual objects having arbitrary shapes and, more particularly, to collision detection and the computation of full-body contact between the two objects.
  • Determining proximity information and computing spatial relationships between virtual objects is a fundamental consideration in many areas of robotics, computer-aided design, computer graphics and virtual reality applications. As virtual objects have no real or physical existence, a virtual object is often characterized by a mathematical and/or geometric representation. As such, simulation of the interactions between virtual objects in a virtual environment is computational in nature.
  • a virtual object is not real, it may have a real counterpart, that is, the representation of the virtual object may describe or model a real object.
  • a robot navigating through a real environment may have an internal representation of that environment such as a mathematical description of itself, where each object is located and what space each object occupies in the environment.
  • models These mathematical descriptions of the objects in the internal representation of the environment, also referred to as models, must explain the physical environment with sufficient accuracy and detail and be responsive to changes in the real environment, such that the robot can navigate and perform the various tasks required of it in the physical environment.
  • the virtual object is sometimes referred to as an avatar.
  • the virtual environment may not have a real or physical counterpart.
  • the objects and environment in the simulation though perhaps modeled to imitate real or physical objects and environments, have no immediate real counterpart.
  • virtual object or simply "object” refers to the representation of the object and not the real object itself.
  • real or “physical” will be used herein when describing objects having a physical existence.
  • the acts involved in mathematically describing, or otherwise modeling virtual objects and environments is referred to generally as "representation.”
  • Such computations may include updating the representation of each virtual object according to motion or movement (e.g., translation, rotation, and/or deformation of the object), detecting collisions and/or at-a-distance forces between virtual objects, determining a response to collision forces, etc.
  • the acts involved in computing relationships and/or interactions between virtual objects in a virtual environment is referred to generally as "simulation.”
  • a simulation is often implemented as a computer program that is executed on one or more processors of a computer and/or computer system.
  • a simulation may be partially guided by input from a user, or otherwise externally instructed.
  • an interface device such as a joystick, mouse, track ball, gimbal, etc.
  • a user may manipulate a joystick to control the movement of a virtual object.
  • the effect of the user input or change enacted on a virtual object through the interface device is computed and rendered as feedback to the user. That is, a user may effect a change in one or more objects in a virtual environment, the representation of the environment may then be updated to reflect the change, interaction between objects caused by the user input detected and simulated, and the results reported back to the user.
  • This type of feedback control loop is commonly used in virtual environments.
  • a simulation is visually rendered to the user. That is, the virtual objects in an environment may be rendered to a display such that a user may observe the effects of his actions through the interface device.
  • a simulation or portions of a simulation may also be haptically rendered.
  • the term "haptic" refers to the sense of touch.
  • a haptic rendering refers to stimuli that is communicated through feel or touch sensations rather than through the more common use of visual stimuli.
  • a virtual environment may employ both haptic and visual renderings of the events in a simulation. Haptic sensations may be communicated, for example, as a force-feedback applied to the interface device manipulated by the user.
  • collision detection is often part of the computations involved in modifying a virtual environment in response to, for example, a user enacted movement, or in updating an environment to ensure that the proper state of the environment is being reported (i.e., updating an environment to ensure that feedback renderings accurately reflect the environment and the objects within the environment) to the user.
  • One embodiment according to the present invention includes a method of determining an extent of an intersection of a first object and a second object.
  • the method comprises act of constructing a first representation of the first object and a second representation of the second object, the second representation defining at least a surface of the second object, determining at least an original Noronoi diagram of the second representation, detecting at least one intersection between the first representation and the second representation, and determining a distance from the at least one point of intersection to the surface of the second object based on a location of the at least one point of intersection within the original Voronoi diagram.
  • Another embodiment according to the present invention includes a method of simulating an interaction between a controlled object and an independent object with six degrees of freedom in a virtual environment including a haptic interface device having a controlled portion moveable with six degrees of freedom.
  • the method comprising acts of receiving control signals from the interface device, the control signals including pose information with six degrees of freedom, updating a first representation of the controlled object according to the pose information, determining a presence or an absence of intersection between the first representation of the controlled object and a second representation of the independent object, and computing at least one rendering signal related to the presence or absence of intersection between the first representation and the second representation.
  • Another embodiment according to the present invention includes a method in a virtual environment of providing full-body contact between a first object and a second object.
  • the method comprising acts of representing a first object by a first representation including a first set of primitives and a second object by a second representation including a second set of primitives, detecting a presence or an absence of an intersection between each of the first set of primitives and the second set of primitives, determining a penetration vector for each present intersection, determining a penetration vector for each present intersection, and computing at least one synthesized vector from the at least one penetration vector, the at least one synthesized vector having at least one translational force component and at least one rotational force component.
  • Another embodiment according to the present invention includes a computer readable medium encoded with a program for execution on at least one processor, the program, when executed on the at least one processor, performing method of determimng an extent of an intersection between a first object and a second object.
  • the method comprises acts of constructing a first representation of the first object and a second representation of the second object, the second representation defining at least a surface of the second object, determining at least an original Noronoi diagram of the second representation, detecting at least one intersection between the first representation and the second representation, and determining a distance from the at least one point of intersection to the surface of the second object based on a location of the at least one point of intersection within the original Noronoi diagram.
  • Another embodiment according to the present invention includes a computer readable medium encoded with a program for execution on at least one processor, the program, when executed on the at least one processor, performing, in a virtual environment including a haptic interface device having a controlled portion moveable with six degrees of freedom, a method of simulating an interaction between a controlled object and an independent object with six degrees of freedom.
  • the method comprises acts of receiving control signals from the interface device, the control signals including pose information with six degrees of freedom, updating a first representation of the controlled object according to the pose information, determining a presence or an absence of intersection between the first representation of the controlled object and a second representation of the independent object, and computing at least one rendering signal related to the presence or absence of intersection between the first representation and the second representation.
  • Another embodiment according to the present invention includes a computer readable medium encoded with a program for execution on at least one processor, the program, when executed on the at least one processor, performing a method of providing full-body contact between a first object and a second object.
  • the method comprises acts of representing a first obj ect by a first representation including a first set of primitives and a second object by a second representation including a second set of primitives, detecting a presence or an absence of an intersection between each of the first set of primitives and the second set of primitives, determining a penetration vector for each present intersection, and computing at least one synthesized vector from the at least one penetration vector, the at least one synthesized vector having at least one translational force component and at least one rotational force component.
  • FIG. 1 illustrates a block diagram of a control feedback loop
  • Fig. 2 illustrates a simulation according to the one embodiment of the present invention
  • Fig. 3 illustrates a block diagram of a bounding volume hierarchy
  • Fig. 4 illustrates a several spatial relationships between virtual objects that may arise during simulations according to one embodiment of the present invention
  • Fig. 5A illustrates a false-negative intersection condition that may arise from loss of depth information
  • Fig. 5B illustrates a solution to the false-negative intersection condition according to one embodiment of the present invention
  • Fig. 6 illustrates an original Voronoi diagram of a virtual object according to one embodiment of the present invention
  • Fig. 7 illustrates a method of determining an extent of an intersection according to one embodiment of the present invention
  • Fig. 8 illustrates a method of collision detection between two virtual objects in a haptic environment according to the present invention.
  • systems that employ rendering as part of a control feedback loop may be vulnerable to loss of synchronization should the simulation fail to observe real- time constraints when computing and generating feedback signals.
  • This loss of synchronized cause and effect can be very disorienting to a user, and may cause the user to make certain erroneous and/or unintentional actions. This may be particularly deleterious in cases where the virtual environment has an immediate physical counterpart.
  • the computational complexity of a simulation is considered closely in order to ensure that real-time constraints are met.
  • Fig. 1 illustrates a control feedback loop that may be employed in virtual reality applications.
  • the control feedback loop allows a user or operator to provide input to a simulation to effect change in a virtual environment and to receive back from the simulation the results of such an action.
  • the components of control feedback loop 15 include a user 10, an interface device 200, and simulation 100.
  • User 10 is often a human operator or participant in the virtual environment.
  • the user may control some portion of the virtual environment via interface device 200.
  • the user manipulates a moveable portion of interface device 200, for example, the handle of a joystick in order to direct the movement of one or more controlled objects.
  • the term "controlled object” or “controlled portion” refers to a virtual object or a portion of a virtual environment that is under direct control of a user or otherwise directly controlled external to the simulation.
  • Interface device 200 may include sensors, electromechanical structures, or any of other various means for tracking the user's motions and translating the motions into control signals 205.
  • Control signals 205 are provided to simulation 100 as instructions on how the controlled portion and/or controlled objects of the virtual environment should be modified and/or updated.
  • the changes in the virtual environment effected by control signals 205 generated by the movements of user 10 may cause the virtual objects in the environment to interact. For instance, the actions of the user may bring the controlled portion of the environment into contact with other portions and/or objects in the environment. As such, the simulation may compute these interactions and report the results back to the user.
  • the acts involved in reporting results of a simulation back to the user are referred to as rendering. Visual and auditory renderings have long been used to notify a user about the state of a virtual environment.
  • Haptic rendering is now recognized by the virtual reality community and others as a valuable method of articulating information about a virtual environment where visual and/or auditory information is insufficient, where tactile information may provide a more complete understanding of a virtual environment, or to supplement other rendering modes to further enhance a virtual experience.
  • a control feedback loop will often exploit more than one rendering technique in order to provide comprehensive information about an environment.
  • simulation 100 computes rendering signals 145 and provides them to the user as stimuli so that the user can interpret and comprehend the virtual environment.
  • a simulation may visually render the environment to a display 147 that the user can view, may haptically render the results of the interaction to the interface device to articulate feel sensations to the user, or both.
  • the control loop may include other types of rendering modes, for example, auditory rendering signals.
  • the feedback loop is not maintained in real time, the virtual environment becomes confusing and may lead to misinterpretation of the environment or to more deleterious effects such as loss of control by the user.
  • most virtual reality simulations are computationally expensive. As such, conventional simulation methods have relied upon certain assumptions about the virtual environment in order to meet the real time constraints of the application.
  • conventional simulations may rely upon various assumptions about the shape of a virtual object.
  • conventional virtual objects may be limited to convex shapes.
  • assumptions about shape symmetry, continuity, and complexity are often relied upon in order to reduce the computational requirements of simulating virtual objects.
  • such conventional methods are not generic, that is, they do not support arbitrary shapes.
  • limiting the shape of an object is often a prohibitive restriction.
  • An important and often times computationally complex component of a simulation is collision detection.
  • Collision detection refers generally to computation involved in determimng the spatial relationship between two or more virtual objects so as to determine whether two or more objects have or will come into contact with one another.
  • contact between objects often must be computed on the order of 1000 Hz or higher. As such, the expense of such computations may determine whether a real-time constraint is met.
  • One embodiment according to the present invention includes a method of collision detection supporting arbitrary shapes, specifically, a method independent of the convexity and/or symmetry of the virtual objects being simulated.
  • Point contact refers to collision detection methods wherein only a single point, a small group of points, or a simplified region is tested for collision and/or contact. Stated differently, in point contact simulations, a virtual object may have a designated active region (often a single point) in which collisions are considered and contact forces computed. Collisions occurring elsewhere on the body of the object are ignored, that is, they are not computed. Point-contact methods allow for the number of collision computations to be greatly reduced, often times to a single intersection and force calculation.
  • point contact collision detection may not provide adequate information about interactions between two objects.
  • point contact force feedback may not feel authentic and may be inappropriate for applications that require more realistic and comprehensive feel sensations.
  • point contact information may not be able to articulate the complex and subtle sensations required by a surgeon to properly negotiate and comprehend the virtual environment with, for instance, a virtual scalpel or probe.
  • a virtual object representing a milling tool is controlled by a haptic interface device to remove portions of a surface to be milled.
  • the force applied to the tool may effect how much material is removed from the surface, the depth of a cut, the torque on the tool, etc.
  • contact computations report back realistic and accurate haptic information such that the operator is aware of how much pressure is being applied, and what part of the tool is in contact with the milling surface.
  • Point contact information may not report back forces that arise everywhere on the body of a virtual object and may be too limited to describe higher order forces such as torque, shear, etc.
  • another embodiment of the present invention includes a method of providing full-body contact collision detection and force calculation.
  • Full-body contact refers generally to collision detection in which each part or portion of a virtual object is considered when determining if collision and/or contact has occurred.
  • a simulation is often divided logically into three interrelated categories; representation, interaction, and rendering. The methods applied to each category and how they interrelate may impact how general a simulation may be and how complex and expensive the computation.
  • the number of degrees of freedom refers to the number of dimensions for which an object is permitted to arbitrarily vary, and is typically related to the number of variable parameters that describe the pose of an object.
  • the "pose" of an object refers to the set of parameters required to describe an object's location in space and, more particularly, the parameters sufficient to locate an object such that the space or portion of the space an object occupies may be computed from its representation.
  • a rigid body may be located in a three dimensional (3-D) space defined by a standard Cartesian coordinate system by six parameters, specifically, by its deviation from the origin of the coordinate system along each of the three axes (e.g., the X, Y, and Z axes) of the coordinate system.
  • the translational deviation from and the rotational deviation about each axis of the coordinate system needs to be specified.
  • an object simulated with six degrees of freedom would permit all six pose parameters to vary arbitrarily.
  • the interface device in a control feedback loop may limit the number of parameters that are allowed to be controlled by the user.
  • a virtual object may be simulated with six degrees of freedom while the interface device is limited to some reduced number of degrees of freedom.
  • the interface device is limited to some reduced number of degrees of freedom.
  • the user may only have direct control over one dimension (i.e., the rotation of the axle of the car), that is, the control signals from the interface device will only contain information related to rotational deviation of the axle about the axis of the car. However, this information may be received by a simulation to determine the location of the representation of the virtual car with six degrees of freedom.
  • the input from the user i.e., one dimensional information regarding the rotation of the axle
  • the input from the user may be used in conjunction with velocity information to compute the car's position on the surface that the car is being driven. Additionally, this new position may lead the car up a ramp or send the car into collision with other objects.
  • These actions may result in the virtual car being simulated in a jump, flip or roll wherein the representation of the car varies with six degrees of freedom.
  • the interface device itself only provides the simulation with control signals having a variable component in a single dimension, that is, the interface device operates with one of freedom.
  • the simulation is computed with fewer degrees of freedom than provided by the interface device.
  • the classic video game Pong in which the "paddle" was only free to move in one direction. Since the paddle was constrained in all other directions, the paddle could be located simply by tracking its deviation along a single axis.
  • the paddle in pong was simulated with one degree of freedom.
  • the paddle in Pong was often controlled by a joystick or trackball capable of providing control signals with more than one degree of freedom. In such a situation, the simulation simply ignores the control signals relating to movement in the unused dimension or dimensions, that is, the dimensions constrained by the simulation.
  • some simulations exploit assumptions about the virtual objects or about the nature of the simulation in order to reduce the number of parameters required to locate an object, thereby reducing the computational complexity of the simulation. For example, assuming that an object is axially symmetric allows rotation about the z- axis to be disregarded since the space occupied by an axially symmetric object will be identical for all rotations about the axis of symmetry. Other assumptions about the simulation may also reduce the dimensionality of locating an object.
  • Applicants have developed and implemented a simulation for use with an interface device capable of providing control signals having variable components in six dimensions, that is, the interface device and the simulation are capable of operating with six degrees of freedom. According to one embodiment of the present invention, the simulation operates without compromising the real-time constraint.
  • Fig. 2 illustrates an embodiment of a simulation according to the present invention.
  • Simulation 100' includes representation 120, interaction 130, and renderer 140.
  • Representation 120 includes a model of the virtual environment and, more particularly, the mathematical and/or geometric description of the objects in the environment.
  • the virtual environment is illustrated as including six virtual objects 20a-20f, at least one of which is controlled by the interface device.
  • the choice of representation may impact the type of shapes that can be modeled and may influence the computational complexity, and ultimately the speed and robustness of the simulation.
  • Virtual objects in an environment may be modeled in a variety of different ways, that is, there are numerous and varied representations of an object.
  • Representations include implicit equations, algebraic and parametric representations, splines, etc. Each representation may have characteristics that benefit a particular application. For instance, implicit representations have well defined algebraic equations making them well suited for fast collision detection due to simple and rapid intersection tests. However, the use of an implicit representation may prevent the simulation from processing arbitrary shapes. That is, the shape of the object typically must be known a priori in order to generate an implicit representation of the object.
  • Another method of object representation includes modeling an object as a set of primitives. The term "primitive" describes any geometric element used as a basis for constructing a shape or object.
  • typical primitives include, but are not limited to, points, lines, polygons, etc.
  • a virtual object may be represented as a point cloud, polygon soup, polygonal surface or mesh, etc.
  • Such primitive-based approaches often describe an object by defining the outer boundary of the object, referred to as a surface representation.
  • objects 20a-20f may be represented as a set of interconnected polygons on the surface of the virtual object, often referred to as a polygonal surface or mesh.
  • each facet will share an edge with at least one other facet.
  • the simplest polygonal surface comprises triangular facets and is referred to as a triangulated surface or, alternatively, a triangulated mesh.
  • object 20a is illustrated in bold to indicate that it is a controlled object manipulated by the user via interface device 200.
  • the movement of object 20a may be computed according to the information provided by control signals 205'.
  • the pose of the virtual object 20a is updated such that it is consistent with the input from the interface device.
  • a user manipulates the interface device in order to direct the movements of object 20a.
  • the control signals generated by the interface device are provided to the simulation and the pose of object 20a is updated.
  • the controls signals include variable components in six dimensions and the pose of object 20a is updated according to each of the six dimensions, that is, object 20a is simulated with six degrees of freedom.
  • the updated pose of object 20a may cause an interaction with one or more of the other objects in the virtual environment.
  • interaction 130 may compute the spatial relationship between the objects in the virtual environment in view of the updated pose of object 20a.
  • interaction 130 may perform collision detection between object 20a and the other objects in the virtual environment (e.g., objects 20b-20g).
  • collisions in a virtual environment are mathematical concepts in that virtual objects have no real physical structure with which to collide or create contact.
  • one method of detecting collisions between two virtual objects is to intersect the representations of the virtual objects.
  • the term "intersection” refers to the computations involved in determining whether the representations of two virtual objects occupy space in common. However, in order to make this determination, the space or a portion of the space that each virtual object occupies must be determined.
  • surface representations e.g., a triangulated surface representation
  • implicit representations information regarding points on the interior and on the exterior of an object are not readily available.
  • Full-body contact refers to computing collisions and/or contacts between two virtual objects wherein the entire representation of the object is considered. That is, full-body intersection is computed and contact forces related to the full-body intersection are then determined.
  • Applicant has developed and implemented a method of collision detection b _ ⁇ etw spicye ⁇ enn virtual objects that supports arbitrary shapes and provides full-body contact.
  • One embodiment according to the present invention includes constructing a bounding hierarchy for objects considered for collision detection in a simulation.
  • a bounding element hierarchy is a structure that encloses a core representation of a virtual object using a simpler representation and/or a fewer number of primitives.
  • the term "core representation" generally refers to the representation of an object that is the least approximate of the underlying shape, that is, the representation having the finest level of resolution.
  • an element is used synonymously with primitive.
  • an element may be a primitive or a collection of primitives.
  • the term element will be used in reference to bounding hierarchies and primitives for core representations for clarity when describing the two types of representations in relation to one another.
  • the two terms do not refer to fundamentally different things.
  • bounding element hierarchies bounding elements are often computationally simpler than the core representation in that they may have known orientations and comprise primitives having known relationships to one another.
  • a bounding element hierarchy forms a progressively refined representation of the virtual object that may be traversed as a binary tree, thus employing the principles and enjoying the complexity reduction of a binary search.
  • a bounding element hierarchy is a multi-resolution object representation.
  • the root level of the hierarchy is typically at the coarsest resolution, that is, the root level is the most imprecise approximation to the underlying shape of an object.
  • the representation is at its highest resolution, that is, it most closely approximates the underlying shape of the object.
  • the underlying shape of an object has a core representation and the bounding hierarchy is built around this core representation and progressively refined down the hierarchy from root to leaf.
  • FIG. 3 illustrates a two-dimensional example of a bounding element hierarchy for an object 20'.
  • Object 20' may be represented, for instance, as a series of connected lines.
  • Boundary hierarchy 50 illustrates four levels 50a-50d of a progressively refined bounding element hierarchy of object 20' .
  • object 20' is enclosed in a single rectangle (bounding element).
  • bounding element As shown, the bounding rectangle 60 completely encloses all eight lines of the core representation of object 20'.
  • the space occupied by the bounding element is a superset of the one or more primitives it encloses.
  • two bounding rectangles are used to enclose object 20'.
  • each bounding rectangle need only enclose a subset, for example, half of the lines comprising object 20' .
  • the bounding hierarchy at the second level approximates the core representation more accurately than at the root level.
  • level 50c of the bounding element hierarchy the representation has been further refined such that each rectangle encompasses only two lines of shape 20'.
  • each bounding rectangle encloses a single primitive (e.g., a line) of the core representation.
  • the hierarchy in Fig. 3 is organized as a binary tree.
  • the root level comprises a single element 60.
  • Element 60 has two children, specifically, element 60a and 60b at the second level of the hierarchy.
  • element 60a has two children 60aa and 60ab
  • element 60b has two children 60ba and 60bb.
  • the structure is similarly refined in such a manner until the leaf level is reached (i.e., a bounding element encloses a single primitive).
  • bounding element hierarchy The leaf level of the hierarchy is the best approximation available to bounding hierarchy 50.
  • other types of bounding elements may be used in a bounding element hierarchy such that the approximation may be more accurate.
  • the bounding elements in hierarchy 50 are Axis Aligned Bounding Boxes (AABBs).
  • AABBs Axis Aligned Bounding Boxes
  • the possible orientations of AABBs are constrained such that they align with one of the axes of the Cartesian space of the environment being modeled.
  • each line forming part of a bounding element in hierarchy 50 either aligns with the x-axis or the y-axis.
  • a bounding hierarchy is formed wherein the bounding elements are allowed orientations in k different directions. This allows a bounding element to more closely approximate the core representation but still retain computational simplicity by having a known number of possible orientations (i.e., k possible orientations).
  • Fig. 4 illustrates a method of traversing a bounding element hierarchy by considering three different spatial relationships between an a pair of objects. Spatial relationships 70a-70c illustrate three arrangements that may occur between two objects 20A and 20B during a simulation. Each spatial relationship is illustrated along with the level of the hierarchy that must be traversed in order to determined if the objects intersect.
  • the objects do not intersect. Since the objects will not intersect if their bounding elements do not intersect, the single computation of intersecting the root level bounding elements 80 and 82 is sufficient to determine that the objects do not intersect.
  • the objects again do not intersect.
  • the root level bounding elements 80 and 82 are a coarse approximation to the core representation, the root levels of the hierarchies in fact intersect.
  • the next level of the hierarchy may need to be tested to determine if the objects intersect.
  • the second level of the hierarchy enclosing primitives of object 2 A includes bounding elements 80a and 80b.
  • the hierarchy enclosing primitives of object 20B includes bounding elements 82a and 82b. In spatial relationship 70b, it is sufficient to test the second level of the hierarchies for intersection to establish that the objects are non-intersecting.
  • the core representations intersect, that is, the virtual objects have intersected one another. Because an intersection occurs at the finest resolution available (i.e., the core representation), at least one path from the root of the hierarchy to one or more leaves of the hierarchy will be traversed.
  • a bounding element encloses a single primitive of the core representation. For clarity, only the leaf level bounding elements detected as intersecting are depicted. As shown, two leaves of the bounding element hierarchy of object 20A are found to be intersecting two respective leaves of the bounding element hierarchy of object 20B, namely, element 80bd intersects element 82ad and element 80bc intersects element 82aa.
  • Traversal of a bounding hierarchy either terminates when all intersection tests at a given level of the hierarchy are false or completes when each bounding element in the leaf level of the hierarchies has been tested for intersection. If one or more of the leaf level bounding elements intersect, then at least some portion of the core representations may intersect.
  • leaf level of a bounding volume hierarchy need not enclose a single primitive of the core representation.
  • the resolution required by the leaf level may be chosen to satisfy the requirements of a particular implementation.
  • a bounding element hierarchy may be structured ways other than a binary tree. For example, each bounding element at a level of the hierarchy may give rise to more than two children.
  • the interior of a surface representation is not easily obtained when the representation of an object is allowed to be arbitrarily complex. For example, if the number of primitives in a core representation is allowed to be arbitrarily large, the computation involved in determining if a point lies on the interior of the object may compromise real-time performance.
  • a bounding element is typically simpler and has fewer primitives than the core representation, the interior of a bounding element is often known or easily obtained.
  • a bounding element may inherently define part of the interior of the core representation by being a superset of the space occupied by the one or more primitives it encloses. For example, in relationship 70a of Fig. 4, the root level bounding element defines the entire interior of the object. Since, the interior of the bounding element is trivially obtained, so is the interior of the object.
  • the bounding elements are mutually defined by the bounding elements. As shown in the leaf level of hierarchy 50 in Fig. 3, there is a region in the interior of object 20' that is not defined by the bounding elements. At levels in the hierarchy near the leaves, the bounding element very closely approximates the core representation. As a bounding element approximates the underlying representation arbitrarily close, the interior of the object defined by the bounding element approaches zero. Stated differently, the closer the bounding volume approximates the surface of an object, the less information is known about the object in directions normal to the surface of the object, that is, there is a loss of depth information.
  • FIG. 5A illustrates such a problem that may arise from loss of depth information.
  • Fig. 5 A illustrates a first object 20A' having an associated bounding element hierarchy. The leaf level representation of the associated bounding element hierarchy is shown. In this example, the bounding elements are AABB.
  • a second object 20B' has a point cloud core representation and an associated bounding element hierarchy of which the leaf level is shown.
  • point 9 is enclosed by bounding element 82a' .
  • Point 9 is an intersecting point, that is, it lies on the interior of object 20A'.
  • primitive 22a' of the representation of object 20A' is near axis-aligned and therefore the leaf level bounding element which is axis-aligned very closely approximates line 22a'.
  • the bounding element 80a' provides only minimal depth information regarding the interior of the object in a direction normal to line 22a'. In the degenerate case where the line is exactly axis-aligned, all depth information is lost.
  • the intersection test between leaf level bounding element 80a' of object A and leaf level bounding element 82a' of object B fails and the objects are erroneously determined to be non-intersecting.
  • Applicants have developed a method of adding depth information into the bounding representation to avoid false-negative intersection detection cases as illustrated in Fig. 5 A.
  • Applicants employ an offset surface in order to incorporate depth information about the interior of an object.
  • the offset surface insures that at least a desired region of the interior of a surface representation in directions normal to the surface will be defined by the associated bounding element hierarchy.
  • An offset surface is described as the locus of points at a signed distance from a progenitor surface along surface normals of the progenitor surface. While surface normals for continuous representations such as algebraic or parametric functions are well understood, surface normals for discrete surfaces may be more complicated to compute and may only be approximate. For example, normals to a triangulated surface may be computed in more than one way.
  • surface normals may be defined at the vertices of the triangles by "averaging" the planar normals of the facets that share the vertices.
  • One such averaging method includes computing a localized quadratic surface approximation at each vertex, and taking the normal to the quadratic surface at the vertex.
  • Another method includes offsetting the planes along the normal to each facet of a polygonal surface, for instance, a triangulated surface.
  • the original surface representation is referred to as the progenitor surface.
  • the offset surface may be calculated, for example, by calculating the normal to each facet of a polygonal surface and defining a new plane at some desired distance (i.e., the desired extent of depth information) along the normal.
  • the offset planes may then be intersected to establish the facets of the offset surface.
  • Fig. 5B illustrates an object wherein an offset surface has been computed by using this approach. As shown in Fig. 5B, an offset surface 44 is shown at an offset distance d from progenitor surface 42 along the normal to each line of the representation of object 20 A.
  • An associated bounding element hierarchy may then be constructed such that the bounding elements enclose the progenitor and offset surfaces, thus retaining depth information related to the extent of the offset. That is, the leaf level bounding elements carry information about the interior of object 20 A' to the extent of the offset distance d. The distance d may then be chosen to satisfy the requirements of a particular implementation. As such, the depth information prevents point 9 from being mischaracterized during intersection detection. The bounding element enclosing point 9 is now within bounding element 22a' ' and will therefore test positive as an intersecting point.
  • a synthesized vector may include components describing the translation force and the rotational force on the virtual objects resulting from the full-body contact.
  • a synthesized vector may include three components describing the translation force on a virtual object and three components describing the moment or torque on the virtual object.
  • the extent of an intersection between two virtual objects may be used to simulate the forces that would be experienced by physical world objects.
  • the extent of intersection between two colliding virtual objects may be related to the forces that resist and/or prevent two physical objects from intersecting in the physical world.
  • the extent of an intersection in the virtual domain may be used to measure the forces that would have arisen in the physical domain to prevent such an intersection.
  • the depth of penetration is calculated for each primitive of an intersecting virtual object. Since bounding elements only approximate the core representation, the traversal of a bounding volume hierarchy only determines whether an intersection is present, but does not describe the precise location of the intersection or, ultimately, the extent of the intersection. Moreover, bounding elements enclosing both a progenitor and an offset surface will have regions where adjacent bounding elements overlap, creating possible ambiguity as to precisely where an intersecting primitive intersected the surface of the intersected object.
  • intersecting and intersected object will be used to distinguish two objects being tested for intersection. While either object can be considered as the intersecting or intersected object, when appropriate, the controlled object will be referred to as the intersecting object.
  • One method of determining the extent of an intersection between virtual objects is to compute the location of each primitive testing positively for intersection during traversal of the bounding volume liierarchy with respect to the surface of the intersected object.
  • the closest primitive on the surface of the intersected object to an intersecting primitive may not be known.
  • proximity information available in a Voronoi diagram representation of the intersected object may be utilized to rapidly identify the primitive on the surface of the intersected object nearest detected intersection.
  • a Voronoi diagram is defined as the locus of points having a minimum distance equidistant to at least two primitives and whose distance to all other primitives is greater than said minimum distance.
  • Voronoi diagram may be referred to as the symmetry set, medial axis, Theissen or Dirichlet tessellations, Delaunay triangulation, etc. While these structures may contain other information about a shape or object, they all contain at least the information in a Voronoi diagram as defined above.
  • FIG. 6 illustrates a two-dimensional closed shape and its associated Voronoi diagram.
  • Shape 20 is formed from a set of connected primitives (e.g., lines) 22a-22m giving rise to Voronoi diagram 30.
  • the Voronoi diagram consists of the primitives on the interior of shape 20.
  • a Voronoi diagram will comprise a set of primitives that depend on the primitives constituting the shape from which the Voronoi diagram is computed.
  • the Voronoi primitives are a set of lines that form the boundaries of a plurality of Voronoi regions.
  • a number of Voronoi primitives are labeled in diagram 30 (e.g., lines 32a-32g).
  • a Voronoi region describes the locus of points having a minimum distance to the same geometric primitive (i.e., all points sharing a closest primitive of the shape or object). For example, every point within Voronoi region 40 is closer to primitive 22i than any other primitive in the set of primitives defining shape 20.
  • each Voronoi region forms a polygon bounded by at least two original Voronoi primitives of the Voronoi diagram.
  • An original Voronoi primitive refers to a primitive in the Voronoi diagram that includes a point having a globally minimum distance from its at least two equidistant primitives.
  • the Voronoi primitives that arise out of the connections between the lines forming object 20 are original Voronoi primitives.
  • Voronoi primitives arise from the intersection of two or more original Voronoi primitives and/or the intersection of two or more original and secondary Voronoi primitives. For clarity, not all of the Voronoi primitives and Voronoi regions are labeled in Fig. 6.
  • Voronoi primitives 32a-32h illustrate several original Voronoi primitives of Voronoi diagram 30.
  • Voronoi primitives 34a-34d illustrate several secondary Voronoi primitives.
  • Voronoi region 40 for example, is bounded by original Voronoi primitives 32f and 32g and secondary Voronoi primitives 34c and 34d.
  • An original Voronoi diagram refers to a Voronoi diagram wherein only original Voronoi primitives are computed. As such, the myriad of intersections that may occur between original and/or secondary Voronoi primitives in the interior of a shape need not be considered.
  • a Voronoi region as defined, is guaranteed to be valid only to a distance defined by the least distance intersection between two original Voronoi primitives. This is because, in an original Voronoi diagram, events that occur beyond this least distance intersection are not computed. The least distance intersection
  • the closest primitive on the surface of object 20 can be determined for any point on the interior of object 20 that lies at a depth less than the valid boundary line
  • the original Voronoi primitives are the angle bisectors between adjacent primitives of shape 20. Accordingly, they may be trivially computed. The extension of the Voronoi diagram into three-dimensional space may be easier stated than depicted.
  • shape 20 may be, for instance, a three dimensional object represented as a polygonal surface. Accordingly, the Voronoi primitives of the Voronoi diagram for such a representation would be planes rather than lines.
  • the original Voronoi planes are the angle bisectors between adjacent facets of the polygonal surface. In a closed shape, for example, the original Voronoi planes would bisect the angle between polygonal facets sharing an edge.
  • the Voronoi region of a facet in an original Voronoi diagram would be the volume bounded by the Voronoi planes associated with each of the edges of a facet. Therefore, the closest facet of an arbitrary point can be determined by computing which Voronoi region an arbitrary point occupies, that is, by determining if the point is bounded by the original Voronoi planes of the facet of the polygonal surface.
  • the types of primitives used to represent a shape are not limited to lines, polygons, and/or triangles as illustrated in FIG. 6 and discussed above, nor do all of the primitives have to be of the same type.
  • the original Voronoi primitive of a line and a point distinct from the line is a parabola.
  • the original Voronoi primitive arising from the three-dimensional counterpart i.e., a plane and a point not in the plane
  • a Voronoi diagram may be computed from any combination, arrangement and type of a set of primitives forming the boundary of an object or shape. As such, the Voronoi calculation makes no assumption about the representation and may be computed for arbitrary shapes.
  • the Voronoi diagram of an object can be used to determine the closest primitive on the surface of an object relative to a point of intersection.
  • the depth of penetration of an intersecting primitive may be computed by determining the distance between the intersecting primitive and the closest primitive on the surface of the intersected object.
  • the depth of penetration is calculated for each primitive of an intersecting object that intersects an intersected object.
  • the depth of penetration refers to a measure related to the distance between a portion of an intersecting object and the surface of the intersected object.
  • a depth of penetration may be the distance from an intersecting primitive to the nearest primitive of the intersected object.
  • the depth of penetration may be the area (2-D) or volume (3-D) of an intersection between an intersecting primitive and the nearest primitive of the intersected object.
  • Fig. 7 illustrates one method of computing full-body contact by determimng the extent of an intersection by computing a depth of penetration at each instance of an intersection between two virtual objects.
  • the measure of a depth of penetration at each instance is the minimum distance between an intersecting primitive of an intersecting object to the closest primitive on the surface of the intersected object.
  • the core representation of the intersecting object 20B' is a point cloud
  • the core representation of the intersected object 20A' is a set of closed connected lines.
  • the distance from the intersecting point to the nearest primitive may be calculated to determine a depth of penetration of the intersecting point.
  • the distance of the point from the nearest intersected primitive may be determined in any number of ways.
  • the perpendicular distance for example, may be used as a measure of the depth of penetration.
  • the depth of penetration may be viewed as a penetration vector whose magnitude is the distance between the intersecting primitive and the nearest primitive and whose direction is the direction, for example, along which the distance was calculated.
  • the direction of the penetration vector indicates a direction in which the primitive should move in order to reduce the magnitude of the penetration.
  • On method includes computing the perpendicular distance from the nearest primitive such that the penetration vector has direction indicating a least distance correction of the penetration.
  • perpendicular distance vectors have been computed for each of the intersecting points 14a-14f.
  • the penetration vectors thus calculated describe the extent of the intersection between intersecting object 20B' and intersected object 20A'. It may be desirable to synthesize this set of penetration vectors into a synthesized vector. For example, in the case of a haptic feedback system, it may be desirable to report the results of intersection back to a user such that the full-body contact between the two objects is rendered. Moreover, it may be important for the simulation to update the pose of the objects based on a synthesized vector. That is, since intersection is a concept in the virtual domain, in order to effectively simulate the physical domain, an intersection may need to be corrected and/or eliminated.
  • the synthesized vector may be used to indicate the direction the intersecting object will need to move in order to correct the penetration.
  • One method of computing a synthesized vector from a set of penetration vectors is to take the weighted sum of the penetration vectors.
  • An alternative method of determining the extent of intersection is to model the intersection as an impulse. This concept suggests that the furthest extent of an intersection should guide the correction. Accordingly, one method of determining a synthesized vector is to base the synthesized vector on the penetration vector having the greatest magnitude. Other methods of determining a synthesized vector will occur to those skilled in the art and are considered to be within the scope of the invention.
  • a location of application refers to a location in space where a vector is to be applied.
  • the location of application is in reference to a virtual object, the controlled portion of a haptic interface device, or both. For instance, if full-body contact is to be accurately rendered, a location of application for a force vector included in rendering signals provided to a haptic interface device may be desirable, particularly haptic interface devices operating with more than three degrees of freedom.
  • a force vector applied to the center of mass of a controlled portion of a haptic interface device will feel different to a user than the same force vector applied off the center of mass in that the latter tends to cause the controlled portion of the haptic interface device to rotate.
  • the location of application may be important in articulating particular sensations such as torque.
  • the location of application may be determined in a number of ways.
  • One method includes taking the centroid of the region of intersection between the two objects.
  • a weighted sum of the locations of each of the intersecting primitives may be computed in order to determine a location of application.
  • the location of application may be related to the location of an intersecting primitive giving rise to a penetration vector having the greatest magnitude.
  • Other methods of determining a location of application will occur to those skilled in the art and are considered to be within the scope of the invention.
  • a synthesized vector may incorporate the rotational force on a virtual object by storing components related to the torque or moment of a virtual object.
  • a moment may be calculated from the penetration vectors and incorporated into the synthesized vector.
  • a synthesized vector may include three components related to the translational force on a virtual object and three components related to the rotational force on a virtual object.
  • the moment may be calculated, for instance, in relation to the center of mass of the virtual object or in relation to the center of mass of the controlled portion of a haptic interface device.
  • a synthesized vector may be used as a measure as to how the objects would have behaved in the physical domain. As such, the synthesized vector may be used to compute a new pose of the virtual objects. In addition, objects may deform upon collision. As such, the set of penetration vectors may be used to deform or modify the representations of one or both of the colliding objects.
  • the penetration vectors may describe the portion of the milling surface that is removed by the tool according to such a contact.
  • the synthesized vector and a location of application may be used to determine a rebound or reflection force.
  • the characteristics of the virtual objects that is, the characteristics of the physical objects being simulated may guide the way in which the penetration vectors and/or the synthesized vector is utilized and how the penetration vectors are synthesized.
  • FIG. 7 illustrates one embodiment according to the present invention including a method of providing three dimensional real-time collision detection, full-body contact of arbitrary shapes with six degrees of freedom in a simulation associated with a haptic interface device. In this embodiment the intersection of two virtual objects is considered.
  • the method of collision detection described below may be repeated for any number of objects in an environment for which it is desired to determine an intersection.
  • the first object is controlled by a user manipulating interface device 200' ' .
  • Interface device 200 is capable of tracking the movements of the user with six degrees of freedom and provides control signals 205" having information related to movement in each of the six dimensions over which the device is free to vary.
  • the first object will be referred to interchangeably as the controlled object and the intersecting object.
  • the second object will be referred to as the independent or intersected object.
  • An independent object is a virtual object that is not directly controlled by a user via an interface device. It should be appreciated that an independent object may be indirectly controlled by user, for example, by manipulating the controlled object such that a collision or contact occurs with the independent object. However, an independent object typically will not be updated according to control signals from the interface device, but rather as a result of interaction computations.
  • controlled object 500 and independent object 600 have triangulated surface representations.
  • the surface representation comprises a set of primitives, namely triangles.
  • the triangles may be stored as the vertices of the triangles.
  • the vertex representation may include information about edges, adjacent triangles etc. It should be appreciated that by storing the vertices (i.e., three dimensional points) of the triangulated surface, a point cloud representation is necessarily implied. That is, a triangulated surface can be thought of as including a point cloud representation with additional information, for example, the connectedness of points, adjacency information, etc.
  • controlled object 500 and independent object 600 also have point cloud core representations.
  • the object representations are generated and stored, for example, in a computer readable medium such as a memory.
  • Object representations are often stored in data abstractions or structures defined by the simulation designer. Data abstractions include, but are not limited to, Java classes, C structs, C++ classes, etc. Any method of generating, storing and providing access to the mathematical and/or geometric description of an object is suitable and contemplated for simulations according to the present invention.
  • an offset surface of the representation of independent object 600 is computed. As such, the independent object has a progenitor triangulated surface 600a representation and an offset triangulated surface representation 600b.
  • step 800c bounding volume hierarchies are computed and stored for the objects in the virtual environment.
  • 24-DOPS may be used to provide a progressively refined boundary volume hierarchy of the virtual objects.
  • the root level of each bounding volume hierarchy is a superset of the entire core representation.
  • the bounding element e.g., a 24-DOP
  • bounding volume hierarchies 500b and 600c are generated for the controlled object and independent object, respectively.
  • the bounding volume hierarchy 600c of the independent object is constructed to enclose the progenitor surface 600a and the offset surface 600b.
  • the leaf level bounding elements enclose both a primitive (e.g., a triangular facet) on the progenitor surface and the associated primitive on the offset surface.
  • the bounding elements at the leaf level incorporate depth information about the interior of the independent object.
  • an original Voronoi diagram 600d is calculated for the triangulated surface representation of independent object 600.
  • an original Voronoi diagram of the intersected object is computed.
  • the Voronoi diagram may need to be computed for each virtual object for which an intersection will be tested.
  • Preprocessing stage 80 may include other computational tasks or initialization procedures required to begin a simulation.
  • the simulation may need to check for and validate a correct communication with interface device 200", synchronize the haptic interface device and one or more rendering modes, etc.
  • the preprocessing phase 800 is outside of the force feedback loop.
  • the feedback control loop is activated and the simulation begins.
  • the interaction is divided into two phases; a broad phase 300 and a narrow phase 400.
  • a user may begin participating in the virtual environment by manipulating the interface device 200".
  • Interface device 200" may be any interface device capable of translating a user's translational and rotational objectives into control signals and providing them to the simulation. In the embodiment illustrated in Fig.
  • interface device 200" provides confrol signals 205" including information related to translation affected by the user in three dimensions and rotations affected by the user in three dimensions. For instance, interface device may indicate the change in pose of the interface device enacted by the user since the last update of control signals 205". Alternatively, interface device 200" may provide an absolute pose of interface device 200" at the instant of issue of control signals 205". The control signals are then provided to the simulation such that the simulation can be updated accordingly.
  • the simulation may update the pose of controlled object 500.
  • simulation 100 may effect a transformation on the core representation of controlled object 500 to simulate the movement of the virtual object according to control signals 205 ' ' .
  • controlled object 500 may be complex and consist of a large number of primitives. As such, transforming each of the primitives of the core representation each time control signals are issued from the interface device may compromise real-time constraints.
  • updating the controlled object according to the control signals may be completed recursively and involving the various representations of the controlled object only at a necessary resolution.
  • updating the pose of the controlled object may be part of, and may be terminated or completed, along with the traversal of the bounding volume hierarchies 500b and 600c.
  • the updating of a controlled object may be accomplished with the same efficiency as the intersection of the object by only requiring portions of an object representation to be transformed as needed, for instance, by transforming only child bounding volumes of intersecting bounding volumes and ignoring children of non-intersecting bounding volumes.
  • control signals 205" are obtained from the interface device 200" and the broad phase loop of intersecting the controlled object (i.e., the intersecting object) and the independent object (i.e., the intersected object) begins.
  • the broad phase includes traversing the bounding volume hierarchies of the controlled and independent objects. As noted above, updating the controlled object is accomplished as an iterative process during the traversal of the bounding volume hierarchies.
  • the core level bounding element of hierarchy 500a i.e., the bounding element that encloses the entire representation of the controlled object
  • the core level bounding element of hierarchy 500a is updated according to control signals 205".
  • a transformed root level bounding element encloses the controlled object as if the controlled object had been transformed itself. Such is the case at every level of the bounding volume hierarchy.
  • each bounding element at level is tested for an intersection. As discussed above, if no intersection occurs between the two root level bounding volumes, no intersection has occurred between the core representation of the controlled object and the independent object. Likewise, at any level; of the bounding volume hierarchy, if no bounding volumes intersect, no intersection between the controlled object and the independent object occurs.
  • the bounding volumes at level; of hierarchy 500b having intersecting parent bounding volumes are updated according to control signals 205' and then intersected with the bounding volumes at level; of hierarchy 600c having intersecting parent bounding volumes.
  • the bounding volumes that are determined as intersecting are flagged and only children of these intersecting bounding volumes are tested for intersection during the next iteration at leveli +1 . Non-intersecting bounding volumes and their children are not considered during subsequent iterations either for update or intersection. If no intersection is detected at levelj, the traversal is terminated and the simulation provides this information to the user. For example, in a haptic rendering, the simulation may provide zero level rendering signals 145' to the interface device 200" to indicate that no collision has occurred and no forces should be articulated to the user.
  • step 350 it is determined if the leaf level of the hierarchy has been intersected as shown in step 350. If the leaf level has not been intersected, i is incremented and the procedure is repeated from step 320. If the leaf level of the hierarchy has been intersected and at least one leaf level bounding volume of hierarchies 500b and 600c intersect, the core representations of the controlled object and the independent object may intersect and the simulation enters a narrow phase 400.
  • each leaf level bounding volume of hierarchy 500b that has tested positive for intersection with at least one leaf level bounding volume of hierarchy 600c will enclose a primitive of the core representation of the controlled object that may need to be considered during the narrow phase 400 in order to accurately compute the extent of intersection or penetration of the controlled object into the independent object.
  • Narrow phase 400 includes identifying for each intersecting primitive in the representation of the controlled object the nearest primitive of the representation of the independent object. Once the nearest primitive has been identified, a distance from the intersecting primitive is determined and a penetration vector is computed for that intersection. After a penetration vector has been computed for each intersecting primitive, a synthesized vector may be generated from the penetration vectors. This synthesized vector may be indicative of a force that will reduce or eliminate the intersection between the controlled and the independent object.
  • an intersecting primitive is taken from the set of intersecting primitives that resulted from the broad phase.
  • the nearest primitive of the representation of the independent object is determined by using the proximity information of the original Voronoi diagram computed from the surface of the independent object.
  • the original Voronoi primitives of a triangulated surface in three dimensions are a set of planes.
  • each facet of the triangulated surface representing independent object 600 has an associated Voronoi plane for each of its three edges.
  • the original Voronoi planes associated with a particular facet bound the region of the interior of the independent object containing all points nearer to the particular facet than any other facet in the triangulated surface, that is, they bound the associated Voronoi region.
  • a primitive of the independent object (intersected primitive) is nearest to an intersecting primitive of the controlled object by comparing the intersecting primitive to the Voronoi planes associated with the intersected primitive. If the intersecting primitive is found to lie in the region bounded by the Voronoi planes, it is nearer the associated intersected primitive than any other primitive of the triangulated surface of the independent object.
  • the selected intersecting primitive is first transformed according to control signals 205". It should be appreciated that only the intersecting primitives need to be updated and not the entire core representation.
  • the intersecting primitive is then compared to the Voronoi planes of the intersected primitive in step 410. In step 420, it is determined if the intersecting primitive is within the Voronoi region of the intersected facet. If the primitive is determined to be in the Voronoi region bounded by the Voronoi planes, the penetration of the primitive is then calculated and a penetration vector associated with that intersection is computed in step 430.
  • the next intersecting primitive is selected from the set of intersecting primitives that resulted from the broad phase and the procedure is repeated at step 410 until all intersecting primitives have been processed.
  • the penetration vector of an intersection may be computed in a number of ways.
  • the core representation of the controlled object includes a point cloud.
  • the intersecting primitives are in fact points.
  • the perpendicular distance between the intersecting point and the nearest facet may be calculated in order to generate a penetration vector indicating the magnitude of the penetration, and a direction of least distance to the surface of the intersected object.
  • a synthesized vector may be computed from the penetration vectors.
  • the synthesized vector may be generated in a number of ways. For example, a weighted sum of the penetration vectors may be computed in order to generate the synthesized vector. Additionally, the synthesized vector may be computed based on the penetration vector having the greatest magnitude.
  • Various alternative methods for synthesizing a vector from the penetration vectors will occur to those skilled in the art and are within the scope of the present invention.
  • the synthesized vector and a location of application are then is then incorporated into rendering signals 145" as a force feedback applied to interface device 200" to indicate the controlled object is in contact with the independent object and to articulate to the user the conditions of the contact and the forces arising from such a spatial relationship.
  • visual rendering signals may also be generated from the force-feedback vector.
  • the representations of the controlled object and the independent object may be transformed to simulate the effects that such a force would have on- the virtual objects.
  • the representation of the independent object may be updated to indicate the material from the independent object removed by the contact of the milling tool (i.e., the controlled object).
  • a user will respond to the feed-back received via the rendering signals by further manipulation of the controlled object.
  • the feedback control loop continues as described until the simulation ends, the user ceases to respond or otherwise indicates that the simulation has been terminated.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • Human Computer Interaction (AREA)
  • Geometry (AREA)
  • Manipulator (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

La présente invention concerne un procédé de détection de collision qui permet une simulation à six degrés de liberté d'objets virtuels de formes arbitraires. Ce procédé de détection de collision est capable de donner un contact intégral des corps par détection des différentes intersections entre la représentation d'un premier objet virtuel et celle d'un second objet virtuel. Le contact intégral des corps peut être fourni sous forme de signaux de rendu haptiques indiquant un contact intégral des corps entre les deux objets virtuels.
PCT/US2002/032235 2001-10-09 2002-10-09 Procede et dispositif de detection et de correction d'une penetration entre des objets WO2003031005A2 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US32794001P 2001-10-09 2001-10-09
US60/327,940 2001-10-09

Publications (2)

Publication Number Publication Date
WO2003031005A2 true WO2003031005A2 (fr) 2003-04-17
WO2003031005A3 WO2003031005A3 (fr) 2003-08-21

Family

ID=23278750

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2002/032235 WO2003031005A2 (fr) 2001-10-09 2002-10-09 Procede et dispositif de detection et de correction d'une penetration entre des objets

Country Status (1)

Country Link
WO (1) WO2003031005A2 (fr)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN100337250C (zh) * 2004-10-26 2007-09-12 中国科学院计算技术研究所 一种人体模型与柔性物体间的冲突检测方法
US20130241957A1 (en) * 2011-08-31 2013-09-19 Rakuten, Inc. Information processing device, control method for information processing device, program, and information storage medium
US9423948B2 (en) 2011-08-31 2016-08-23 Rakuten, Inc. Information processing device, control method for information processing device, program, and information storage medium for determining collision between objects on a display screen
CN111179405A (zh) * 2018-11-09 2020-05-19 中移(杭州)信息技术有限公司 3d场景漫游碰撞的检测方法及装置
CN111540053A (zh) * 2020-03-12 2020-08-14 浙江工业大学 一种针对种植牙修复体设计的嵌入回缩方法
CN112771582A (zh) * 2018-10-02 2021-05-07 索尼公司 图像处理设备和方法

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
A. GREGORY ET AL: "Six degree-of-freedom haptic display of polygonal models" PROCEEDINGS OF THE CONFERENCE ON VISUALIZATION '00 , 2000, pages 139-146, XP002238309 Salt Lake City, Utah, *
KLOSOWSKI J T ; HELD M ; MITCHELL J S B ; SOWIZRAL H ; ZIKAN K : " Efficient collision detection using bounding volume hierarchies of k-DOPs " IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, vol. 4, no. 1, January 1998 (1998-01) - March 1998 (1998-03), pages 21-36, XP002238310 *
PONAMGI M K ET AL: "INCREMENTAL ALGORITHMS FOR COLLISION DETECTION BETWEEN POLYGONAL MODELS" IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS, IEEE SERVICE CENTER, PISCATAWAY, NJ, US, vol. 3, no. 1, 1997, pages 51-64, XP000685845 ISSN: 1077-2626 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN100337250C (zh) * 2004-10-26 2007-09-12 中国科学院计算技术研究所 一种人体模型与柔性物体间的冲突检测方法
US20130241957A1 (en) * 2011-08-31 2013-09-19 Rakuten, Inc. Information processing device, control method for information processing device, program, and information storage medium
US8952994B2 (en) * 2011-08-31 2015-02-10 Rakuten, Inc. Information processing device, control method for information processing device, program, and information storage medium
US9423948B2 (en) 2011-08-31 2016-08-23 Rakuten, Inc. Information processing device, control method for information processing device, program, and information storage medium for determining collision between objects on a display screen
US9619134B2 (en) 2011-08-31 2017-04-11 Rakuten, Inc. Information processing device, control method for information processing device, program, and information storage medium
US11568602B2 (en) 2018-10-02 2023-01-31 Sony Corporation Image processing apparatus and method using point cloud generation and a surface of a mesh
CN112771582A (zh) * 2018-10-02 2021-05-07 索尼公司 图像处理设备和方法
EP3843046A4 (fr) * 2018-10-02 2021-10-27 Sony Group Corporation Dispositif et procédé de traitement d'image
CN112771582B (zh) * 2018-10-02 2024-01-12 索尼公司 图像处理设备和方法
US11922579B2 (en) 2018-10-02 2024-03-05 Sony Corporation Image processing apparatus and method for image processing by deriving voxel and mesh data to generate point cloud data
CN111179405A (zh) * 2018-11-09 2020-05-19 中移(杭州)信息技术有限公司 3d场景漫游碰撞的检测方法及装置
CN111179405B (zh) * 2018-11-09 2023-07-04 中移(杭州)信息技术有限公司 3d场景漫游碰撞的检测方法及装置
CN111540053A (zh) * 2020-03-12 2020-08-14 浙江工业大学 一种针对种植牙修复体设计的嵌入回缩方法
CN111540053B (zh) * 2020-03-12 2023-05-23 浙江工业大学 一种针对种植牙修复体设计的嵌入回缩方法

Also Published As

Publication number Publication date
WO2003031005A3 (fr) 2003-08-21

Similar Documents

Publication Publication Date Title
US7084869B2 (en) Methods and apparatus for detecting and correcting penetration between objects
US6704694B1 (en) Ray based interaction system
Gregory et al. Six degree-of-freedom haptic display of polygonal models
US6084587A (en) Method and apparatus for generating and interfacing with a haptic virtual reality environment
US6191796B1 (en) Method and apparatus for generating and interfacing with rigid and deformable surfaces in a haptic virtual reality environment
JP2726811B2 (ja) ボクセルに基づくオブジェクト表現のための衝突回避システム
US20040010346A1 (en) Method of real-time collision detection between solid geometric models
Johnson et al. Six degree-of-freedom haptic rendering of complex polygonal models
Ho et al. Ray-based haptic rendering: Force and torque interactions between a line probe and 3d objects in virtual environments
US7403880B2 (en) Moving a virtual articulated object in a virtual environment while avoiding internal collisions between the articulated elements of the articulated object
RU2308764C2 (ru) Перемещение виртуального сочлененного объекта в виртуальном пространстве с предотвращением столкновений сочлененного объекта с элементами окружающего пространства
Ruspini et al. Haptic display for human interaction with virtual dynamic environments
Fares et al. Collision detection for rigid bodies: A state of the art review
Hadap et al. Collision detection and proximity queries
WO2003031005A2 (fr) Procede et dispositif de detection et de correction d'une penetration entre des objets
Vlasov et al. Haptic rendering of volume data with collision detection guarantee using path finding
EP0545684A2 (fr) Système à faible coût pour modeler la position et le mouvement d'un joueur dans un système de réalité virtuelle
Peng et al. Interactive Solid Modelling in a Virtual Environment with a Haptic Interface
US6873944B1 (en) Method of real time collision detection between geometric models
CN112486319A (zh) 基于触觉渲染设备的vr交互方法、装置、设备及介质
O'Sullivan et al. Collision Handling for Virtual Environments.
Basdogan et al. 3-DOF haptic rendering
Peng et al. A virtual sculpting system based on triple Dexel models with haptics
Yu et al. Six degree-of-freedom haptic simulation of sharp geometric features using a hybrid sphere-tree model
Laycock et al. The haptic rendering of polygonal models involving deformable tools

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): CA JP

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): AT BE BG CH CY CZ DE DK EE ES FR GB GR IE IT LU MC NL PT SE SK TR

121 Ep: the epo has been informed by wipo that ep was designated in this application
122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: JP

WWW Wipo information: withdrawn in national office

Country of ref document: JP