WO2005069755A2 - Computing environment that produces realistic motions for an animatronic figure - Google Patents

Computing environment that produces realistic motions for an animatronic figure Download PDF

Info

Publication number
WO2005069755A2
WO2005069755A2 PCT/US2004/026115 US2004026115W WO2005069755A2 WO 2005069755 A2 WO2005069755 A2 WO 2005069755A2 US 2004026115 W US2004026115 W US 2004026115W WO 2005069755 A2 WO2005069755 A2 WO 2005069755A2
Authority
WO
WIPO (PCT)
Prior art keywords
animatronic
movement
instruction
fixed show
motion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/US2004/026115
Other languages
English (en)
French (fr)
Other versions
WO2005069755A3 (en
Inventor
Alexis P. Wieland
Holger Irmler
Akhil Jiten Madhani
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Disney Enterprises Inc
Original Assignee
Disney Enterprises Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Disney Enterprises Inc filed Critical Disney Enterprises Inc
Priority to EP04780885.2A priority Critical patent/EP1711241A4/en
Priority to JP2006549231A priority patent/JP2007517585A/ja
Publication of WO2005069755A2 publication Critical patent/WO2005069755A2/en
Publication of WO2005069755A3 publication Critical patent/WO2005069755A3/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63HTOYS, e.g. TOPS, DOLLS, HOOPS OR BUILDING BLOCKS
    • A63H11/00Self-movable toy figures
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B62LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
    • B62DMOTOR VEHICLES; TRAILERS
    • B62D57/00Vehicles characterised by having other propulsion or other ground- engaging means than wheels or endless track, alone or in addition to wheels or endless track
    • B62D57/02Vehicles characterised by having other propulsion or other ground- engaging means than wheels or endless track, alone or in addition to wheels or endless track with ground-engaging propulsion means, e.g. walking members
    • B62D57/028Vehicles characterised by having other propulsion or other ground- engaging means than wheels or endless track, alone or in addition to wheels or endless track with ground-engaging propulsion means, e.g. walking members having wheels and mechanical legs
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B62LAND VEHICLES FOR TRAVELLING OTHERWISE THAN ON RAILS
    • B62DMOTOR VEHICLES; TRAILERS
    • B62D57/00Vehicles characterised by having other propulsion or other ground- engaging means than wheels or endless track, alone or in addition to wheels or endless track
    • B62D57/02Vehicles characterised by having other propulsion or other ground- engaging means than wheels or endless track, alone or in addition to wheels or endless track with ground-engaging propulsion means, e.g. walking members
    • B62D57/032Vehicles characterised by having other propulsion or other ground- engaging means than wheels or endless track, alone or in addition to wheels or endless track with ground-engaging propulsion means, e.g. walking members with alternately or sequentially lifted supporting base and legs; with alternately or sequentially lifted feet or skid

Definitions

  • a system and method are disclosed, which generally relate to robotic figures, and more specifically to animatronic figures.
  • An animatronic figure is a robotic figure, puppet, or other movable object that is animated via one or more electromechanical devices.
  • the term "animated” is meant to be interpreted as to move to action.
  • the electromechanical devices include electronics, mechanical, hydraulic, and/or pneumatic parts.
  • Animatronic figures are popular in entertainment venues such as theme parks. For example, animatronic characters can be seen in shows, rides, and/or other events in a theme park.
  • the animatronic character's body parts, such as the head and the arms, may generally move freely. However, the animatronic character is usually incapable of freely roaming or walking from one place to another.
  • a reception module receives a fixed show selection input and a puppetted input from an operator.
  • the fixed show selection input is associated with at least one fixed show instruction.
  • the puppetted input provides a puppetted instruction.
  • a translation software module translates the at least one fixed show instruction associated with the fixed show selection input into at least one fixed show physical movement instruction and translates the received puppetted instruction into at least one puppetted physical movement instruction.
  • a motion software module receives the at least one fixed show physical movement instruction, receives the at least one puppetted physical movement instruction, and calculates a composite animated instruction from the at least one fixed show physical movement instruction and the at least one puppetted physical movement instruction so that at least one actuator can effectuate at least one component of the animatronic figure in a life-like manner.
  • An audio software module provides an instruction to an audio device to output an audio signal when the motion of the at least one component of the animatronic figure is effectuated.
  • An automatic fixed show software module automatically provides the fixed show selection input to the reception module at a predetermined time.
  • a predetermined time coincides with a time that the operator provides the puppetted input.
  • an animatronic system there is an animatronic system.
  • the operator presses a button to provide the fixed show selection input.
  • a sensor is operably connected to the animatronic figure that determines the occurrence of an event.
  • an animatronic system there is an animatronic system.
  • the occurrence of an event triggers the fixed show selection input to be inputted to the reception module.
  • an animatronic system there is an animatronic system.
  • the operator turns a dial to provide the fixed show selection input to the reception module.
  • a method that produces motion of an animatronic figure.
  • a puppetting instruction is provided to the animatronic figure to perform a puppetting movement.
  • a fixed show selection is associated with at least one fixed show instruction to the animatronic figure to perform a fixed show movement.
  • the puppetting instruction is combined with the at least one fixed show instruction to form a combined instruction.
  • the animatronic figure is instructed to perform the combined instruction.
  • the puppetted movement is to be performed by a different component of the animatronic figure than the fixed show movement.
  • an animatronic system there is an animatronic figure.
  • a reception module receives a first command and a second command from an operator.
  • the reception module is operably connected to the animatronic figure.
  • the reception module communicates with the animatronic figure.
  • the animatronic figure makes a first movement according to the first command.
  • the animatronic figure makes a second movement according to the second command.
  • a filter module filters the first command through a first filter and filters the second command through a second filter to coordinate a first movement of the animatronic figure resulting from the first command with a second movement of the animatronic figure resulting from the second command so that the motion of the animatronic figure provides a life-like appearance.
  • the first command is a puppetted instruction.
  • the second command is a fixed show selection that is associated with at least one fixed show instruction.
  • the first filter is a low pass filter.
  • the second filter is a high pass filter.
  • the second filter is a band pass filter.
  • an animatronic system there is an animatronic system.
  • the first filter and the second filter are used on different components of the same body part of the animatronic figure.
  • an animatronic system there is an animatronic system.
  • the first filter and the second filter are used on different body parts of the animatronic figure.
  • an animatronic figure there is a leg.
  • a leg actuator is operably connected to the leg, wherein the leg actuator effectuates movement of the leg.
  • a wheel actuator is operably connected to the wheel, wherein the wheel actuator effectuates movement of the wheel.
  • a processor determines a leg motion and a wheel motion to effectuate movement of the animatronic figure. The processor sends the leg motion to the leg actuator. The processor sends the wheel motion to the wheel actuator.
  • a position sensor determines a first current position of the animatronic figure.
  • an animatronic figure there is an animatronic figure.
  • An incremental sensor measures a second current relative position of the animatronic figure by incrementing the distance traveled from an initial position of the animatronic figure.
  • a clipping module determines if the difference between the second current position of the animatronic figure and the first current position of the animatronic figure has reached a positional clipping limit.
  • a clipping module determines if the velocity of the animatronic figure in moving from the first current position to the second current position has reached a velocity-clipping limit in addition to reaching the positional clipping limit.
  • a clipping module determines if the velocity of the animatronic figure in moving from the first current position to the second current position has reached a clipping limit.
  • a clipping module determines if the acceleration of the animatronic figure in moving from the first current position to the second current position has reached a clipping limit, separately or in conjunction with a position-clipping limit and/or with a velocity- clipping limit.
  • the clipping module shortens a trajectory of the animatronic figure if the clipping limit has been reached.
  • the clipping module reduces a velocity of the animatronic figure if the clipping limit has been reached.
  • the clipping module reduces an acceleration of the animatronic figure if the clipping limit has been reached.
  • a puppetted instruction is provided to the animatronic figure to perform a puppetted movement.
  • a fixed show selection command is provided to the animatronic figure to perform at least one fixed show movement associated with the fixed show selection.
  • the puppetted movement and the fixed show movement are combined into a composite movement.
  • the animatronic figure is instructed to perform the composite movement.
  • the fixed show selection is provided by the user.
  • the puppetted movement and the at least one fixed show movement is a composite of the puppetted movement and the fixed show movement.
  • a pre-determined limit reduces the composite of the puppetted movement and the fixed show movement.
  • an animatronic system there is an animatronic figure.
  • a reception module receives a first command from an operator.
  • the reception module is operably connected to the animatronic figure.
  • the reception module communicates with the animatronic figure.
  • the first command Is used by the animatronic system as part of a calculation to determine a first movement.
  • An algorithmic response module provides a second command to the animatronic figure based on the occurrence of the first command being provided to the animatronic figure.
  • the algorithmic response module receives the first command from the reception module.
  • the animatronic figure makes a second movement according to the second command.
  • the first command is a puppetted instruction.
  • the second command is a fixed show selection input associated with at least one fixed show instruction.
  • an animatronic system there is an animatronic system.
  • the second command is an animatronic fixed show instruction that is based on the occurrence of a sensor operably connected to the animatronic figure detecting a stimulus.
  • a filter module filters the first movement through a first filter and filters the second movement through a second filter to coordinate the first movement with the second movement so that the motion of the animatronic figure provides a life-like appearance.
  • Figure 1 illustrates an animatronic system, which is configured to produce different types of shows in real-time and in a life-like manner.
  • Figure 2 illustrates a process that produces motion of the animatronic figure.
  • Figure 3A illustrates an algorithmic configuration, which determines movements of the animatronic figure according to an algorithm.
  • Figure 3B illustrates one embodiment of the algorithmic configuration 300 in which the sensor provides a sensed condition to the algorithmic response module.
  • Figure 3C illustrates another embodiment of the algorithmic configuration in which reception module receives the sensed condition from the sensor and receives the user-inputted motion from the operator.
  • Figure 4 illustrates a process that combines a user-inputted instruction with an algorithmically determined instruction.
  • Figure 5 illustrates a filtering animatronic configuration in which movements produced by the animatronic figure are filtered.
  • Figure 6A illustrates a graph for unfiltered motion of the head of the animatronic figure.
  • Figure 6B illustrates a graph for filtered motion of the head of the animatronic figure.
  • Figure 7 illustrates an animatronic filter configuration, which filters user-inputted motions and algorithmically determined motions in a coordinated manner to produce life-like motions for the animatronic figure.
  • Figure 8A illustrates a graph depicting filters for an animatronic figure.
  • Figure 8B illustrates a graph that represents the rotation of body parts of the animatronic figure that were filtered by the filter module.
  • Figure 9A illustrates a process for filtering motions of the animatronic figure based on a user-inputted command and a command determined by the occurrence of the user-inputted command.
  • Figure 9B illustrates a process for filtering motions of the animatronic figure based on a user-inputted command a command that is determined from a stimulus.
  • Figure 10 illustrates a layered system of software modules.
  • Figure 11 illustrates a detailed view of the trajectory planner.
  • Figure 12 shows an example of a simple composite trajectory.
  • Figure 13 illustrates a curve described by a plurality of keyframes.
  • Figure 14 illustrates another embodiment in which a curve described by a plurality of keyframes.
  • Figure 15 illustrates a graph with a sequence of points.
  • Figure 16A illustrates another graph with an interpolation of data.
  • Figure 16B shows the corresponding velocities of the graph depicted in Figure 16A.
  • Figure 16C shows the corresponding accelerations of the graph depicted in Figure 16A.
  • Figure 17 is an example script for the simple four point keyframe trajectory shown in Figure 15 used to describe smooth keyframes.
  • Figure 18 illustrates an example script for the composite trajectory shown in Figure 12.
  • a computing environment is disclosed, which provides greater flexibility for controlling animatronic systems than previously seen and provides the ability to produce real-time composite motions.
  • This computing environment can be applied to robotic systems generally and is not limited to only animatronic systems. Further, this computing environment is not limited to any particular programming language. In one embodiment, a scripting language is used to provide a programmer with a high level tool that instructs the animatronic figure to create and produce real, life-like motions.
  • the computing environment provides resources for different types of shows that the animatronic figure can perform, the combination and sequencing of the movements in these shows to produce life-like movements in real-time, calculations of trajectories for real-time movements, and/or filtering of the animatronic figure's movements.
  • Figure 1 illustrates an animatronic system 100, which is configured to produce different types of shows in real-time and in a life-like manner.
  • a show is a sequence of movements of an animatronic figure 108 animated by the animatronic system 100.
  • One type of show is a puppetted show.
  • the puppetted show is a sequence of movements that are operator-controlled.
  • an operator 101 manually inputs the desired movements of the animatronic figure 108 into the animatronic system 100
  • the animatronic figure 108 produces a corresponding motion in a real-time fashion.
  • the animatronic system 100 instructs the animatronic figure 108 to produce the desired motion instantly or in a short time period after the user manually inputs the desired motion command. Therefore, the animatronic figure 108 appears to be moving according to the desired motion as the user is entering the desired motion into the animatronic system 100.
  • puppetted shows such as the MUPPETS or the puppets in the live stage adaptation of Disney's "The Lion King” which originally played in the New Amsterdam Theater in New York, NY in 1997.
  • the user inputs the puppetted motions through a joystick or other analog input device.
  • the user inputs the puppetted motions through a keyboard.
  • the user inputs the puppetted motions through sensors attached to the user's body.
  • face tracking is used to input the puppetted motions. Face tracking detects movements of the face and/or head. A camera can be focused on the face and/or head to receive the puppetted motion. Sensors can also be placed on the head to receive the puppetted motion. For instance, the operator 101 can wear a headband that has sensors on it for detecting movement of the head.
  • the user inputs the puppetted motions through voice commands into a microphone that operates in conjunction with voice recognition software.
  • Another type of show is a fixed show.
  • the fixed show is a recordation of pre-animated sequences that can be played back either once or continuously in a loop.
  • the operator 101 can simply select a fixed show and animate the animatronic figure 108 without having to input each movement while the show is playing, as the operator 101 would do if providing a puppetted instruction.
  • One of ordinary skill in the art will recognize exampies of fixed shows such as the "Enchanted Tiki Room” shown at Disneyland in 1963 and "Great Moments with Mr. Lincoln” originally shown at the New York world's fair in 1964 and moved to Disneyland in 1965.
  • the fixed show is a recordation of the user puppetting a sequence of movements.
  • the fixed show is a recordation of movements that a user enters through a graphical user interface ("GUI").
  • GUI graphical user interface
  • motions are derived from other sources, such as deriving mouth positions from analyzing recorded speech.
  • a combination of these approaches is used.
  • the instructions for the fixed show are stored on a computer-readable medium.
  • the user inputs the selection of the fixed show through a button (not shown) that instructs the animatronic system 100 to animate the animatronic figure 108 according to the pre-recorded motions of the fixed show that is selected.
  • a plurality of buttons can be provided with each button representing a different fixed show selection.
  • the operator 101 can have a variety of fixed show selections from which to choose.
  • Each of the fixed show selections is associated with a set of pre-recorded movements.
  • the animatronic system 100 performs the set of pre-recorded movements associated with the fixed show selection.
  • the operator 101 only has to provide one command to the animatronic system 100 to choose a fixed show selection.
  • the animatronic system instructs the animatronic figure 108 to perform the instructions associated with the fixed show selection made by the operator 101.
  • the user inputs the selection of the fixed show with a touch screen display.
  • the user inputs the selection of the fixed show through a dial.
  • the user inputs the selection of the fixed show through voice commands into a microphone that operates in conjunction with voice recognition software.
  • the animatronic system 100 provides the user with the ability to animate the animatronic figure 108 according to a fixed show and a puppetted show simultaneously. If a fixed show provides an instruction to one actuator while the puppetted sequence provides an instruction to a different actuator, both instructions are performed simultaneously to give the appearance that two different body parts are moving simultaneously.
  • a composite motion is calculated.
  • the composite motion is calculated so that the composite motion appears life-like. For example, if a fixed show is a sneezing motion and a puppetted motion is moving the head forward, a simple superimposition of the fixed show instruction and the puppetted motion would lean the head forward more so than if only the puppetted instruction was given.
  • the composite motion is formed by multiplying motions to increase or reduce a component motion. For example, the amplitude that the tail wags can be increased when a laugh is played.
  • the composite motion is formed by modulating component motions. For example, the frequency of the breathing motion can increase when a strenuous motion is played. This combination parallels the motion of a living creature that would have, in the sneezing example, an exaggerated movement of the head forward if a sneeze occurred during the forward motion of the head.
  • the composite motion may appear to be too exaggerated of a motion to appear life-like.
  • the head may move too far forward during the combined forward motion and the sneeze to appear life-like.
  • the composite motion can be clipped.
  • the term clipping refers to the reduction of a value to fall within a predetermined limit.
  • the trajectory of the composite motion of the animatronic figure 108 can be clipped.
  • the forward trajectory of the composite motion of the head in the sneezing example can be shortened to meet the predetermined limit of how far the head can move forward.
  • An example of the predetermined limit would be stopping the head of the animatronic figure 108 from touching the ground.
  • clipping ensures that the range limits for the actuators of the animatronic figure 108 that perform the movement are complied with. For instance, if the operator 101 provides a puppetted instruction to move the head of the animatronic figure 108 too far backwards, a hard stop will be encountered. By the term hard stop, one of ordinary skill in the art will recognize that the actuators will hit a physical end stop or barrier restricting the motion of the robot. A wide variety of problems may result from hitting a hard stop. For instance, it may be possible to hit the hard stop with enough force to physically damage the robotic figure. Furthermore, other actuators that depend on the movements of the affected actuators may then also malfunction. In essence, hitting a hard stop could potentially lead to a complete malfunction of the animatronic figure 108.
  • a clipping module determines if the difference between a second current position of the animatronic figure 108 and a first current position of the animatronic figure 108 has reached a positional clipping limit.
  • the velocity of the composite motion of the animatronic figure 108 can be clipped.
  • the speed at which a body part of the animatronic figure 108 is moving can be reduced to fall within a predetermined limit.
  • the predetermined limit can be set to one hundred degrees per second to assure that the sideways movement of the head does not exceed a potentially dangerous speed.
  • clipping prevents structural constraints from being reached. For instance, a motor has a limit on the velocity the motor can produce. If the animatronic figure 108 attempts to reach or surpass this velocity limit, the motor may simply malfunction.
  • the clipping module determines if a velocity clipping limit has been reached. In yet another embodiment, the clipping module determines if the positional clipping limit and the velocity clipping limit has been reached.
  • the acceleration of the animatronic figure 108 can be clipped.
  • the change in the speed at which a body part is moving can be reduced to fall within a predetermined limit. Since the required motor torque is proportional to acceleration, this can be used to effectively limit the torque required by the actuators as well as make motions that are smooth and life-like.
  • Clipping range of a joint's motion by simply limiting the allowed range will cause the joint to suddenly stop when it reaches the end of its range. This sudden stop may cause an acceleration that exceeds the predetermined acceleration limit.
  • a joint has clipping simultaneously for both position and acceleration it is necessary to reduce the velocity in the direction of the position-clipping limit as the joint reaches that position-clipping limit.
  • the clipping module determines if an acceleration clipping limit has been reached.
  • the clipping module can also determine any combination of position, velocity, and/or acceleration being clipped.
  • a reception module 102 receives operator input commands for controlling the animatronic figure 108.
  • the reception module 102 is a software program that receives user inputs from a hardware device such as a joystick.
  • the joystick sends control inputs for adjusting the character's movements.
  • the control inputs are, for example, analog and/or digital signals for adjusting the character's puppeting motions.
  • the reception module 102 is a software program that receives user inputs from a hardware device such as a keyboard that sends control inputs, such as analog and/or digital signals, to adjust the character's movements.
  • an operator presses a button to provide control inputs.
  • the operator turns a dial to provide control signals to the animatronic figure.
  • the control signals are input into the translation software module 104.
  • an operator of an amusement park ride rotates a joystick to control the head and the neck of the animatronic figure 108 for the viewing enjoyment of park patrons.
  • an operator of the animatronic figure 108 enters a keyboard command to move the legs of the animatronic figure 108, altering the direction of travel of the animatronic figure 108 from a forward direction to a right turn to avoid a collision course with a park patron.
  • the use of the keyboard may be particularly helpful in a debugging environment for debugging the animatronic figure 108.
  • the user can debug the animatronic figure 108 by entering commands to the keyboard one command at a time to analyze individual motions. Similarly, the user can also provide a plurality of commands through the keyboard to analyze the individual motions.
  • the translation software module 104 converts the fixed show instruction associated with the fixed show selection into at least one physical movement instruction. Further, the translation software module can also convert the puppetted instruction into at least one physical movement instruction. First, the translation software module 104 evaluates a fixed show instruction associated with the fixed show selection. In one embodiment, a computer program is loaded from a memory device.
  • the reception module 102, the translation software module 104, and the motion software module are all stored on different computers.
  • the reception module 102 does not need to be stored on a computer. Rather, the reception module 102 can be a simple input device.
  • the reception module 102 and the translation software module 104 are stored on the same computer but a different computer than the computer on which the motion software module 106 is stored.
  • the reception module 102 and the motion software module 106 are stored on the same computer but on a different computer than the computer on which the translation software module 104 is stored.
  • Figure 2 illustrates a process 200 that produces motion of the animatronic figure 108.
  • the user provides a puppetted instruction to the animatronic system 100 to perform a puppetted movement.
  • the user provides a fixed show selection associated with at least one fixed show instruction to the animatronic system 100 to perform a fixed show movement.
  • the fixed show instruction includes a control signal for making the animatronic figure growl.
  • a combined instruction results from the combination of the puppetted instruction and the fixed show instruction.
  • a combined instruction results from the superimposition of the puppetted instruction and the fixed show instruction.
  • the animatronic system 100 instructs the animatronic figure 108 to move according to the combined instruction.
  • a combined instruction results from multiplying the puppetted instruction and the fixed show instruction.
  • a combined instruction results from the modulation of the puppetted instruction by the fixed show instruction.
  • an operator provides, through keyboard input, a puppetted instruction to raise the animatronic character's leg.
  • the puppetted input is stored in a memory device within a computer.
  • the operator presses a button operably connected to the animatronic character to request a fixed show selection that is associated with at least one fixed show instruction for the character to growl.
  • a processor within the computer superimposes the instruction to raise the animatronic character's 108 leg with the instruction to growl.
  • the animatronic system 100 then causes the animatronic figure 108 to raise its leg and to growl.
  • Figure 3A illustrates an algorithmic configuration 300, which determines movements of the animatronic figure 108 according to an algorithm.
  • the animatronic system 100 analyzes a variety of conditions to determine if the algorithm should be run and/or what motion should be produced.
  • the condition can be an environmental occurrence, a motion resulting from a puppetted instruction, or a motion resulting from a fixed show selection. If one of the conditions is met, the algorithm performs a real time determination or modification of the motion that should be produced by the animatronic figure 108.
  • an algorithmic response module 308 determines whether a condition has been met so that an algorithm can be run.
  • the algorithmic response module accesses a database that stores a condition and an algorithm to be run if the condition is met.
  • the database can store a condition and an instruction to be performed if the condition is met.
  • the algorithmic response module 308 determines the direction of the resulting motion.
  • the algorithmic response module 308 determines the magnitude of the resulting motion.
  • the algorithmic response module 308 determines the duration of the resulting motion.
  • the algorithmic response module 308 determines the sequence of motions to follow or how many times an individual motion repeats.
  • a clock 302 provides the time to the algorithmic response module 308.
  • One of the conditions can be that the animatronic figure 108 wags its tail at a particular time or at a particular frequency. For instance, an algorithm can be run so that the animatronic figure 108 wags its tail at one o'clock p.m. every day.
  • the algorithmic response module 308 can receive the time from the clock 302 and determine that the time condition for wagging the tail of the animatronic figure 108 has been satisfied at one o'clock p.m.
  • the tail can be set to wag at one Hertz.
  • the algorithmic response module 308 can receive the time from the clock and use that to determine the current position based on the one Hertz wagging motion.
  • the algorithmic response module then provides an instruction and associated data to the motion software module 106 for the animatronic figure 108 to wag its tail.
  • the algorithmic response module 308 performs the calculations for position, velocity, and acceleration and provides the data to the motion software module 106.
  • the motion software module 106 performs the calculations for position, velocity, and acceleration.
  • the algorithmic response module 308 is the same module as the motion software module 106.
  • an algorithm is provided that calculates a breathing motion for the animatronic figure 108. In other words, a breathing noise can be produced and the chest of the animatronic figure 108 can heave in and out. The breathing motion can be based on the time input of the clock 302 to the animatronic response module 308.
  • a random number generator 304 provides an input to the algorithmic response module 308.
  • An algorithm can be set to receive random numbers from the random number generator 304 to instruct the animatronic figure 108 to perform a motion at random times. For instance, the animatronic figure 108 can be instructed to blink at random times.
  • the user of the random number generator 304 provides a more life-like appearance to the movement of the animatronic figure 108 as blinking is naturally a random movement.
  • the random number generator 304 may provide a random amplitude for a ifiotion.
  • the random number generator 304 may provide a random direction for a motion.
  • the random number generator 304 may provide a random component of a motion. An example of a random component would be a breathing motion that occurs at a randomly varying frequency, amplitude, and offset.
  • a sensor 306 provides data to the algorithmic response module 308.
  • the data is environmental data.
  • the environmental data can include a detection of a body part of the animatronic figure 108 moving too far in one direction towards an end stop. If the animatronic figure 108 is about to sneeze, while the head and neck are turned and therefore near the sideways end stop, the sneezing motion may cause the neck to turn further and clipping of the motion occur. It is preferable in that case for the sneeze to turn the head away from the end stop, which is toward the center of the neck's range of motion, so that clipping does not occur.
  • the direction and the magnitude of the head turn during the sneeze would be the algorithmic response to the neck position sensor.
  • the sensed condition can be the occurrence of the motion of any body part regardless of whether the motion results from a puppetted instruction or from a fixed show selection.
  • a condition is established such that the motion of a first body part invokes an algorithm to produce a motion for a second body part in the same vicinity.
  • the algorithm calculates the second motion such that the first motion and the second motion appear to provide a coordinated life-like movement.
  • the condition can be a downward motion of the head of the animatronic figure 108.
  • the algorithmic response module 308 can then calculate a motion, which causes the animatronic figure 108 to lean forward and crouch slightly in a coordinated manner with the motion of the head of the animatronic figure 108.
  • Figure 3B illustrates one embodiment of the algorithmic configuration 300 in which the sensor 306 provides a sensed condition to the algorithmic response module 308.
  • the sensor 306 receives the condition from the animatronic figure 108.
  • the sensor 306 provides the sensed condition to the algorithmic response module 308, which then calculates a motion according to an algorithm.
  • the motion is then provided to the translation software module 104.
  • the reception module 102 receives an input from a user 101.
  • the translation software module 104 translates the user-inputted motion, combines this with the algorithmically computed response, and produces at least one physical movement instruction.
  • the motion software module then instructs the animatronic figure 108 to move according to the physical movement instruction.
  • the motion software module 106 also performs the functions of the translation software module 104.
  • the translation software module 104 forms a composite motion by combining the user-inputted motion with the algorithmically calculated motion.
  • the animatronic figure 108 performs the composite motion in a life-like manner that coordinates both the user-inputted motion and the algorithmically calculated motion.
  • FIG. 3C illustrates another embodiment of the algorithmic configuration 300 in which reception module 102 receives the sensed condition from the sensor 306 and receives the user-inputted motion from the operator 101.
  • the reception module 102 determines the type of instruction that is receive and distributes the instruction to an appropriate module. For instance, when the reception module 102 receives a user-inputted instruction, the reception module 102 distributes the user- inputted instruction to the translation software module 104. If the reception module 102 receives a sensed condition, the reception module 102 provides the sensed condition to the algorithmic response module 308.
  • the algorithmic response module computes at least one motion that is sent to the translation software module 104. In one embodiment, the motion software module 106 also performs the functions of the translation software module 104 and/or the algorithmic response module 308. Therefore, the reception module 102 can send all the instructions to one module.
  • Figure 4 illustrates a process 400 that combines a user-inputted instruction with an algorithmically determined instruction.
  • a user-inputted instruction is received.
  • the user-inputted instruction is received form the operator 101 at the reception module 102.
  • a stimulus is received at the sensor 306 at a process block 404.
  • the stimulus can be any one of the environmental conditions.
  • the stimulus is provided as an input to an algorithm.
  • the stimulus is provided to the 11 algorithmic response module 308.
  • the algorithm is performed to determine algorithmic movement of the animatronic figure 108.
  • the algorithmic response module 308 runs the algorithm to determine the algorithmic movement.
  • the user-inputted motion is combined with the algorithmically determined motion.
  • the motion software module 106 performs this combination.
  • the composite motion resulting from the combination provides the animatronic figure 108 with a coordinated life-like motion.
  • Figure 5 illustrates a filtering animatronic configuration 500 in which movements produced by the animatronic figure 108 are filtered.
  • the user 101 provides a user-inputted motion to the reception module 102.
  • the reception module 102 provides the user-inputted motion to a filter module 502.
  • the filter module 502 can receive motions or motion instructions that are not necessarily user-inputted.
  • the filter module 502 passes the filtered movement commands to the motion software module 106, which produces physical movement commands for the animatronic figure 108.
  • the filter module 502 could connect directly to the animatronic figure 108.
  • a motion of the animatronic figure 108 has an associated frequency content.
  • a single figure motion of the animatronic figure 108 can have an associated frequency content. If the motion of the animatronic figure 108 is changes quickly, the at least a portion of the frequency content includes a high frequency. If the motion of the animatronic figure 108 changes slowly, then at least a portion of the frequency content includes a low frequency.
  • the filter module 502 filters the user-inputted motion according to one of a variety of filters. For instance, the filter module 502 can filter the user-inputted motion according to a low-pass filter, which blocks high frequency content and passes the lower frequency content. The filter module 502 can also filter the user- inputted motion according to a high-pass filter, which blocks low frequency content and passes the higher frequency content. Further, the filter module 502 can filter the user-inputted motion according to a band-pass filter, which passes a limited range of frequencies. The band-pass filter can be a combination of a low-pass filter and a high-pass filter to only pass through certain low frequency content and certain high frequency content.
  • filters can be used by the filter module 502.
  • the filter module 502 filters motions to appear more life-like. For instance, when the user 101 inputs an instruction to move the head of the animatronic figure 108 in a particular direction, the head may normally move so as to reflect both the gross low-frequency turning motion as well as any high frequency jerking or stuttering motion the operator inadvertently added.
  • the filter module 502 filters the motion of the head so that high frequency content is blocked and low frequency content is passed through. Therefore, the head of the animatronic figure 108 will move in a relatively smooth motion. The head of the animatronic figure 108 can still move through the entire range of motion, but will do so more smoothly.
  • Figure 6A illustrates a graph 600 of the filter for unfiltered motion of the head of the animatronic figure 108.
  • the unfiltered motion appears as a horizontal bar 602 that has a constant amplitude at different portions of the frequency content.
  • the horizontal bar 602 represents the motion of the animatronic figure 108, approximately tracking the input of the user 1010. For instance, if the user 101 moves a joystick quickly from left to right, the head of the animatronic figure 108 will move at a movement close to that of the joystick movement.
  • Figure 6B illustrates a graph 610 of the filter for filtered motion of the head of the animatronic figure 108.
  • the portion of the higher frequency content is filtered so that only the lower frequency content passes through.
  • a curve 620 is produced from filtering the horizontal bar 602. As illustrated in the graph 610, the curve 620 has the amplitude in the lower frequency portion that the animatronic figure 108 would produce if filtering did not take place at all. However, the amplitude decreases in the higher frequency portions.
  • the head of the animatronic figure 108 moves without any high frequency jerking or stuttering inadvertently provided by the user 101.
  • the resulting motion provides a life-like motion like the head of a living creature.
  • Figure 7 illustrates an animatronic filter configuration 700, which filters user-inputted motions and algorithmically determined motions in a coordinated manner to produce life-like motions for the animatronic figure 108.
  • a filter module 502 receives a user-inputted instruction from the reception module 102, which receives the user-inputted instruction from the user 101.
  • the algorithmic response module 308 receives a condition from the clock 302, the random number generator 304, and/or the sensor 306. The algorithmic response module 308 utilizes the condition to algorithmically calculate an algorithmically determined motion. Further, the algorithmic response module 308 provides the algorithmically determined motion to the filter module 502.
  • the filter module 502 applies a first filter to the user-inputted motion and applies a second filter to the algorithmically-determined motion in an attempt to coordinate movements of the animatronic figure 108. For instance, a living creature would likely move its eyes in a direction that its head is going to move prior to the head's movement, referred to as an anticipation motion.
  • the filter module 502 can apply a band-pass filter or a high pass filter to the eyes of the animatronic figure 108 to produce this quick anticipatory motion of the eyes. Further, the filter module 502 can apply a low pass filter to the head to produce a relatively slow smooth motion of the head.
  • the filter module 502 then provides the filtered motion of the eyes and the filtered motion of the head to the motion software module 106.
  • the motion software module 106 instructs the animatronic figure 108 to produce the filtered motions.
  • Figure 8A illustrates a graph 800 depicting the filters for the different motions of the animatronic figure 108.
  • a curve 802 represents the filter for the motion of the eyes of the animatronic figure 108.
  • the filter module 502 applies a band-pass filter to the motion of the eyes of the animatronic figure 108. As can be seen from the curve 802, a portion of high frequency content and a portion of low frequency content are blocked. As a result, the remaining mid- frequency content is passed through.
  • the curve 804 is illustrated to represent of the low-pass filter for the head of the animatronic figure 108.
  • a low-pass filter is applied to the motion of the head of the animatronic figure 108.
  • the low-pass filter blocks the high frequency content and pass through the low frequency content. Therefore, the head of the animatronic figure 108 moves in a relatively slow smooth motion compared with the eyes of the animatronic figure 108. As a result, the eyes move first in the direction that the head is about to move to produce a coordinated life-like motion in the animatronic figure 108.
  • a filter is applied to the motion of the middle portion of the neck of the animatronic figure 108.
  • the filter is a slightly lower low- pass filter.
  • a curve 806 represents the low-pass filter of the middle portion of the neck. The low-pass filter blocks out a greater portion of the high frequency content on the middle portion of the neck than is blocked out on the head. Accordingly, the head moves a little before and a little more quickly than the middle portion of the neck to coordinate the head and neck motion in a life-like manner.
  • a filter is also applied to the motion of the base of the neck of the animatronic figure 108.
  • the filter is an even slightly lower low- pass filter.
  • a curve 808 represents the low-pass filter of the base of the neck. The low-pass filter blocks out a greater portion of the high frequency content on the base of the neck than is blocked out on the middle portion of the neck. Accordingly, the middle portion of the neck moves a little before and a little more quickly than the base of the neck to coordinate the head and neck in a life-like manner.
  • Figure 8B illustrates a graph 820 that represents the resulting rotational motion of the body parts of the animatronic figure 108 that were filtered by the filter module 502 using the filters shown in Figure 8A.
  • the step input 822 represents an idealized extreme unfiltered motion that is inputted by the user 101.
  • a curve 824 represents the filtered eye motion.
  • a curve 826 represents the filtered head motion
  • a curve 828 represents the filtered mid-neck motion
  • a curve 830 represents the base of the neck motion.
  • the curve 824 has a very quick jump in rotation because the eyes quickly move in the direction that the other body parts will eventually move. Accordingly, the head, mid-neck, and base of the neck move more slowly.
  • Figure 9A illustrates a process 900 for filtering motions of the animatronic figure 108 based on a user-inputted command and a command determined by the occurrence of the user-inputted command.
  • a first command is received from an operator 101.
  • a second command is determined based on the occurrence of the first command being received.
  • the algorithmic response module 308 algorithmically determines a movement based upon the occurrence of a condition. For instance, the algorithmic response module 308 can provide a command for the animatronic figure 108 to move the eyes of the animatronic figure 108 upon the occurrence of the user 101 providing a command to move the head of the animatronic figure 108. In one embodiment, the eyes move in the direction of the user-inputted motion.
  • a first movement and a second movement are determined.
  • the movements are determined from the respective commands.
  • the translation software module 104 translates the commands into the movements.
  • Figure 9B illustrates a process 920 for filtering motions of the animatronic figure 108 based on a user-inputted command a command that is determined from a stimulus.
  • a first command is received from an operator 101.
  • a stimulus is received from the sensor 306 at the process block 924.
  • a second command is determined based on receiving a stimulus.
  • the algorithmic response module 308 algorithmically determines a movement based upon the occurrence of stimulus. For instance, the algorithmic response module 308 can provide a command for the animatronic to move its chest in. a breathing motion according to a time on the clock 302.
  • a first movement and a second movement are determined.
  • the movements are determined from the respective commands.
  • the translation software module 104 translates the commands into the movements.
  • the first movement and the second movement are filtered to produce a coordinated life-like motion. Accordingly, at a process block 910, the first and second movements are provided to the animatronic figure 108.
  • Show elements are theatric elements that can range from small motions and gestures up to complex interactive motions such as walking. These show elements contain sets of motions for the animatronic figure 108 as well as audio and other computer controlled events such as lighting and video. Show elements are typically built up, using the scripting language, by combining simpler component show elements to produce complex life-like motions.
  • the scripting language and its execution environment are specialized for physical animation and run in a strictly real-time framework.
  • the specialization for physical animation makes common animation tasks simple and efficient.
  • the real-time framework assures that the resulting motions do not stutter or pause, maintaining consistent control and safety.
  • Figure 10 illustrates a layered system 1000 of software modules.
  • the layered system 1000 controls the animatronic figure 108.
  • An operator interface 1002 allows the operator 101 to control the system.
  • a trajectory planner 1004 decides the path that the motors of the animatronic figure 108 should take.
  • a servo loop 1006 directs the motors other actuators to perform the desired motions. The servo loop 1006 also monitors the motors and other actuators to make sure that the motors and other actuators are performing within specified limits.
  • a plurality of low-level hardware drivers 1008 communicated directly with hardware 1010 of the animatronic figure 108.
  • Figure 11 illustrates a detailed view of the trajectory planner 1004.
  • the trajectory planner 1004 plans and implements motions of the animatronic figure 108.
  • the script elements mentioned above are implemented by directing the actuators of the animatronic figure 108 to follow specific trajectories. These trajectories are not necessarily fixed or pre-determined, but typically, change based on external inputs and operator control. This allows spontaneous interaction with guests, responses to the environment, and triggered script elements. In one embodiment, continuously evolving trajectories are produced in a timely manner and are smooth.
  • trajectories are implemented as computing objects, effectively computer programs, which are evaluated on a precise real-time clock. They are created, installed, and then run until the resulting motion completes or is stopped. While a trajectory is running, it is continually being evaluated to determine animatronic figure's next target positions. [00134] Trajectories are evaluated on a strictly periodic cycle. It is not acceptable for the evaluation to pause or not return a value in its allotted time. Such a pause would be indistinguishable from a system failure, and not returning a value after the allotted time would result in motion that is not smooth. Therefore, elements of the computation that could cause the calculation to falter are necessarily moved elsewhere.
  • the trajectory planner 1004 has a trajectory following and execution module 1102. Accordingly, the trajectory following and execution module 1102 handles the real-time task of evaluating trajectories and determining target positions. The trajectory following and execution module 1102 periodically wakes up, evaluates the active trajectories, and returns target positions.
  • the trajectory planner 1004 has a trajectory creation and management module 1104. Accordingly, the trajectory creation and management module 1104 handles all the non-real time needs, particularly the creation of trajectory objects and managing of resources. When a trajectory completes, it and all its resources are handed back to the manager to be cleaned up, that is memory and devices are freed back to the computer to be used elsewhere. This approach of having a non real-time management thread, cooperating program that shares and manages a common memory space, is different and more powerful than typically used elsewhere for resource management.
  • Trajectory objects can be atomic or composite.
  • An atomic trajectory is capable of computing targets position independent of other trajectories.
  • the atomic trajectory uses time or external inputs.
  • a composite trajectory computes its target positions based on the outputs of component trajectories.
  • Figure 12 shows an example of a simple composite trajectory. The atomic sine wave 1202 and envelope trajectories 1204 are multiplied and then the atomic joystick input 1206 is filtered and then added in. This composite trajectory object outputs its result as a single value, the next target position.
  • the trajectory objects in this embodiment provide a generalized computing environment.
  • the example in Figure 12 shows the use of math operators such as multiply 1208 and sum 1212, basic functions such as the atomic sine wave 1202, keyframed data such as the envelope trajectories 1204, and a filter 1210.
  • trajectories can sequence, composite, and transition trajectories, perform logical operators, and handle starting, stopping, and interrupting trajectories. Trajectories can produce multiple outputs, be transformed, loop, and be tested. Finally, trajectories can define an internal computational environment where internal variables, functions, and trajectories can be defined.
  • Keyframes based curves are a common and important tool for animation, computer aided design, and elsewhere.
  • Standard approaches used for keyframe animation are powerful and simple to use but lack the high degree of mathematical smoothness required by physical animation, that is the life-like motion of an animatronic figure or robot.
  • an atomic keyframe trajectory that is used is C 2 continuous.
  • C 2 continuous it is meant that the positions as well as the first two derivatives, the velocity and acceleration, are continuous and smooth.
  • smooth keyframe based curves are implemented. Accordingly, the curve is C 2 continuous. Further, data is interpolated, i.e. the curve passes through, not just approximates, all specified data.
  • keyframes may be non-uniformly spaced. In other words, the time between adjacent pairs of keyframes may be different.
  • Individual keyframes may contain position data alone, position and velocity data, or position, velocity, and acceleration data, all of which is interpolated. Different kinds of keyframes, such as position, position/velocity, or position/velocity/acceleration, may be arbitrarily mixed.
  • Standard approaches to keyframe animation often use Bezier curves or the generalized class of Hermite splines, both of which in their standard form are only C 1 continuous. Higher order Bezier and Hermite splines exist but are not well suited to simple curve editing.
  • the most common Hermite spline for curve editing is the Catmull-Rom method that uses adjacent points to compute the additional needed constraints, i.e., the velocities, at each point. Additional controls can also be added, probably the most common example being Kochanek-Bartels splines, as seen in AUTODESK'S 3D-STUDIO MAX and NEWTEK'S LIGHTWAVE.
  • FIG. 13 illustrates a curve 1300 described by a plurality of keyframes.
  • the curve is described by a set of N keyframes k -i, 2, ... k n.
  • Each keyframe kj contains a time and a position ⁇ .;, x ⁇ , and may optionally also contain a velocity ⁇ tj, X
  • Catmull-Rom computes velocity by simply drawing a line through the neighboring points:
  • Catmull-Rom can perform poorly if the times between keyframes are not fairly equal, sometime requiring curves with high accelerations.
  • Figure 14 illustrates another embodiment in which a curve 1400 described by a plurality of keyframes. A weighted average of two adjacent velocities is implemented. The following equations describe this weighted average:
  • the equation a A + a B a i 2 represents this computation.
  • the component accelerations a A and a B are computed in the following manner.
  • the previous keyframe kn does not have a specified acceleration and is not the first keyframe.
  • the third-order polynomial f a (t) a 0 + ai t + a 2 t 2 + a 3 1 3 such that it passes through ⁇ tn, v, ⁇ and then compute the resulting acceleration at tj:
  • the subsequent keyframe kj+ ⁇ does not have a specified acceleration and is not the last keyframe.
  • the third-order polynomial f b (t) b 0 + bi t + b 2 1 2 + b 3 1 3 such that it passes through ⁇ tj, Xj, Vi ⁇ and ⁇ tj + i, x i+1 , v i+ ⁇ and then compute the resulting acceleration at tj:
  • each keyframe has a position, velocity, and acceleration, either from being specified or from being computed based on the values of adjacent keyframes.
  • These fully populated keyframes ⁇ tj, Xj, Vj, ai ⁇ are now used to compute the resulting curve.
  • FIG 16A illustrates another graph 1600 with an interpolation of data.
  • the graph 1600 shows the standard Catmull-Rom keyframe spline as well as the smooth keyframe, both with and without the acceleration smoothing.
  • Figure 16B shows the corresponding velocities of the graph 1610.
  • acceleration averaging typically reduces the maximum acceleration, and thereby eventual torque required by the robot to realize the motion.
  • trajectory objects Not only should atomic and composite trajectory objects produce smooth motions, but transitions between curves, curves that are interrupted, and curves that start from unanticipated positions must do so as well. This is done by computing and passing software objects, called PVA objects (for position, velocity, and acceleration) that describe the trajectory's current motion. Trajectory objects use PVAs to initialize their motion and return them when they stop or are interrupted. Additionally, it was found that PVA objects can be used to produce trajectories that combine motions in different transform or description spaces.
  • PVA objects for position, velocity, and acceleration
  • Cross fades can be used for any trajectory. It is often useful to transition an internal or transformed trajectory. This causes a smooth transition from one complex motion to a different complex motion in an often unintuitive but powerful way.
  • Some trajectories are able to start from an arbitrary initial position or motion. All trajectory objects are initialized during execution before being evaluated for the first time. This initialization sets a number of parameters such as the start time and the start PVA. Keyframe trajectories, if they have been specified to do so, can use this start PVA as their initial keyframe, a subsequent keyframe in the curve, or as a computational component used by that trajectory object. Thus, trajectories that start with a non-predetermined motion can still compute and follow a smooth trajectory.
  • FIG. 1 Another example is the movement of an animatronic dinosaur, which involves feet positions during walking. While a foot is on the ground, its position is described in terms of ground coordinates, in particular, a foot on the ground should not move or slide along the ground. While a foot is swinging forward in the air during a step, it is described in terms of the body's position.
  • the ability to describe and rapidly try out trajectories provides much of the power and flexibility.
  • the scripting language allows trajectory objects to be described at a high level and then have the computer manage them. The computer is also able to perform some optimizations on the scripts as they are read in, or parsed. These optimizations result in smaller trajectory objects that evaluate more efficiently.
  • the scripting language used in the preferred embodiment is a text-based language that allows trajectories to be defined and manipulated. As a way of explanation, scripts for the trajectories illustrated previously will be shown. The syntax is reminiscent of the computer language C++.
  • Figure 17 is an example script for the simple four point keyframe trajectory shown in Figure 15 used to describe smooth keyframes.
  • the trajectory is just a list of the keyframe positions interspersed with the time between pairs of keyframes.
  • the text to the right of the // are comments ignored by the computer.
  • the resulting trajectory will run for three seconds.
  • Figure 18 illustrates an example script for the composite trajectory shown in Figure 12.
  • a value "joystick” and a trajectory "envelope” are first defined that are then used as components in the final composite trajectory. Further, the second keyframe in the trajectory envelope contains both a position and a velocity term.
  • the animatronic system 100 is intended to cover any system that deals with an unconstrained environment where the operator 101 can come in and either modify or completely replace some or all of the behavior and then potentially hand control back possibly even at a new point.
  • the animatronic system 100 provides the ability to mix and to switch between canned (pre-recorded), computed (typically sensor responses, but also planned), and live (operator controlled) motion to produce a seamless amalgam.
  • the animatronic system 100 can be applied to a variety of different configurations.
  • the animatronic figure 108 can be a surgical robot where a doctor supervises and takes over if the surgical robot overlooks something or otherwise goes astray. If the doctor takes control, the resulting motion appears fluid. If the doctor's intervention was not coordinated with the surgical robot's control of the movements, then a resulting motion may be produced in the transition that may be harmful to the patient.
  • the animatronic figure 108 can be a transportation vehicle, such as an automobile, airplane, space shuttle, or train.
  • An automatic pilot feature in the transportation vehicle can allow an operator to override the automatic pilot with a fluid transition motion that ensures the safety of the vehicle and its passengers.
  • the animatronic system 100 can also handle neglected inputs or mistakes.
  • the animatronic figure 108 can also be a household robot where an owner can redirect the household's robot to different tasks. For example, the owner can redirect the robot to greet guests as opposed to performing a chore. Once again, the transition motion can appear to be life-like.
  • the animatronic figure 108 can also be a hazardous waste cleanup robot used to clean chemical, nuclear, or biological waste, spills or weapons fallout.
  • the user may direct the hazardous waste robot as to which element of waste material is to be placed in a sealed container, but allow the robot to finish the task on its own.
  • the transition of movement between user and robot control is sufficiently smooth to prevent the robot from dropping or otherwise mishandling dangerous waste materials.
  • the animatronic figure 108 can be also be a bomb disposal robot.
  • the user may wish to directly control/guide particular movements of the robot while performing a particularly sensitive defusal operation, but allow the robot to finish disposing of the bomb without user guidance. Again, the transition of movement between user and robot control is smooth to prevent accidental and/or unintentional motions that might disrupt sensitive and possibly dangerous tasks.

Landscapes

  • Engineering & Computer Science (AREA)
  • Chemical & Material Sciences (AREA)
  • Combustion & Propulsion (AREA)
  • Transportation (AREA)
  • Mechanical Engineering (AREA)
  • Toys (AREA)
  • Manipulator (AREA)
  • User Interface Of Digital Computer (AREA)
  • Prostheses (AREA)
  • Rehabilitation Tools (AREA)
PCT/US2004/026115 2004-01-14 2004-08-12 Computing environment that produces realistic motions for an animatronic figure Ceased WO2005069755A2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP04780885.2A EP1711241A4 (en) 2004-01-14 2004-08-12 CALCULATION ENVIRONMENT PRODUCING REALISTIC MOVEMENTS FOR ANIMATRONIC FIGURE
JP2006549231A JP2007517585A (ja) 2004-01-14 2004-08-12 アニマトロニクス・フィギュアの現実的な動作を生成する計算環境

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US10/757,797 2004-01-14
US10/757,797 US7238079B2 (en) 2003-01-14 2004-01-14 Animatronic supported walking system

Publications (2)

Publication Number Publication Date
WO2005069755A2 true WO2005069755A2 (en) 2005-08-04
WO2005069755A3 WO2005069755A3 (en) 2006-03-30

Family

ID=34740094

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2004/026115 Ceased WO2005069755A2 (en) 2004-01-14 2004-08-12 Computing environment that produces realistic motions for an animatronic figure

Country Status (4)

Country Link
US (4) US7238079B2 (enExample)
EP (1) EP1711241A4 (enExample)
JP (1) JP2007517585A (enExample)
WO (1) WO2005069755A2 (enExample)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008259573A (ja) * 2007-04-10 2008-10-30 Tomy Co Ltd 遠隔操作装置及び玩具
US9057508B1 (en) 2014-10-22 2015-06-16 Codeshelf Modular hanging lasers to enable real-time control in a distribution center
US9262741B1 (en) 2015-04-28 2016-02-16 Codeshelf Continuous barcode tape based inventory location tracking
US9327397B1 (en) 2015-04-09 2016-05-03 Codeshelf Telepresence based inventory pick and place operations through robotic arms affixed to each row of a shelf

Families Citing this family (40)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7238079B2 (en) * 2003-01-14 2007-07-03 Disney Enterprise, Inc. Animatronic supported walking system
US8368700B1 (en) * 2006-12-18 2013-02-05 Pixar Animatronics animation method and apparatus
US8060255B2 (en) * 2007-09-12 2011-11-15 Disney Enterprises, Inc. System and method of distributed control of an interactive animatronic show
US7706920B2 (en) * 2007-09-19 2010-04-27 Disney Enterprises, Inc. Smooth clipping of data streams
CN101411946B (zh) * 2007-10-19 2012-03-28 鸿富锦精密工业(深圳)有限公司 玩具恐龙
WO2009079514A1 (en) * 2007-12-17 2009-06-25 Pixar Methods and apparatus for designing animatronics units from articulated computer generated characters
US8390629B2 (en) * 2007-12-17 2013-03-05 Pixar Methods and apparatus for designing animatronics units from articulated computer generated characters
US8232998B2 (en) * 2007-12-17 2012-07-31 Pixar Methods and apparatus for estimating and controlling behavior of animatronics units
US8564547B2 (en) * 2008-04-15 2013-10-22 Mattel, Inc. Touch screen remote control device for use with a toy
US8336420B2 (en) 2010-06-02 2012-12-25 Disney Enterprises, Inc. Three-axis robotic joint using four-bar linkages to drive differential side gears
US8956198B1 (en) 2010-09-27 2015-02-17 Mattel, Inc. Walking toy
US8761927B2 (en) 2010-12-17 2014-06-24 Disney Enterprises, Inc. Externally actuated figure
US9345976B2 (en) * 2011-08-29 2016-05-24 Mattel, Inc. Toy figurine with removable features
US20130109272A1 (en) * 2011-10-31 2013-05-02 Stephen M. RINDLISBACHER Method of Controlling a Vehicle or Toy via a Motion-Sensing Device and/or Touch Screen
US20140099857A1 (en) * 2012-10-04 2014-04-10 Mattel, Inc. Toy Figurine with Projectiles
US9130492B2 (en) * 2013-04-22 2015-09-08 Thermadyne, Inc. Animatronic system with unlimited axes
JP6539035B2 (ja) * 2014-01-08 2019-07-03 ローム株式会社 チップ部品
US10034717B2 (en) 2014-03-17 2018-07-31 Intuitive Surgical Operations, Inc. System and method for breakaway clutching in an articulated arm
US9457466B2 (en) 2014-08-15 2016-10-04 Disney Enterprises, Inc. Pneumatically actuated and safely compliant skeletal joints for robotic characters
US10112119B2 (en) * 2015-11-09 2018-10-30 Disney Enterprises, Inc. Method for modifying local properties of materials
CN105751248B (zh) * 2016-05-09 2017-10-13 上海理工大学 仿生液压四足机器恐龙头颈部结构
US20180016806A1 (en) * 2016-07-14 2018-01-18 Anissa Bullard Animatronic Agitating Assembly
CN107351060B (zh) * 2017-07-10 2020-07-14 上海理工大学 一种仿生机器人头颈部柔性结构
KR102509678B1 (ko) * 2017-12-29 2023-03-14 프라운호퍼 게젤샤프트 쭈르 푀르데룽 데어 안겐반텐 포르슝 에. 베. 대상물 모션 예측 방법, 모션 모델 보정 방법, 기정 양 도출 방법 및 가상 현실 뷰 생성 방법
US10967508B2 (en) * 2018-02-15 2021-04-06 DMAI, Inc. System and method for dynamic robot configuration for enhanced digital experiences
WO2019160613A1 (en) * 2018-02-15 2019-08-22 DMAI, Inc. System and method for dynamic program configuration
JP2021142572A (ja) * 2018-06-12 2021-09-24 ソニーグループ株式会社 制御装置、制御方法及びプログラム
US11007451B2 (en) 2019-01-10 2021-05-18 Universal City Studios Llc Interactive character control system
US11541549B2 (en) 2019-02-14 2023-01-03 Universal City Studios Llc Mobile character control system
US12099997B1 (en) 2020-01-31 2024-09-24 Steven Mark Hoffberg Tokenized fungible liabilities
US11995943B2 (en) 2020-08-11 2024-05-28 ScooterBug, Inc. Methods of and systems for controlling access to networked devices provided with machine-readable codes scanned by mobile phones and computing devices
US12479521B2 (en) 2020-08-11 2025-11-25 ScooterBug, Inc. Wireless network, systems and methods for configuration, operation and use of motorized, wheeled and steerable passenger-carrying carts for entertainment purposes
US11790722B2 (en) 2020-08-11 2023-10-17 Best Lockers, Llc Single-sided storage locker systems accessed and controlled using machine-readable codes scanned by mobile phones and computing devices
US11631295B2 (en) 2020-08-11 2023-04-18 ScooterBug, Inc. Wireless network, mobile systems and methods for controlling access to lockers, strollers, wheel chairs and electronic convenience vehicles provided with machine-readable codes scanned by mobile phones and computing devices
US11998856B2 (en) 2020-12-04 2024-06-04 Universal City Studios Llc Conical-shaped articulated member
US12029998B2 (en) 2020-12-07 2024-07-09 Universal City Studios Llc Flexible shell structure for show elements
US12501225B2 (en) 2022-01-02 2025-12-16 Poltorak Technologies Llc Bluetooth enabled intercom with hearing aid functionality
US20230390653A1 (en) * 2022-06-03 2023-12-07 Universal City Studios Llc Smoothing server for processing user interactions to control an interactive asset
CN117262067B (zh) * 2023-09-20 2025-02-07 哈尔滨工业大学 髋关节结构、腿部结构及六自由度低惯量的机器人仿生腿
CN119705668B (zh) * 2024-12-23 2025-11-21 浙江工业大学 一种可适应非结构化环境的足底稳定装置及机器人

Family Cites Families (137)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US202433A (en) * 1878-04-16 Improvement in distance and velocity recorders
US6337552B1 (en) * 1999-01-20 2002-01-08 Sony Corporation Robot apparatus
US2108229A (en) * 1935-07-05 1938-02-15 Martha M Metz Anatomical skeleton
US4202423A (en) 1978-04-20 1980-05-13 Soto Jose M Land vehicle with articulated legs
FR2512410A1 (fr) 1981-09-04 1983-03-11 Kroczynski Patrice Systeme de robots a pattes
JPS5851090A (ja) * 1981-09-21 1983-03-25 株式会社東芝 関節装置
US4511011A (en) 1983-03-18 1985-04-16 Odetics, Inc. Leg mechanism for walking machine
US4527650A (en) 1983-03-18 1985-07-09 Odetics, Inc. Walking machine
US4503924A (en) 1983-03-18 1985-03-12 Odetics, Inc. Horizontal actuator mechanism for the legs of a walking machine
JPS6055985A (ja) * 1983-09-05 1985-04-01 株式会社トミー 音声認識玩具
US4558758A (en) 1983-12-02 1985-12-17 Erwin Littman Prime mover
DE3417868C2 (de) * 1984-05-14 1986-08-07 Deutsche Forschungs- und Versuchsanstalt für Luft- und Raumfahrt e.V., 5000 Köln Verfahren zum Steuern von Roboterbewegungen
US4712184A (en) * 1984-09-12 1987-12-08 Haugerud Albert R Computer controllable robotic educational toy
US4747127A (en) * 1985-12-23 1988-05-24 American Telephone And Telegraph Company, At&T Bell Laboratories Customer programmable real-time system
US4695977A (en) 1985-12-23 1987-09-22 American Telephone And Telegraph Company And At&T Bell Laboratories Control of real-time systems utilizing a nonprocedural language
US4821463A (en) * 1986-01-24 1989-04-18 The Boeing Company Drill bit sharpening apparatus
JP2525358B2 (ja) * 1986-01-25 1996-08-21 フアナツク株式会社 ロボツト制御装置
JPS62232006A (ja) * 1986-04-02 1987-10-12 Yokogawa Electric Corp ロボツト・システム
JPS6318403A (ja) * 1986-07-10 1988-01-26 Fanuc Ltd オフライン制御実行方法
GB8625144D0 (en) 1986-10-21 1986-11-26 Hennequin J R Pneumatic/hydraulic artificial muscle
US4774445A (en) * 1986-11-20 1988-09-27 Unimation, Inc. Multiaxis robot control having capability for executing timed moves
US4868474A (en) * 1986-11-20 1989-09-19 Westinghouse Electric Corp. Multiprocessor position/velocity servo control for multiaxis digital robot control system
US4835730A (en) * 1987-02-27 1989-05-30 Adept Technology, Inc. Database driven robot programming system and method
JPS63256381A (ja) * 1987-04-10 1988-10-24 フアナツク株式会社 ロボツト制御装置
US4840242A (en) 1988-03-03 1989-06-20 Chih Liu P Convertible animal toy
US4923428A (en) * 1988-05-05 1990-05-08 Cal R & D, Inc. Interactive talking toy
GB8906541D0 (en) 1989-03-21 1989-05-04 Portsmouth Tech Consult Robot devices
JP2520019B2 (ja) 1989-06-29 1996-07-31 本田技研工業株式会社 脚式移動ロボットの駆動制御装置
US5021878A (en) * 1989-09-20 1991-06-04 Semborg-Recrob, Corp. Animated character system with real-time control
JP2592340B2 (ja) 1989-12-14 1997-03-19 本田技研工業株式会社 脚式歩行ロボットの関節構造
JPH0390922U (enExample) * 1989-12-28 1991-09-17
US5303384A (en) * 1990-01-02 1994-04-12 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration High level language-based robotic control system
EP0446395B1 (de) * 1990-03-15 1994-12-07 Martin Dr. Mohr Verfahren und Schaltungsanordnung zur Realisierung der mit vitalen Gestalten-Gattungsmimiken und Gestalten-Gattungs-Choreographien gleichwertigen, künstlichen Gestalten-Gattungs-Mimiken und Gestalten-Gattungs-Choreographien bei der Animation von künstlich bewegbaren Gestalten-Gattungen
FI87171C (fi) 1990-08-14 1992-12-10 Plustech Oy Svaengarm
DE69124486T2 (de) 1990-11-30 1997-05-15 Honda Motor Co Ltd System zur Steuerung der Fortbewegung eines Schreitroboters mit Beinen
US5088953A (en) * 1991-01-10 1992-02-18 Robert Richman Gravity activated walking toy
JP2921132B2 (ja) * 1991-01-31 1999-07-19 株式会社安川電機 産業用ロボットの手首機構
US5158493A (en) * 1991-05-30 1992-10-27 Richard Morgrey Remote controlled, multi-legged, walking robot
US5157316A (en) 1991-06-26 1992-10-20 Glovier Lloyd H Robotic joint movement device
US5302148A (en) * 1991-08-16 1994-04-12 Ted Heinz Rotatable demountable blocks of several shapes on a central elastic anchor
US5355064A (en) * 1992-03-04 1994-10-11 Honda Giken Kogyo Kabushiki Kaisha Control system for legged mobile robot
US5270480A (en) * 1992-06-25 1993-12-14 Victor Company Of Japan, Ltd. Toy acting in response to a MIDI signal
US5297443A (en) 1992-07-07 1994-03-29 Wentz John D Flexible positioning appendage
US5404086A (en) * 1992-07-20 1995-04-04 Honda Giken Kogyo Kabushiki Kaisha System for controlling locomotion of legged mobile robot and correcting inclinometer's output thereof
US5363935A (en) * 1993-05-14 1994-11-15 Carnegie Mellon University Reconfigurable mobile vehicle with magnetic tracks
US5434489A (en) * 1993-07-30 1995-07-18 Fanuc Robotics North America, Inc. Method and system for path planning in cartesian space
JPH07100309B2 (ja) 1993-09-30 1995-11-01 工業技術院長 歩行ロボットのための腕に転用可能な脚機構
US5511147A (en) * 1994-01-12 1996-04-23 Uti Corporation Graphical interface for robot
US5644204A (en) 1994-11-03 1997-07-01 Nagle; John Anti-slip control for a legged robot and realisitc simulation of a legged creature
JP2560264B2 (ja) 1994-12-02 1996-12-04 工業技術院長 多足歩行装置
JPH08174378A (ja) 1994-12-27 1996-07-09 Nippon Telegr & Teleph Corp <Ntt> ロボット制御装置
US5724074A (en) * 1995-02-06 1998-03-03 Microsoft Corporation Method and system for graphically programming mobile toys
US5674127A (en) * 1995-03-07 1997-10-07 Habilas, Inc. Multisite multiplayer interactive electronic entertainment system having a partially player defined universe
US5913727A (en) * 1995-06-02 1999-06-22 Ahdoot; Ned Interactive movement and contact simulation game
DE19521358C1 (de) * 1995-06-12 1996-09-05 Siemens Ag Verfahren zur Schlupfermittlung einer autonomen mobilen Einheit mit Dreiradkinematik
US5808433A (en) 1995-09-29 1998-09-15 Honda Giken Kogyo Kabushiki Kaisha Method of generating gait of legged walking robot and system for controlling its locomotion
US20010031652A1 (en) * 1995-11-20 2001-10-18 Creator Ltd. 1*doll
US6368177B1 (en) * 1995-11-20 2002-04-09 Creator, Ltd. Method for using a toy to conduct sales over a network
US5752880A (en) * 1995-11-20 1998-05-19 Creator Ltd. Interactive doll
US5784541A (en) * 1996-01-18 1998-07-21 Ruff; John D. System for controlling multiple controllable devices according to a script transmitted from a personal computer
US6230078B1 (en) * 1996-01-18 2001-05-08 John D. Ruff Simplified animatronic and CNC system
JP3662996B2 (ja) 1996-01-25 2005-06-22 本田技研工業株式会社 脚式移動ロボットの歩行制御装置
US5628667A (en) * 1996-02-26 1997-05-13 Levi; Shimon Sinuous toy
US6553410B2 (en) * 1996-02-27 2003-04-22 Inpro Licensing Sarl Tailoring data and transmission protocol for efficient interactive data transactions over wide-area networks
US5746602A (en) * 1996-02-27 1998-05-05 Kikinis; Dan PC peripheral interactive doll
JPH10137439A (ja) 1996-11-08 1998-05-26 Sony Corp ロボツト装置
US5929585A (en) 1996-11-19 1999-07-27 Sony Corporation Robot system and its control method
US6249278B1 (en) * 1997-01-07 2001-06-19 The Animated Animations Company Llc. Modular dynamic dialogue animated display device
CA2225060A1 (en) * 1997-04-09 1998-10-09 Peter Suilun Fong Interactive talking dolls
CA2290527A1 (en) * 1997-05-19 1998-11-26 Creator Ltd. Programmable assembly toy
IL120857A (en) * 1997-05-19 2003-03-12 Creator Ltd Programmable assembly toy
IL120856A0 (en) * 1997-05-19 1997-09-30 Creator Ltd Controllable toy system operative in conjunction with a household audio entertainment player
IL121574A0 (en) * 1997-08-18 1998-02-08 Creator Ltd Techniques and apparatus for entertainment sites amusement parks and other information and/or entertainment dispensing sites
CA2213287A1 (en) 1997-08-18 1999-02-18 Clement Tessier Apparatus for shielding an articulated structure
IL121642A0 (en) * 1997-08-27 1998-02-08 Creator Ltd Interactive talking toy
CN1113678C (zh) 1997-10-15 2003-07-09 Cj合伙有限公司 带有可活动部件和模制柔软表层的玩具模型
JP3936749B2 (ja) * 1998-04-16 2007-06-27 クリエイター・リミテッド 対話型玩具
US6056618A (en) * 1998-05-26 2000-05-02 Larian; Isaac Toy character with electronic activities-oriented game unit
KR100529287B1 (ko) * 1998-06-09 2005-11-17 소니 가부시끼 가이샤 로봇 장치 및 그 자세 제어 방법
US6053797A (en) * 1998-07-17 2000-04-25 Eastgate Innovations Incorporated Interactive toy
JP2000153476A (ja) * 1998-09-14 2000-06-06 Honda Motor Co Ltd 脚式移動ロボット
US6192215B1 (en) 1998-10-23 2001-02-20 Mai Wang Interactive and animated mini-theater and method of use
JP2000135384A (ja) * 1998-10-30 2000-05-16 Fujitsu Ltd 情報処理装置及び擬似生物機器
US6149490A (en) * 1998-12-15 2000-11-21 Tiger Electronics, Ltd. Interactive toy
CN1293607A (zh) * 1999-01-28 2001-05-02 索尼公司 机器人装置用关节装置及脚式步行机器人装置
JP2000254360A (ja) * 1999-03-11 2000-09-19 Toybox:Kk 対話型玩具
US6629087B1 (en) * 1999-03-18 2003-09-30 Nativeminds, Inc. Methods for creating and editing topics for virtual robots conversing in natural language
US6198247B1 (en) * 1999-04-20 2001-03-06 Steven Barr Servo-articulated modules and robotic assemblies incorporating them
US6264521B1 (en) * 1999-06-15 2001-07-24 HERNáNDEZ JOSé M. Doll with video control
US6663393B1 (en) * 1999-07-10 2003-12-16 Nabil N. Ghaly Interactive play device and method
JP3443077B2 (ja) * 1999-09-20 2003-09-02 ソニー株式会社 ロボットの運動パターン生成装置及び運動パターン生成方法、並びにロボット
JP3555107B2 (ja) * 1999-11-24 2004-08-18 ソニー株式会社 脚式移動ロボット及び脚式移動ロボットの動作制御方法
JP2001154681A (ja) * 1999-11-30 2001-06-08 Sony Corp 音声処理装置および音声処理方法、並びに記録媒体
JP2001157976A (ja) * 1999-11-30 2001-06-12 Sony Corp ロボット制御装置およびロボット制御方法、並びに記録媒体
US6937289B1 (en) * 1999-12-30 2005-08-30 Microsoft Corporation Method and system for downloading and storing interactive device content using the horizontal overscan portion of a video signal
US6719604B2 (en) * 2000-01-04 2004-04-13 Thinking Technology, Inc. Interactive dress-up toy
US6620024B2 (en) * 2000-02-02 2003-09-16 Silverlit Toys Manufactory, Ltd. Computerized toy
US6377281B1 (en) * 2000-02-17 2002-04-23 The Jim Henson Company Live performance control of computer graphic characters
KR100375699B1 (ko) * 2000-03-10 2003-03-15 연규범 토이와 연계한 인터넷 서비스 시스템
JP2001260063A (ja) 2000-03-21 2001-09-25 Sony Corp 多関節型ロボット及びその動作制御方法
ATE540347T1 (de) * 2000-05-01 2012-01-15 Irobot Corp Verfahren und vorrichtung zur steuerung eines mobilen roboters
US6462498B1 (en) * 2000-05-09 2002-10-08 Andrew J. Filo Self-stabilizing walking apparatus that is capable of being reprogrammed or puppeteered
JP2001322079A (ja) * 2000-05-15 2001-11-20 Sony Corp 脚式移動ロボット及びその動作教示方法
US20020029388A1 (en) * 2000-06-22 2002-03-07 Bernd Heisele Interactive toy system
JP2002018146A (ja) * 2000-07-04 2002-01-22 Tomy Co Ltd 対話型玩具、反応行動パターン生成装置および反応行動パターン生成方法
TW477237U (en) * 2000-07-05 2002-02-21 Elan Microelectronics Corp Interactive toy device using ultrasound for transmitting signals
US6537130B1 (en) 2000-09-07 2003-03-25 C.J. Associates, Ltd. Jointed support system and method of constructing same
KR20020067921A (ko) * 2000-10-23 2002-08-24 소니 가부시끼 가이샤 각식 로봇 및 각식 로봇의 행동 제어 방법, 및 기억 매체
JP4119080B2 (ja) * 2000-11-17 2008-07-16 本田技研工業株式会社 人間型ロボットの腕構造
JP3655233B2 (ja) 2000-11-17 2005-06-02 本田技研工業株式会社 二足ロボットの遠隔操作装置
US6669527B2 (en) * 2001-01-04 2003-12-30 Thinking Technology, Inc. Doll or toy character adapted to recognize or generate whispers
US6871563B2 (en) * 2001-02-26 2005-03-29 Howie Choset Orientation preserving angular swivel joint
AU2002258422A1 (en) * 2001-02-27 2002-09-12 Anthrotronix, Inc. Robotic apparatus and wireless communication system
US6543307B2 (en) 2001-04-06 2003-04-08 Metrica, Inc. Robotic system
US6507773B2 (en) 2001-06-14 2003-01-14 Sharper Image Corporation Multi-functional robot with remote and video system
US6454624B1 (en) * 2001-08-24 2002-09-24 Xerox Corporation Robotic toy with posable joints
US20030110040A1 (en) * 2001-12-07 2003-06-12 Creative Logic Solutions Inc. System and method for dynamically changing software programs by voice commands
JP2003181784A (ja) * 2001-12-19 2003-07-02 Sanyo Electric Co Ltd 自動応答ロボット
US6800013B2 (en) * 2001-12-28 2004-10-05 Shu-Ming Liu Interactive toy system
JP3674789B2 (ja) * 2001-12-28 2005-07-20 本田技研工業株式会社 脚式移動ロボットの歩容生成装置
DE20320263U1 (de) * 2002-02-08 2004-04-15 Educo International, Inc. Drahtlabyrinthspielzeug
US6773327B1 (en) 2002-02-12 2004-08-10 Hasbro, Inc. Apparatus for actuating a toy
US20030162475A1 (en) * 2002-02-28 2003-08-28 Pratte Warren D. Interactive toy and method of control thereof
CN100333813C (zh) 2002-05-31 2007-08-29 马特尔公司 可变形的玩偶和可摆姿势的动作人物造型
JP4133188B2 (ja) * 2002-10-07 2008-08-13 株式会社ハーモニック・ドライブ・システムズ ロボットハンドの指ユニット
KR100500842B1 (ko) * 2002-10-31 2005-07-12 삼성광주전자 주식회사 로봇청소기와, 그 시스템 및 제어방법
US7245989B2 (en) * 2002-12-20 2007-07-17 Brooks Automation, Inc. Three-degree-of-freedom parallel robot arm
US7238079B2 (en) * 2003-01-14 2007-07-03 Disney Enterprise, Inc. Animatronic supported walking system
USD496952S1 (en) 2003-01-31 2004-10-05 Td Industrial Coverings Robot cover
USD491964S1 (en) 2003-01-31 2004-06-22 Td Industrial Coverings Robot cover
US6808437B2 (en) * 2003-02-04 2004-10-26 Senario Llc Extendible link soft or plush toy
US7113848B2 (en) 2003-06-09 2006-09-26 Hanson David F Human emulation robot system
US20050283043A1 (en) * 2003-11-06 2005-12-22 Sisk Bradley G Self-contained, submersible, autonomous, speaking android
US7734375B2 (en) * 2004-06-09 2010-06-08 Boston Dynamics Robot and robot leg mechanism
TWI257214B (en) * 2004-12-10 2006-06-21 Univ Nat Chiao Tung Brainwave-controlled embedded Internet robot agent architecture
KR100619786B1 (ko) * 2005-05-17 2006-09-06 엘지전자 주식회사 로봇 청소기의 미끄럼 방지 장치
US7837531B2 (en) * 2005-10-31 2010-11-23 Les Friedland Toy doll

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
See references of EP1711241A4 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2008259573A (ja) * 2007-04-10 2008-10-30 Tomy Co Ltd 遠隔操作装置及び玩具
US9057508B1 (en) 2014-10-22 2015-06-16 Codeshelf Modular hanging lasers to enable real-time control in a distribution center
US9157617B1 (en) 2014-10-22 2015-10-13 Codeshelf Modular hanging lasers to provide easy installation in a distribution center
US9327397B1 (en) 2015-04-09 2016-05-03 Codeshelf Telepresence based inventory pick and place operations through robotic arms affixed to each row of a shelf
US9262741B1 (en) 2015-04-28 2016-02-16 Codeshelf Continuous barcode tape based inventory location tracking

Also Published As

Publication number Publication date
EP1711241A4 (en) 2013-11-20
US7238079B2 (en) 2007-07-03
US20070021032A1 (en) 2007-01-25
US20040219861A1 (en) 2004-11-04
US20070021031A1 (en) 2007-01-25
WO2005069755A3 (en) 2006-03-30
JP2007517585A (ja) 2007-07-05
EP1711241A2 (en) 2006-10-18
US20050153624A1 (en) 2005-07-14
US7819719B2 (en) 2010-10-26
US8374724B2 (en) 2013-02-12

Similar Documents

Publication Publication Date Title
US8374724B2 (en) Computing environment that produces realistic motions for an animatronic figure
Shapiro Building a character animation system
Laszlo et al. Interactive control for physically-based animation
US6057859A (en) Limb coordination system for interactive computer animation of articulated characters with blended motion data
US6088042A (en) Interactive motion data animation system
Pan et al. Realistic and interactive robot gaze
WO1997024696A3 (en) Computer-assisted animation construction system and method and user interface
CA2968589C (en) System and method for placing a character animation at a location in a game environment
EP1012791A1 (en) Limb coordination system for interactive computer animation of articulated characters with blended motion data
JP2005342873A (ja) ロボット装置及びその動作制御方法
US20180001201A1 (en) System and method for determining curved path of travel for a character in cover mode in a game environment
Lamberti et al. Designing interactive robotic games based on mixed reality technology
Wu et al. Simple steps for simply stepping
Martín et al. Humanoid soccer player design
Zannatha et al. Behavior control for a humanoid soccer player using Webots
JP2003136439A (ja) ロボット装置及びロボット装置の歩行制御方法並びにロボット装置の歩行制御プログラム
Cunha et al. Learning robotic soccer controllers with the Q-Batch update-rule
EP4570452A2 (en) Rapid design and animation of freely-walking robotic devices
Martins et al. Sensor Kinect in a telepresence application
Wu Goal-Directed Biped Stepping and Push Recovery with Momentum Control
EP3254742A1 (en) System and method for placing a character animation at a location in a game environment
Saito et al. Animating Various Characters using Arm Gestures in Virtual Reality Environment
JP2025518327A (ja) ユーザインタラクションを処理してインタラクティブアセットを制御する平滑化サーバ
KR20240057416A (ko) 제어가능 객체의 적어도 하나의 특성을 제어하기 위한 방법, 관련된 시스템 및 관련된 디바이스
Jenkins et al. Primitives and Behavior-Based Architectures for Interactive Entertainment

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

WWE Wipo information: entry into national phase

Ref document number: 2006549231

Country of ref document: JP

NENP Non-entry into the national phase

Ref country code: DE

WWW Wipo information: withdrawn in national office

Country of ref document: DE

WWE Wipo information: entry into national phase

Ref document number: 2004780885

Country of ref document: EP

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWP Wipo information: published in national office

Ref document number: 2004780885

Country of ref document: EP