US9638497B2 - Improvised explosive device defeat system - Google Patents
Improvised explosive device defeat system Download PDFInfo
- Publication number
- US9638497B2 US9638497B2 US14/667,048 US201514667048A US9638497B2 US 9638497 B2 US9638497 B2 US 9638497B2 US 201514667048 A US201514667048 A US 201514667048A US 9638497 B2 US9638497 B2 US 9638497B2
- Authority
- US
- United States
- Prior art keywords
- robotic arm
- macro
- micro
- control
- movement
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
- 239000002360 explosive Substances 0.000 title description 6
- 238000000034 method Methods 0.000 claims description 45
- 238000003032 molecular docking Methods 0.000 claims description 12
- 230000007704 transition Effects 0.000 claims description 7
- 238000004891 communication Methods 0.000 description 15
- 210000003811 finger Anatomy 0.000 description 7
- 230000004044 response Effects 0.000 description 5
- 230000006870 function Effects 0.000 description 4
- 238000012545 processing Methods 0.000 description 4
- 238000007789 sealing Methods 0.000 description 4
- 230000007246 mechanism Effects 0.000 description 3
- 101000963759 Homo sapiens Melanocortin-2 receptor accessory protein Proteins 0.000 description 2
- 102100040147 Melanocortin-2 receptor accessory protein Human genes 0.000 description 2
- 230000009471 action Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 238000013519 translation Methods 0.000 description 2
- 230000006399 behavior Effects 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 239000012530 fluid Substances 0.000 description 1
- 230000007274 generation of a signal involved in cell-cell signaling Effects 0.000 description 1
- 210000004247 hand Anatomy 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000000691 measurement method Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000001681 protective effect Effects 0.000 description 1
- 210000003813 thumb Anatomy 0.000 description 1
Images
Classifications
-
- F—MECHANICAL ENGINEERING; LIGHTING; HEATING; WEAPONS; BLASTING
- F41—WEAPONS
- F41H—ARMOUR; ARMOURED TURRETS; ARMOURED OR ARMED VEHICLES; MEANS OF ATTACK OR DEFENCE, e.g. CAMOUFLAGE, IN GENERAL
- F41H11/00—Defence installations; Defence devices
- F41H11/12—Means for clearing land minefields; Systems specially adapted for detection of landmines
- F41H11/16—Self-propelled mine-clearing vehicles; Mine-clearing devices attachable to vehicles
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J3/00—Manipulators of master-slave type, i.e. both controlling unit and controlled unit perform corresponding spatial movements
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1628—Programme controls characterised by the control loop
- B25J9/1633—Programme controls characterised by the control loop compliant, force, torque control, e.g. combined with position control
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1679—Programme controls characterised by the tasks executed
- B25J9/1689—Teleoperation
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05D—SYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
- G05D1/00—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
- G05D1/0011—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots associated with a remote control arrangement
- G05D1/0016—Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots associated with a remote control arrangement characterised by the operator's input device
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/40—Robotics, robotics mapping to robotics vision
- G05B2219/40404—Separate master controls macro and microslave manipulator
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S901/00—Robots
- Y10S901/01—Mobile robot
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S901/00—Robots
- Y10S901/02—Arm motion controller
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S901/00—Robots
- Y10S901/14—Arm movement, spatial
- Y10S901/15—Jointed arm
Definitions
- the invention concerns explosive devices, and more particularly systems and methods for defeating improvised explosive devices (IEDs) using robotic equipment.
- IEDs improvised explosive devices
- Unmanned ground vehicles are commonly used for defeating explosive devices, and more specifically, IED devices.
- Such UGVs commonly include a remotely operated battery powered vehicle which is capable of moving under its own power.
- One problem with UGVs is that they frequently have trouble navigating in cluttered environments. In such environments, the UGV can experience a loss of mobility when obstacles block its path or otherwise inhibit its movement.
- UGVs When UGVs are used for ordinance disposal, they usually include a robotic arm and a video camera which allow a user to disable an IED from a safe distance.
- conventional UGVs tend to have relatively primitive control systems for controlling movement of the distal end of the robotic arm, where a gripping device is typically located. Such control systems commonly allow the robotic arm to be moved with four or five degrees of freedom, but often require the operator to separately control each joint. Consequently, these robot arm control systems are not very intuitive, and contribute to making the robot arm difficult to control precisely.
- MRAP mine resistant ambush protected
- Buffalo which is manufactured by Force Protection, Inc.
- the arm which is similar to the articulated arm of a backhoe, can be used to allow personnel to dispose of IEDs while remaining enclosed in the relative safety of the vehicle.
- the robot arms used on the MRAPs suffer from many of the same problems as the robot arm of the UGV, and are even more difficult to control with precision.
- the invention concerns a robot system, including a control system.
- the control system includes a robot control interface having a control interface grip.
- the robot system includes at least two robotic arms, namely a macro robotic arm and a micro robotic arm.
- the macro robotic arm has a first set of articulating joints.
- the robot system is arranged such that the macro robotic arm will respond, in a first control system state, to movement of the control interface grip.
- the macro robotic arm will move in a plurality of directions responsive to corresponding movement of the interface grip.
- the micro robotic arm has a second set of articulating joints distinct from the first set.
- the robot system is arranged so that the micro robotic arm will respond, in a second control system state, to movement of the control interface grip.
- the micro robotic arm will move in a plurality of directions responsive to corresponding movement of the interface grip.
- the invention also concerns a method for controlling a robot system.
- the method includes controlling a movement of a macro robotic arm having a first set of articulating joints in a first system state.
- the macro robotic arm is responsive to a user manipulation of a control interface grip.
- the macro robotic arm will move in a plurality of directions in response to corresponding movement of the control interface grip.
- the method further involves controlling in a second system state a movement of a micro robotic arm having a second set of articulating joints distinct from the first set.
- the micro robotic arm will move in a plurality of directions in response to corresponding movement of the control interface grip.
- the method includes controlling a movement of a vehicle base to which the micro robotic arm is attached responsive to a user manipulation of the control interface grip.
- the invention concerns a method for performing robotic operations on a work piece.
- the method includes communicating with a macro control system to control a pose of a macro robotic arm.
- the method also includes selectively varying the pose of the macro robotic arm to transport a micro robotic arm to a work location.
- the method continues while the micro robotic arm is at such work location, by communicating with a micro control system separate from said macro control system to control a pose of said micro robotic arm.
- the invention includes performing robotic operations on a work piece.
- the method involves communicating with a macro control system to control a pose of a macro robotic arm, and then selectively varying the pose of said macro robotic arm to move an unmanned ground vehicle (UGV) from a first location to a second location. While at the second location, the UGV is released from a docking station of the macro robotic arm. Thereafter, the method continues with operating a motorized vehicle base of the UGV to move the UGV from the second location to a third location. Finally, while the UGV is in the third location, a pose of a micro robotic arm attached to said UGV is controlled to perform a robotic task.
- the method can include controlling UGV movement using the same interface as used to control the macro and micro robot arms.
- FIG. 1 is a drawing that shows a control system for controlling movement of a macro robotic arm mounted to a shelter, and a UGV having a micro robotic arm.
- FIG. 2 is a drawing that is useful for understanding how the macro robotic arm of FIG. 1 can facilitate a positioning of a micro robotic arm disposed on the UGV.
- FIG. 3 is a drawing that shows how the UGV of FIG. 1 can be deployed at a worksite by using the macro robotic arm of FIG. 1 .
- FIG. 4 is a drawing which is useful for understanding an alternative embodiment of the invention in which the macro robotic arm of FIG. 1 can position a micro robotic arm at a worksite when access is limited by an obstruction.
- FIG. 5 is a drawing that is useful for understanding a control system that can be used with the robot system in FIGS. 1-4 .
- FIGS. 6A and 6B are drawings which show certain details of the control system in FIG. 5 .
- FIG. 7 is a drawing that is useful for understanding certain control circuitry associated with the control system in FIG. 5 .
- the invention concerns a robot system 50 and a method of using same.
- the robot system includes a control system 101 .
- the control system including a control interface having a control interface grip 102 .
- the control interface grip 102 and associated control system 101 is arranged for controlling two or more robotic systems as described herein.
- the controlled robotic systems can be selected to include two or more selected from the group consisting of a macro robotic arm, a micro robotic arm and a UGV.
- the control system 101 communicates with a macro control system 51 of a macro robotic arm 54 .
- the macro robotic arm has a first set of articulating joints 66 a , 66 b , 66 c , 66 d .
- the macro control system 51 can be any system operatively associated with the macro robotic arm 54 which facilitates control of actuators 67 a , 67 b , 67 c , 67 d in response to control system 101 .
- the actuators cause various elongated portions 75 a , 75 b , 75 c of the robotic arm 54 to move relative to one another at the articulating joints.
- the macro robotic arm 54 is thus movable in two or more directions responsive to movement of the control interface grip 102 .
- the control system 101 communicates with micro control system 53 to control motion of a micro robotic arm 60 .
- the micro robotic arm 60 has a second set of articulating joints 68 a , 68 b , 68 c which are distinct from those of the macro robotic arm 54 .
- the micro control system 53 can be any system operatively associated with the micro robotic arm which controls a plurality of actuators 57 a , 57 b , 57 c for moving various elongated portions 76 a , 76 b , 76 c relative to one another at the articulating joints.
- the micro robotic arm 60 is movable in a plurality of directions (e.g.
- the control system 101 is configured to transition between the first control system state and the second control system state responsive to at least one user input to the control system.
- a push-button switch 210 disposed on the grip 102 could be used to cause such transition in some embodiments.
- control interface grip can be used to control the movement of a UGV.
- the micro robotic arm described herein can be mounted on the UGV and the same control is used in three different control system states to control each of the three robotic systems. UGVs and robot operations involving the UGV will be described in further detail as the discussion progresses.
- the controller 101 can be disposed inside of a shelter 52 , which can be a vehicle.
- the macro robotic arm 54 can be mounted to a portion of the shelter as shown in FIG. 1 . Accordingly, an operator can remain inside the shelter while controlling the movement of the macro robotic arm 54 and the micro robotic arm 60 .
- the foregoing arrangement can be advantageous for protecting an operator when the robot system is used to defeat an improvised explosive device.
- the macro and micro robotic arms 54 , 60 in FIG. 1 each includes three elongated sections.
- the invention is not limited in this regard. More or fewer elongated portions in each arm 54 , 60 are also possible.
- a variety of different types of actuators can be used for effecting such movement of the elongated portions of each arm 54 , 60 about the articulating joints. For example, electric, hydraulic, and pneumatic actuators are possible.
- the robotic arms can also have grasping devices 55 , 70 which can be used to manipulate various objects.
- Robotic grasping devices are well known in the art and therefore will not be described here in detail.
- movement of one or both of the grasping devices 55 , 70 is controlled by control system 101 .
- the control system 101 includes one or more interface controls (e.g. a trigger disposed on the grip 102 ), which permit a user to cause one or more robot fingers on the grasping devices to open or close so as to effect grasping of portions of a work piece 62 .
- rotation of the grip 102 can in some embodiments be configured to cause rotation of the grasping device when the control system is in the first or second state.
- the macro robotic arm has a macro range of movement which exceeds a micro range of movement of the micro robotic arm.
- the macro robotic arm can have an overall length of about 30 feet, whereas the micro robotic arm can have an overall length of about 3 feet. Regardless of the exact length of each arm, the macro robotic arm can have a reach or range of movement which is approximately five to twenty times larger than that of the micro robotic arm. Still, the invention is not limited in this regard and the relative difference in length of the two arms can be larger or smaller. Further, the macro robotic arm can have a macro precision of movement which is less precise than a micro precision of movement of the micro robotic arm.
- the motion of the micro robotic arm can be controlled with much greater precision as compared to the motion of the macro robotic arm.
- the greater precision associated with the micro robotic arm can be achieved using any arrangement as would be known to one skilled in the art.
- the greater precision in the micro robotic arm can be achieved by controlling the tolerances of the mechanical linkages, and through more precise control over the motion of the actuators.
- the control system 101 can communicate with the macro control system 51 and/or micro control system 53 by means of any suitable communication link.
- a wired or wireless link can be provided for effecting such communications.
- a wired connection link 64 is shown as the means for communicating with macro control system 51
- a wireless link is used to communicate with the micro control system 53 .
- the wireless link can be effected using wireless data transceivers (not shown) and suitable antennas 114 .
- the invention is not limited in this regard, and the control system 101 can communicate with the respective robotic arms using any suitable method now known or known in the future.
- the control system 101 may not communicate with control systems 51 , 53 but may instead communicate directly with the actuators associated with the particular robotic arm.
- the micro robotic arm 54 is mounted to a vehicle base 58 so as to form part of an unmanned ground vehicle (UGV) 56 .
- UGVs are well known in the art and therefore will not be described here in detail.
- a UGV is a small unmanned robotically controlled vehicle.
- the UGV generally includes a motorized base 58 powered by an onboard power source (e.g. a battery), tracks or wheels coupled to the motorized base to facilitate movement of the UGV across varied types of terrain, and a control system that is responsive to a remote control unit for controlling the movement of the motorized base.
- an onboard power source e.g. a battery
- tracks or wheels coupled to the motorized base to facilitate movement of the UGV across varied types of terrain
- a control system that is responsive to a remote control unit for controlling the movement of the motorized base.
- the invention is not limited in this regard and other UGV types are also possible.
- control system 101 and micro control system 53 can be used to control the movement of the vehicle base 58 .
- the motion of the vehicle base can be controlled responsive to a user manipulation of the control interface grip 102 in a third control system state.
- the single control system 101 interface is sufficient to direct all aspects of operating the robot arms 54 , 60 and the motorized vehicle base 58 .
- the control interface also facilitates intuitive and precise control over the robotic arm(s) in three dimensions.
- the system described herein provides significant improvements over prior art controllers that employ multiple joysticks and/or buttons to control a plurality of robotic arms.
- the invention also has advantages over conventional robotic arm control systems that require the use of two hands, and transitioning from a first set of controls, such as a joystick, to a second set of controls physically removed from the first set, such as buttons.
- the arrangement described herein offers a significant increase of convenience for the user, as the user does not need to jump from one set of controls to another. Instead, the user can simply continue to grasp a single interface control grip to direct the robotic arm(s) and vehicle base in all aspects of their operation.
- control system interface includes haptic feedback (force/tactile feedback to the operator) mechanisms that act upon the user's hand, fingers or both.
- Force sensors (not shown) present on the micro and/or macro robotic arm are used to generate signals for such haptic feedback. These signals are processed by micro control system 53 and/or macro control system 51 , and communicated to control system 101 . The control system 101 can then generate the appropriate haptic feedback force. In this manner, the interface permits the user to sense when the robotic arm 54 , 60 is striking an object, how hard the robotic arm is pushing or pulling on an object, and/o how hard the grasping device of each arm is bearing down upon an object.
- control system 101 can be configured to generate in the first control system state a first haptic force upon the interface grip 102 responsive to a detected force exerted upon the macro robotic arm 54 .
- control system 101 can be configured to generate in the second control system state a second haptic force upon the control interface grip 102 responsive to a detected force exerted upon the micro robotic arm 60 .
- an image sensor 69 , 71 is disposed on the robotic arm 54 , 60 respectively, and a video display 65 is positioned at a location near the control system 101 .
- the system is operatively arranged so that image sensor signals are communicated to the video display device using an appropriate communication link. Consequently, images detected by the image sensor can be displayed on the video display device 65 .
- the image sensors 69 , 71 are advantageously disposed near a distal end of the robotic arm 54 , 60 , proximate to a location where a robotic gripper 55 , 70 is mounted.
- the video display device and associated control circuitry can be arranged so that the image generated from image sensor 69 is automatically displayed on the video display device 65 when the control unit is in its first operational state (i.e., controlling the macro robotic arm 54 ). Conversely, the image generated from sensor 71 can be automatically displayed when the control unit is in its second operational state (i.e., controlling micro robotic arm 60 ). Further, the displayed image can be manually controlled by an operator.
- a mobility of the vehicle base 58 may be restricted by obstacles such as debris, fences or structures. Such obstacles can prevent the vehicle base 58 from moving the micro robotic arm 60 in proximity to the work piece 62 . Consequently, the micro robotic arm may not be able to reach the work piece. In such scenarios, the macro robotic arm 54 and the micro robotic arm 60 can function cooperatively to overcome the limitations of the vehicle base 58 .
- the invention concerns a method for performing robotic operations on a work piece 62 .
- the method includes communicating with the macro control system 51 to control a pose of a macro robotic arm 54 .
- the method also includes selectively varying the pose of the macro robotic arm 54 to transport a micro robotic arm 60 to a work location (e.g. a location in proximity to work piece 62 ).
- the method continues while the micro robotic arm is at such work location, by communicating with the micro control system 53 to control a pose of the micro robotic arm.
- the micro robotic arm 60 is configured so that it is supported at a distal end 80 of the macro robotic arm 54 .
- Such an arrangement is advantageous to facilitate positioning the micro robotic arm 60 in proximity to an object or work piece 62 upon which some robot action is to be performed.
- Such an arrangement is particularly advantageous when the access of the UGV to a work piece 62 is restricted by obstacles such as fences 78 and debris 77 .
- the entire UGV 56 is supported in a docking station 82 at distal end 80 of the macro robotic arm 54 .
- the docking station 82 can be a grasping device 55 or can be any other mechanical structure configured to securely engage the UGV 56 .
- the docking station can include a power and/or a communication interface for communicating power and/or control signals to the UGV.
- the UGV can operate entirely independently of the macro robotic arm while being supported in the docking station. For example, communications signals can be communicated to the UGV wirelessly as previously described in relation to FIG. 1 , and the power to the UGV can be provided by a battery pack.
- the vehicle base 58 of the UGV is omitted and micro robotic arm 60 is instead carried in the docking station 82 in a non-motorized robot arm base unit 84 .
- the non-motorized robot arm base unit 84 can include any necessary power supply, communication and control circuitry for the micro robot arm 60 that is normally provided as part of the UGV.
- the control system 101 can be used in the first operational state to control macro robotic arm 54 so that the micro robot arm 60 is moved in close proximity to a work piece 62 .
- an operating location of the micro robotic arm 60 can be moved (by controlling the macro robotic arm) so that the micro robotic arm is in reach of the work piece.
- the control system 101 can be transitioned to its second operational state in which it controls the movement of the micro robotic arm 60 .
- the foregoing arrangement has two advantages; it offers the extended reach of the macro robotic arm, and the greater precision of movement of the micro robotic arm.
- control system 101 is configured to automatically transition between the first and second control system states responsive to movement of the control interface grip.
- control system 101 can be configured to have an operating state in which the control system automatically selectively generates control signals which control the micro robotic arm 60 and/or the macro robotic arm 54 .
- Such state is sometimes referred to herein as a dual use operating state
- the dual-use operating state is primarily used to control the micro robotic arm 60 .
- control system 101 will automatically generate movement commands for the macro robotic arm 54 to cause the distal end 80 to move in the direction specified by the user for the micro robotic arm 60 .
- the control system 101 can generate movement commands for the macro robotic arm 54 that causes the distal end to move closer to the object 62 .
- This movement of the macro robotic arm 54 can stop when, for example, the user pulls the grip 102 back a predetermined distance along the Y-direction. In that case, the macro robotic arm becomes substantially closer along the Y-direction with respect to the target position, thereby affording the micro robotic arm 60 an ability to reach the target position.
- a macro robotic arm 54 cannot position the micro robotic arm 60 within reach of the work piece 62 because obstacles also prevent the shelter 52 from approaching the work piece.
- the work piece can be an improvised explosive device and the reach of the macro robotic arm can be insufficient to provide a safe stand off distance from the perspective a human operator.
- the docking station 82 can be configured to releasably secure the UGV at the distal end of the macro robotic arm 54 as shown in FIG. 3 .
- the invention can also include other methods for performing robotic operations on a work piece 62 .
- the method can involve communicating with a macro control system 51 to control a pose of a macro robotic arm 54 , and then selectively varying the pose of the macro robotic arm to move an unmanned ground vehicle (UGV) 56 from a first location to a second location. While at the second location, the UGV can be released from the docking station 82 of the macro robotic arm. Thereafter, the motorized vehicle base 58 of the UGV is operated to move the UGV from the second location to a third location proximate to the work piece 62 . Finally, while the UGV is in the third location, a pose of a micro robotic arm 60 attached to the UGV is controlled to perform a robotic task.
- UGV unmanned ground vehicle
- the macro robotic arm 54 can first be used to transport the UGV 56 over or across terrain so as to overcome any obstacles which would limit the mobility of the UGV.
- the macro robotic arm can also be used to place the UGV in relatively close proximity to the work piece 62 .
- locations accessible by the macro robotic arm 54 may not be sufficiently close to work piece 62 so as to permit access by the robotic arm 60 .
- shelter 52 may be prevented from approaching the work piece by a fence 78 or other debris 77 . Accordingly, once the macro robotic arm has been used to transport the UGV over any obstacles, the UGV can be releasably deployed from docking station 82 .
- the UGV can rely upon the vehicle base 58 to move independently of the macro robotic arm 54 .
- the vehicle base 58 can be used to transport the micro robotic arm to a location proximate the work piece.
- the UGV 56 can be returned to a location accessible by the macro robotic arm 54 .
- the vehicle base 58 can be used to return the UGV to the docking station 82 , after which the UGV can be transported across obstacles 77 , 78 and recovered.
- the movement of the macro robotic arm, vehicle base, micro robotic arm, and the grasping device are all under the control of a single control interface grip 102 associated with control system 101 .
- the control system 101 includes a robot control interface which senses human control input gestures. For example, these input control gestures can be applied to the interface through the grip 102 .
- the control system responds to these input gestures by generating signals which are used directly or indirectly to control vehicle base 58 and robotic arms 54 , 60 .
- the control interface grip 102 is coupled via a ball joint 105 to a plurality of articulating arms 107 a , 107 b , 108 a , 108 b , and 109 a , 109 b .
- Each of articulating arms is mounted at a base end 113 , 115 , 117 to an interface base 110 and coupled at an opposing end to a housing 106 .
- Each base end 113 , 115 , 117 pivots on a single defined axis (e.g. 124 , 125 ) in a hinge-like manner.
- the articulating arms 107 a - 109 b are assembled from a plurality of gimbal joints which are operable to permit translational movement of the housing 106 in three orthogonal directions (x, y and z).
- a center portion of the articulating arms 107 a - 109 b can comprise a plurality of joints 119 a , 119 b , joints 121 a , 121 b , and joints 123 a , 123 b which permit arm rotation about two transverse axes.
- joints 126 a , 126 b , 127 a , 127 b , and 128 a , 128 b also permit rotation about two transverse axes.
- the arrangement of the arms and related joints permit linear translation movement of housing 106 with respect to the x, y, and z orthogonal axes.
- the joints of the articulating arms are arranged to permit such linear translational movement, they are arranged such that the housing 106 is prevented from rotating.
- pitch, roll and yaw of the housing 106 is generally precluded by the attachment of the articulating arms 107 a , 107 b , 108 a , 108 b , and 109 a , 109 b to the housing.
- This type of mechanism is known in the art as a “Delta” mechanism, and the translation-only behavior of the housing has been shown in Clavel, R., “Conception d′un robot Sheffielder à 4 degrés deception,” Ph.D. Thesis, EPFL, Lausanne, Switzerland, 1991.
- At least one sensor of a first type is provided for measuring the linear translational movement of the housing 106 .
- Any suitable means may be employed to generate the current x, y and z movement information of the housing 106 with respect to a reference frame defined by the base 110 , including both optical and electro-mechanical sensors.
- the first type sensors are a plurality of angle sensors 116 employed to respectively determine the angular movement of each joint.
- one group of angle sensors 116 determines the respective angular positions of the base ends 113 , 115 , 117 with respect to the base 110 .
- Another group of angle sensors 116 measures the angle of the joints 119 a , 121 a , 123 a , while another group measures the angles of the joints 119 b , 121 b , 123 b .
- Knowledge of the location of each joint 113 , 115 , 117 directly coupled to the base 110 combined with knowledge of the lengths of the arms 107 a - 109 b and the angular positions of the joints 119 , 121 , 123 is used to compute the location, movement or both of the housing 106 with respect to the base 110 , using known trigonometric functions. Determining the movement, position or both of the housing 106 from the angle measurements provided by the joint angle sensors 116 is known in the art (see Clavel 1991 ). It should be understood that this is only one possible method of measuring the translational movement of the housing 106 and the invention is not limited with respect to the measurement method used.
- a pivot member is disposed in the housing 106 and is arranged to pivot with regard to three angular axes about a single pivot point.
- the pivot member is configured to roll, pitch and yaw about the pivot point. Any suitable mechanical arrangement can be used to form such a pivot member.
- the pivot member comprises a ball 105 including a ball surface which is at least partially spherical.
- the housing 106 includes a socket member for receiving the ball 105 .
- a ball and socket arranged of this kind is described in further detail with respect to FIGS. 6A and 6B .
- a pin 103 is fixed to ball 105 and extends in a generally radial direction from the ball 105 .
- the pin 103 is securely attached at one end to the grip 102 so that movement of the grip 102 is communicated to the ball 105 . It will be appreciated that grip 102 is provided and is attached to the pivot member so that a user, upon grasping the grip 102 , can cause the pivot member to rotate within the housing 106 .
- One or more second type sensors are mounted within housing 106 and can detect rotational motion of the ball 105 with regard to as many as three orthogonal axes (i.e., pitch, roll and yaw).
- the second type sensors can be any type sensor which is suitable to generate output signals which in combination allow the speed, direction or both of rotation of the ball 105 to be determined.
- Output signals from the first and second type sensors are advantageously used to generate data regarding movement of the grip 102 with regard to any one of six directions. This data is then used to generate output control signals for a robotic arm 54 , 60 or vehicle base 58 , as discussed above.
- the control system includes control circuitry which is responsive to a plurality of sensor signals produced by the first and second type sensors for generating output control signals.
- this control circuitry can be disposed within the base 110 ; in other embodiments the control unit may be disposed within shelter 52 .
- the output control signals generated by the control circuitry are configured to control one or more of the robotic arms 54 , 60 and vehicle base 58 .
- the pivot member includes housing 106 and ball 105 mounted therein so as to form a ball joint 201 .
- the housing 106 is shown partially cut away in FIG. 6B to reveal a socket 202 defined within the housing 106 that is configured to accept ball 105 .
- the socket 202 is at least partially spherical in shape.
- the ball 105 has a shape that is at least partially spherical in shape.
- the pivot member can pivot about a spherical center of the ball 105 .
- the external surface of the ball 105 is in direct mechanical contact or engagement with an inner surface 204 defined by the socket 202 .
- the ball 105 is generally conformal to the inner surface 204 of the socket 202 and preferably fits snugly within the inner surface 204 .
- the inner surface 204 forms a conjugate surface with respect to the outer surface of ball 105 .
- the housing 106 has an opening 206 , which in some embodiments is circular; although it will be appreciated that other shapes can be used. Such an arrangement can be useful when desiring to limit movement of the ball 105 to an extent defined by the opening 206 .
- the pin 103 extends through the opening 206 as shown.
- the diameter of the opening 206 is advantageously formed somewhat smaller than the diameter of the ball 105 in order to retain the ball 105 inside the socket 202 .
- the size of the opening 206 can be larger than the diameter of the ball 105
- the ball joint 201 can further include a sealing cup (not shown) fitted around the opening 206 and having an opening that is smaller than the diameter of the ball 105 .
- the sealing cup can be tapered in shape and sealingly connect with the sidewalls of the pin 103 so as to provide a protective cover that prevents debris from entering into the opening 206 .
- the housing opening 206 , the opening in the sealing cup, or both are dimensioned to allow the ball 105 to freely rotate within the limits defined by opening 206 , the sealing cup, or both. In other words the ball 105 can roll, pitch and yaw within the socket 202 .
- the inner surface 204 can include a bearing surface (not shown) disposed inside the socket 202 to facilitate movement of the ball 105 in relation to the housing 106 .
- the bearing surface in such embodiments can be detachably connected to the housing 106 or can be an integral part of the housing 106 .
- the bearing surface reduces friction between the ball 105 and the housing 106 , which decreases wear and tear of the parts and extends the useful life of the ball joint 201 .
- the bearing surface is a low friction lining disposed between the socket 202 and the ball 105 .
- other known means for reducing friction such as bearings or fluids, can also be used to reduce friction in the devices disclosed herein instead of or in addition to the low friction lining.
- Sensors 208 are provided for sensing and measuring the movement of the ball 105 relative to the housing 106 .
- the sensors 208 are arranged so that they are collectively able to sense and measure rotation of the ball 105 with respect to as many as three orthogonal axes (roll, pitch and yaw).
- the ball joint 201 includes at least two sensors 208 .
- Each sensor 208 is preferably capable of sensing ball rotation in at least two orthogonal directions. When two or more sensors 208 are used in this way, it is preferable that they are spaced apart around the circumference of the ball 105 as shown. Such angular spacing can ensure that there is no ambiguity in the sensed rotation direction of the ball 105 .
- the sensors 208 can be spaced apart by an angle of about 90° with respect to the center of the ball 105 , but of course the invention is not limited in this regard.
- Optical sensors are particularly well suited for this purpose, but any sensor type may be employed, provided that it is capable of performing as described herein.
- the first type sensors 116 and second type sensors 208 are electrically connected to control circuitry disposed, for example, within base 110 .
- control circuitry advantageously generates movement data for grip 102 by using information concerning rotation of ball 105 within the housing 106 and information regarding the movement of the articulating arms 107 b , 108 b , and 109 b . This information is then used to generate control signals for controlling one or more of the robotic arm(s) 54 , 60 , and motorized vehicle base 58 .
- a wired or wireless data link can be used for the purpose of communicating motion control command signals to the robotic arms and/or vehicle base.
- the data link can also be used to communicate haptic information from the robotic arms (as determined from corresponding force sensors on the robotic arm) to the control circuitry.
- the control system 101 can have an integrated data transceiver contained therein, including antenna 114 .
- wired connections are also possible, and may be preferred in some cases to avoid electromagnetic interference disrupting operations of the robotic arm(s).
- control system 101 can include an embodiment control unit 304 .
- Control unit 304 can comprise the control circuitry used in the present invention. User input signals generated by movement of the grip 102 can be transmitted, such as by wires, to the control unit 304 .
- the control unit 304 can be disposed within base 110 , or at any suitable location in the shelter 52 .
- the control unit 304 comprises a system interface 424 , a Central Processing Unit (CPU) 414 , a system bus 412 , a memory 416 connected to and accessible by other portions of the control unit 304 through system bus 412 , and hardware entities 418 connected to system bus 412 .
- CPU Central Processing Unit
- System interface 424 allows the control unit 304 to receive data from first and second type sensors 116 , 208 and from thumb buttons 118 and trigger 199 .
- the system interface 424 can also be configured to allow control unit 304 to communicate with data transceiver 306 , network equipment and other computing devices via a wired, wireless or both communications link or links.
- At least some of the hardware entities 418 perform actions involving access to and use of memory 416 , which may be a random access memory (RAM), a disk drive, and/or a compact disc read only memory (CD-ROM).
- RAM random access memory
- CD-ROM compact disc read only memory
- Hardware entities 418 can include microprocessors, application specific integrated circuits (ASICs) and other hardware. Hardware entities 418 may include a microprocessor programmed for facilitating the provision of data communication services and robot arm control processing as described herein. In this regard, it should be understood that the microprocessor can access and run robotic arm control processing as described herein.
- the communication operations can include, but are not limited to, signal receiving operations, signal processing operations, signal generation operations, and signal communication operations. Such communication operations can facilitate communications with the macro robotic arm 54 , micro robotic arm 60 and vehicle base 58 to control their respective operation.
- the hardware entities 418 can include a disk drive unit 426 or other non-volatile memory (e.g., flash memory or ROM) comprising a computer-readable storage medium 428 on which is stored one or more sets of instructions 420 (e.g., software code).
- the instructions 420 can be configured to implement one or more of the methodologies, procedures, or functions described herein, such as generation of the output control signals for motion control of the robotic arms and/or motorized vehicle base 58 , based upon inputs obtained from the control system 101 .
- the instructions are also configured for sending haptic control signals to the control system 101 based upon data obtained from the robotic arm.
- the instructions 420 can also reside, completely or at least partially, within the memory 416 and/or within the CPU 414 during execution thereof.
- the memory 416 and the CPU 414 also can constitute machine-readable media.
- the term “machine-readable media”, as used here, also refers to any medium that is capable of storing, encoding or carrying a set of instructions 420 for execution that cause the control unit 304 to perform any one or more of the methodologies of the present disclosure.
- the control unit 304 implements one or more method embodiments of the present invention, and in particular enables a user to control movement of the robotic arms 54 , 60 , and/or the movement of a motorized vehicle base 58 .
- the control unit 304 can communicate with the robotic arms 54 , 60 and/or motorized vehicle base 58 by using wired or wireless connections.
- a suitable data transceiver can be used for this purpose.
- a wireless connection is preferred, a wireless data transceiver 306 can be provided.
- the data transceiver 306 may include more or less components than those shown in FIG. 7 . However, the components shown are sufficient to disclose an illustrative embodiment implementing the present invention.
- the data transceiver 306 can comprise an antenna 402 (shown as element 114 in FIG. 1 ) for receiving and transmitting Radio Frequency (RF) signals.
- a receive/transmit (Rx/Tx) switch 404 selectively couples the antenna 402 to the transmitter circuitry 406 and receiver circuitry 408 in a manner familiar to those skilled in the art.
- the receiver circuitry 408 decodes the RF signals received from the robotic arm 60 to derive information therefrom; this includes haptic information that can be processed to provide haptic feedback to the user.
- the receiver circuitry 408 is coupled to control unit 304 via an electrical connection 434 .
- the receiver circuitry 408 provides decoded RF signal information to the control unit 304 .
- the control unit 304 uses the decoded RF signal information in accordance with the function(s) of the invention as described herein.
- a wired connection is made with one or more of the robotic arms and/or motorized vehicle base, as known in the art.
- a wired connection may be more suitable for communications associated with control and haptic feedback from macro robotic arm 54 .
- Such a wired connection would be functionally equivalent to a wireless connection shown in FIG. 3 but is easier to implement with respect to the communications circuitry 404 , 406 , 408 .
- the macro robotic arm 54 and micro robotic arm 60 are responsive to control signals for moving a grasping device 55 , 70 in a plurality of directions.
- the robotic arms 54 , 60 can be capable of moving the respective grasping devices in as many as three orthogonal linear directions, and in a plurality of rotational directions defined by as many as three orthogonal axes of rotation.
- Robotic arms as described herein are known in the art and therefore will not be described here in detail.
- each of the macro and micro robotic arms 54 , 60 can include a plurality of articulated joints 66 a - 66 d and 68 a - 68 c , each with a corresponding motor or driving device that sets the angular position of that joint as instructed by the control system 101 . More specifically, by transmitting appropriate output control signals the control system can direct the controls actuators to control the spatial positioning of the tip of a grasping device 55 , 70 .
- the position of the grasping device 55 , 70 is preferably in accordance with the spatial positioning of the housing 106 .
- the angular positioning of the robotic grasping devices 55 , 70 is preferably controlled in accordance with the angular positioning of the grip 102 .
- Power used to activate the articulated joints 68 a - 68 c can be provided, for example, from the battery packs (not shown) associated with the robotic arm 60 .
- Power used to activate the articulated joints 67 a - 67 d can be provided by a power source associated with shelter 52 .
- an onboard hydraulic pump (not shown) included with shelter 52 can be used for this purpose.
- the control unit 304 can obtain power from a battery or alternator associated with the shelter 52 .
- One or both of the macro robotic arm and micro robotic arm can include a microprocessor based control unit and a data transceiver configured to respond to commands from the control system 101 .
- a microprocessor based control unit configured to respond to commands from the control system 101 .
- the UGV can have an onboard wireless data transceiver similar to data transceiver 306 , and appropriate control circuitry for responding to control signals.
- One or both of the grasping devices 55 , 70 can be a force-sensing grasping device, such as the one disclosed in U.S. application Ser. No. 13/049,982, entitled “Robotic Grasping Device With Multi-Force Sensing At Base Of Fingers,” filed on Mar. 17, 2011, the contents of which are incorporated herein by reference.
- the grasping device can include fingers, the angular positions of which can be controlled to open or close relative to each other to grasp and release objects.
- the grasping devices are arranged to send haptic signals back to the control unit 304 , as measured by, for example, the fingers.
- the haptic signals are used by the control system to control motive devices in the control system 101 to create a corresponding haptic feedback response for the user.
- This haptic feedback response can indicate, for example, the magnitude and direction of a force present on the grasping device 55 and/or 70 and/or the forces upon one or both of the robotic arms 54 , 60 .
- This force can be felt by the user in the form of a corresponding resistance to motion of the grip 102 in the direction of this force, the resistance being proportional to the magnitude of the force impinging on the robotic arm.
- a user can sense, for example, the weight of an object, how hard the grasping device is pressing against a surface, and so forth.
- a haptic signal in the form of resistance to depression of the trigger 199 can be generated that is proportional to how hard the fingers are bearing down upon an object being gripped. In this manner a user can determine how firmly an object is being grasped.
- One or more actuators, motors or servo motors can be disposed in relation to the control system 101 to provide these haptic feedback signals via the grip 102 and trigger 199 .
- actuators or the like in the controller base 110 can apply a force to the articulating arms 107 b , 108 b , and 109 b , which is then experienced in the grip 102 .
- actuators or motors can be disposed in motor housings 112 which are provided as part of base 110 .
- the force generated by the motors is used to provide haptic feedback to the user.
- Additional actuators for providing haptic feedback to a user can be provided within grip 102 itself. For example, such an actuator can exert a force upon the trigger 199 .
Landscapes
- Engineering & Computer Science (AREA)
- Robotics (AREA)
- Mechanical Engineering (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Mathematical Physics (AREA)
- Aviation & Aerospace Engineering (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- General Physics & Mathematics (AREA)
- Automation & Control Theory (AREA)
- Computing Systems (AREA)
- Manipulator (AREA)
Abstract
A robot system (50) includes a control system (101) having a control interface grip (102). The robot system includes a macro robotic arm (54) and a micro robotic arm (60). The robot system is arranged such that the macro robotic arm will respond, in a first control system state, to movement of the control interface grip. In particular, the macro robotic arm will move in a plurality of directions responsive to corresponding movement of the interface grip. The micro robotic arm will respond, in a second control system state, to movement of the control interface grip. In particular, the micro robotic arm will move in a plurality of directions responsive to corresponding movement of the interface grip.
Description
This application is a divisional application of and claims priority to co-pending non-provisional U.S. patent application Ser. No. 13/267,357 filed on Oct. 6, 2011, which is hereby incorporated by reference in its entirety.
Statement of the Technical Field
The invention concerns explosive devices, and more particularly systems and methods for defeating improvised explosive devices (IEDs) using robotic equipment.
Description of the Related Art
Unmanned ground vehicles (UGVs) are commonly used for defeating explosive devices, and more specifically, IED devices. Such UGVs commonly include a remotely operated battery powered vehicle which is capable of moving under its own power. One problem with UGVs is that they frequently have trouble navigating in cluttered environments. In such environments, the UGV can experience a loss of mobility when obstacles block its path or otherwise inhibit its movement. When UGVs are used for ordinance disposal, they usually include a robotic arm and a video camera which allow a user to disable an IED from a safe distance. Still, conventional UGVs tend to have relatively primitive control systems for controlling movement of the distal end of the robotic arm, where a gripping device is typically located. Such control systems commonly allow the robotic arm to be moved with four or five degrees of freedom, but often require the operator to separately control each joint. Consequently, these robot arm control systems are not very intuitive, and contribute to making the robot arm difficult to control precisely.
Another type of vehicle which has been developed to address the IED problem is the mine resistant ambush protected (MRAP) family of vehicle. The MRAP is a heavily armored fighting vehicle which is designed to transport soldiers, while also protecting them from IED attacks. In addition to providing protection, certain types of MRAPs are fitted with systems used to assist with ordinance disposal. For example, the Buffalo, which is manufactured by Force Protection, Inc., is a six wheeled vehicle that has a large 30 foot long robotic arm mounted to the frame of the vehicle. The arm, which is similar to the articulated arm of a backhoe, can be used to allow personnel to dispose of IEDs while remaining enclosed in the relative safety of the vehicle. The robot arms used on the MRAPs suffer from many of the same problems as the robot arm of the UGV, and are even more difficult to control with precision.
The invention concerns a robot system, including a control system. The control system includes a robot control interface having a control interface grip. The robot system includes at least two robotic arms, namely a macro robotic arm and a micro robotic arm. The macro robotic arm has a first set of articulating joints. The robot system is arranged such that the macro robotic arm will respond, in a first control system state, to movement of the control interface grip. In particular, the macro robotic arm will move in a plurality of directions responsive to corresponding movement of the interface grip. The micro robotic arm has a second set of articulating joints distinct from the first set. The robot system is arranged so that the micro robotic arm will respond, in a second control system state, to movement of the control interface grip. In particular, the micro robotic arm will move in a plurality of directions responsive to corresponding movement of the interface grip.
The invention also concerns a method for controlling a robot system. The method includes controlling a movement of a macro robotic arm having a first set of articulating joints in a first system state. In such first system state, the macro robotic arm is responsive to a user manipulation of a control interface grip. In particular, the macro robotic arm will move in a plurality of directions in response to corresponding movement of the control interface grip. The method further involves controlling in a second system state a movement of a micro robotic arm having a second set of articulating joints distinct from the first set. In such second system state, the micro robotic arm will move in a plurality of directions in response to corresponding movement of the control interface grip. In a third system state, the method includes controlling a movement of a vehicle base to which the micro robotic arm is attached responsive to a user manipulation of the control interface grip.
According to another aspect, the invention concerns a method for performing robotic operations on a work piece. The method includes communicating with a macro control system to control a pose of a macro robotic arm. The method also includes selectively varying the pose of the macro robotic arm to transport a micro robotic arm to a work location. The method continues while the micro robotic arm is at such work location, by communicating with a micro control system separate from said macro control system to control a pose of said micro robotic arm.
According to another aspect, the invention includes performing robotic operations on a work piece. The method involves communicating with a macro control system to control a pose of a macro robotic arm, and then selectively varying the pose of said macro robotic arm to move an unmanned ground vehicle (UGV) from a first location to a second location. While at the second location, the UGV is released from a docking station of the macro robotic arm. Thereafter, the method continues with operating a motorized vehicle base of the UGV to move the UGV from the second location to a third location. Finally, while the UGV is in the third location, a pose of a micro robotic arm attached to said UGV is controlled to perform a robotic task. The method can include controlling UGV movement using the same interface as used to control the macro and micro robot arms.
Embodiments will be described with reference to the following drawing figures, in which like numerals represent like items throughout the figures, and in which:
The present invention is described with reference to the attached figures. The figures are not drawn to scale and they are provided merely to illustrate the instant invention. Several aspects of the invention are described below with reference to example applications for illustration. It should be understood that numerous specific details, relationships, and methods are set forth to provide a full understanding of the invention. One having ordinary skill in the relevant art, however, will readily recognize that the invention can be practiced without one or more of the specific details or with other methods. In other instances, well-known structures or operation are not shown in detail to avoid obscuring the invention. The present invention is not limited by the illustrated ordering of acts or events, as some acts may occur in different orders and/or concurrently with other acts or events. Furthermore, not all illustrated acts or events are required to implement a methodology in accordance with the present invention.
The invention concerns a robot system 50 and a method of using same. The robot system includes a control system 101. The control system including a control interface having a control interface grip 102. The control interface grip 102 and associated control system 101 is arranged for controlling two or more robotic systems as described herein. The controlled robotic systems can be selected to include two or more selected from the group consisting of a macro robotic arm, a micro robotic arm and a UGV. For example, in a first control system state, the control system 101 communicates with a macro control system 51 of a macro robotic arm 54. The macro robotic arm has a first set of articulating joints 66 a, 66 b, 66 c, 66 d. The macro control system 51 can be any system operatively associated with the macro robotic arm 54 which facilitates control of actuators 67 a, 67 b, 67 c, 67 d in response to control system 101. The actuators cause various elongated portions 75 a, 75 b, 75 c of the robotic arm 54 to move relative to one another at the articulating joints. The macro robotic arm 54 is thus movable in two or more directions responsive to movement of the control interface grip 102.
In a second system state, the control system 101 communicates with micro control system 53 to control motion of a micro robotic arm 60. The micro robotic arm 60 has a second set of articulating joints 68 a, 68 b, 68 c which are distinct from those of the macro robotic arm 54. The micro control system 53 can be any system operatively associated with the micro robotic arm which controls a plurality of actuators 57 a, 57 b, 57 c for moving various elongated portions 76 a, 76 b, 76 c relative to one another at the articulating joints. As such, the micro robotic arm 60 is movable in a plurality of directions (e.g. orthogonal directions) responsive to movement of the control interface grip 102 in corresponding directions. The control system 101 is configured to transition between the first control system state and the second control system state responsive to at least one user input to the control system. For example, a push-button switch 210 disposed on the grip 102 could be used to cause such transition in some embodiments.
In a third system state, the control interface grip can be used to control the movement of a UGV. In some embodiments the micro robotic arm described herein can be mounted on the UGV and the same control is used in three different control system states to control each of the three robotic systems. UGVs and robot operations involving the UGV will be described in further detail as the discussion progresses.
The controller 101 can be disposed inside of a shelter 52, which can be a vehicle. In a preferred embodiment, the macro robotic arm 54 can be mounted to a portion of the shelter as shown in FIG. 1 . Accordingly, an operator can remain inside the shelter while controlling the movement of the macro robotic arm 54 and the micro robotic arm 60. The foregoing arrangement can be advantageous for protecting an operator when the robot system is used to defeat an improvised explosive device.
Robotic arms are well known in the art and therefore will not be described here in detail In general, the macro and micro robotic arms described herein can be implemented using any suitable arrangement as would be known to one skilled in the art now, or in the future. In one embodiment, the macro and micro robotic arms 54, 60 in FIG. 1 each includes three elongated sections. However, the invention is not limited in this regard. More or fewer elongated portions in each arm 54, 60 are also possible. A variety of different types of actuators can be used for effecting such movement of the elongated portions of each arm 54, 60 about the articulating joints. For example, electric, hydraulic, and pneumatic actuators are possible.
The robotic arms can also have grasping devices 55, 70 which can be used to manipulate various objects. Robotic grasping devices are well known in the art and therefore will not be described here in detail. In some embodiments, movement of one or both of the grasping devices 55, 70 is controlled by control system 101. For example, the control system 101 includes one or more interface controls (e.g. a trigger disposed on the grip 102), which permit a user to cause one or more robot fingers on the grasping devices to open or close so as to effect grasping of portions of a work piece 62. Similarly, rotation of the grip 102 can in some embodiments be configured to cause rotation of the grasping device when the control system is in the first or second state.
According to a preferred embodiment, the macro robotic arm has a macro range of movement which exceeds a micro range of movement of the micro robotic arm. In one example, the macro robotic arm can have an overall length of about 30 feet, whereas the micro robotic arm can have an overall length of about 3 feet. Regardless of the exact length of each arm, the macro robotic arm can have a reach or range of movement which is approximately five to twenty times larger than that of the micro robotic arm. Still, the invention is not limited in this regard and the relative difference in length of the two arms can be larger or smaller. Further, the macro robotic arm can have a macro precision of movement which is less precise than a micro precision of movement of the micro robotic arm. Stated differently, the motion of the micro robotic arm can be controlled with much greater precision as compared to the motion of the macro robotic arm. The greater precision associated with the micro robotic arm can be achieved using any arrangement as would be known to one skilled in the art. For example, the greater precision in the micro robotic arm can be achieved by controlling the tolerances of the mechanical linkages, and through more precise control over the motion of the actuators.
The control system 101 can communicate with the macro control system 51 and/or micro control system 53 by means of any suitable communication link. For example, a wired or wireless link can be provided for effecting such communications. In FIG. 1 , a wired connection link 64 is shown as the means for communicating with macro control system 51, and a wireless link is used to communicate with the micro control system 53. The wireless link can be effected using wireless data transceivers (not shown) and suitable antennas 114. Still, the invention is not limited in this regard, and the control system 101 can communicate with the respective robotic arms using any suitable method now known or known in the future. In some embodiments, the control system 101 may not communicate with control systems 51, 53 but may instead communicate directly with the actuators associated with the particular robotic arm.
In an embodiment of the invention, the micro robotic arm 54 is mounted to a vehicle base 58 so as to form part of an unmanned ground vehicle (UGV) 56. UGVs are well known in the art and therefore will not be described here in detail. However, it should be appreciated that a UGV is a small unmanned robotically controlled vehicle. As such, the UGV generally includes a motorized base 58 powered by an onboard power source (e.g. a battery), tracks or wheels coupled to the motorized base to facilitate movement of the UGV across varied types of terrain, and a control system that is responsive to a remote control unit for controlling the movement of the motorized base. Still, the invention is not limited in this regard and other UGV types are also possible. All such UGV systems known now or in the future are intended to be included within the scope of the present invention. In an embodiment of the invention, the control system 101 and micro control system 53 can be used to control the movement of the vehicle base 58. For example, the motion of the vehicle base can be controlled responsive to a user manipulation of the control interface grip 102 in a third control system state.
In a preferred embodiment, the single control system 101 interface is sufficient to direct all aspects of operating the robot arms 54, 60 and the motorized vehicle base 58. The control interface also facilitates intuitive and precise control over the robotic arm(s) in three dimensions. As such, the system described herein provides significant improvements over prior art controllers that employ multiple joysticks and/or buttons to control a plurality of robotic arms. The invention also has advantages over conventional robotic arm control systems that require the use of two hands, and transitioning from a first set of controls, such as a joystick, to a second set of controls physically removed from the first set, such as buttons. The arrangement described herein offers a significant increase of convenience for the user, as the user does not need to jump from one set of controls to another. Instead, the user can simply continue to grasp a single interface control grip to direct the robotic arm(s) and vehicle base in all aspects of their operation.
In preferred embodiments the control system interface includes haptic feedback (force/tactile feedback to the operator) mechanisms that act upon the user's hand, fingers or both. Force sensors (not shown) present on the micro and/or macro robotic arm are used to generate signals for such haptic feedback. These signals are processed by micro control system 53 and/or macro control system 51, and communicated to control system 101. The control system 101 can then generate the appropriate haptic feedback force. In this manner, the interface permits the user to sense when the robotic arm 54, 60 is striking an object, how hard the robotic arm is pushing or pulling on an object, and/o how hard the grasping device of each arm is bearing down upon an object. For example, the control system 101 can be configured to generate in the first control system state a first haptic force upon the interface grip 102 responsive to a detected force exerted upon the macro robotic arm 54. Similarly, the control system 101 can be configured to generate in the second control system state a second haptic force upon the control interface grip 102 responsive to a detected force exerted upon the micro robotic arm 60.
In an embodiment of the invention, an image sensor 69, 71 is disposed on the robotic arm 54, 60 respectively, and a video display 65 is positioned at a location near the control system 101. The system is operatively arranged so that image sensor signals are communicated to the video display device using an appropriate communication link. Consequently, images detected by the image sensor can be displayed on the video display device 65. The image sensors 69, 71 are advantageously disposed near a distal end of the robotic arm 54, 60, proximate to a location where a robotic gripper 55, 70 is mounted. The video display device and associated control circuitry can be arranged so that the image generated from image sensor 69 is automatically displayed on the video display device 65 when the control unit is in its first operational state (i.e., controlling the macro robotic arm 54). Conversely, the image generated from sensor 71 can be automatically displayed when the control unit is in its second operational state (i.e., controlling micro robotic arm 60). Further, the displayed image can be manually controlled by an operator.
Referring once again to FIG. 1 , there may be certain situations in which a mobility of the vehicle base 58 may be restricted by obstacles such as debris, fences or structures. Such obstacles can prevent the vehicle base 58 from moving the micro robotic arm 60 in proximity to the work piece 62. Consequently, the micro robotic arm may not be able to reach the work piece. In such scenarios, the macro robotic arm 54 and the micro robotic arm 60 can function cooperatively to overcome the limitations of the vehicle base 58.
According to another aspect of the invention, the invention concerns a method for performing robotic operations on a work piece 62. The method includes communicating with the macro control system 51 to control a pose of a macro robotic arm 54. The method also includes selectively varying the pose of the macro robotic arm 54 to transport a micro robotic arm 60 to a work location (e.g. a location in proximity to work piece 62). The method continues while the micro robotic arm is at such work location, by communicating with the micro control system 53 to control a pose of the micro robotic arm. This aspect of the invention will now be described in further detail.
As shown in FIGS. 2 and 4 , the micro robotic arm 60 is configured so that it is supported at a distal end 80 of the macro robotic arm 54. Such an arrangement is advantageous to facilitate positioning the micro robotic arm 60 in proximity to an object or work piece 62 upon which some robot action is to be performed. Such an arrangement is particularly advantageous when the access of the UGV to a work piece 62 is restricted by obstacles such as fences 78 and debris 77.
In an embodiment of the invention shown in FIG. 2 , the entire UGV 56, including the vehicle base 58, is supported in a docking station 82 at distal end 80 of the macro robotic arm 54. The docking station 82 can be a grasping device 55 or can be any other mechanical structure configured to securely engage the UGV 56. In some embodiments, the docking station can include a power and/or a communication interface for communicating power and/or control signals to the UGV. Still, the UGV can operate entirely independently of the macro robotic arm while being supported in the docking station. For example, communications signals can be communicated to the UGV wirelessly as previously described in relation to FIG. 1 , and the power to the UGV can be provided by a battery pack.
In the embodiment shown in FIG. 4 , the vehicle base 58 of the UGV is omitted and micro robotic arm 60 is instead carried in the docking station 82 in a non-motorized robot arm base unit 84. In such an embodiment, the non-motorized robot arm base unit 84 can include any necessary power supply, communication and control circuitry for the micro robot arm 60 that is normally provided as part of the UGV.
With the arrangement shown in FIGS. 2 and 4 , the control system 101 can be used in the first operational state to control macro robotic arm 54 so that the micro robot arm 60 is moved in close proximity to a work piece 62. For example, an operating location of the micro robotic arm 60 can be moved (by controlling the macro robotic arm) so that the micro robotic arm is in reach of the work piece. Once the micro robotic arm is positioned in this way, the control system 101 can be transitioned to its second operational state in which it controls the movement of the micro robotic arm 60. The foregoing arrangement has two advantages; it offers the extended reach of the macro robotic arm, and the greater precision of movement of the micro robotic arm.
In an alternative embodiment of the invention shown in FIGS. 2 and 4 , the control system 101 is configured to automatically transition between the first and second control system states responsive to movement of the control interface grip. With such an arrangement, the control system 101 can be configured to have an operating state in which the control system automatically selectively generates control signals which control the micro robotic arm 60 and/or the macro robotic arm 54. Such state is sometimes referred to herein as a dual use operating state The dual-use operating state is primarily used to control the micro robotic arm 60. In this mode, if the user attempts to move or extend the micro robotic arm 60 to a location that is beyond the maximum reach of the robotic arm 60, then control system 101 will automatically generate movement commands for the macro robotic arm 54 to cause the distal end 80 to move in the direction specified by the user for the micro robotic arm 60.
By way of example, consider the case where the user is pushing the interface grip 102 forward along the direction of the Y-axis to cause the grasping device 70 of micro robotic arm 60 to reach a target position on object 62. Assume also that sensors indicate that the micro robotic arm 60 is incapable of further forward movement towards the target position. In other words, assume that the target position is beyond the maximum extended reach of the micro robotic arm 60. In this scenario, the control system 101 can generate movement commands for the macro robotic arm 54 that causes the distal end to move closer to the object 62. This movement of the macro robotic arm 54 can stop when, for example, the user pulls the grip 102 back a predetermined distance along the Y-direction. In that case, the macro robotic arm becomes substantially closer along the Y-direction with respect to the target position, thereby affording the micro robotic arm 60 an ability to reach the target position.
There may be certain situations in which obstacles prevent a UGV 56 from approaching a work piece 62 when relying exclusively upon the mobility of a vehicle base 58, and the problem cannot be overcome by using the greater reach of the macro robotic arm. For example, in some scenarios a macro robotic arm 54 cannot position the micro robotic arm 60 within reach of the work piece 62 because obstacles also prevent the shelter 52 from approaching the work piece. In other scenarios, the work piece can be an improvised explosive device and the reach of the macro robotic arm can be insufficient to provide a safe stand off distance from the perspective a human operator. In such scenarios, the docking station 82 can be configured to releasably secure the UGV at the distal end of the macro robotic arm 54 as shown in FIG. 3 .
In such an embodiment, the invention can also include other methods for performing robotic operations on a work piece 62. For example, the method can involve communicating with a macro control system 51 to control a pose of a macro robotic arm 54, and then selectively varying the pose of the macro robotic arm to move an unmanned ground vehicle (UGV) 56 from a first location to a second location. While at the second location, the UGV can be released from the docking station 82 of the macro robotic arm. Thereafter, the motorized vehicle base 58 of the UGV is operated to move the UGV from the second location to a third location proximate to the work piece 62. Finally, while the UGV is in the third location, a pose of a micro robotic arm 60 attached to the UGV is controlled to perform a robotic task.
In the arrangement shown in FIG. 3 , the macro robotic arm 54 can first be used to transport the UGV 56 over or across terrain so as to overcome any obstacles which would limit the mobility of the UGV. The macro robotic arm can also be used to place the UGV in relatively close proximity to the work piece 62. Still, locations accessible by the macro robotic arm 54 may not be sufficiently close to work piece 62 so as to permit access by the robotic arm 60. For example, shelter 52 may be prevented from approaching the work piece by a fence 78 or other debris 77. Accordingly, once the macro robotic arm has been used to transport the UGV over any obstacles, the UGV can be releasably deployed from docking station 82. At this point, the UGV can rely upon the vehicle base 58 to move independently of the macro robotic arm 54. For example, the vehicle base 58 can be used to transport the micro robotic arm to a location proximate the work piece. When such activity is completed, the UGV 56 can be returned to a location accessible by the macro robotic arm 54. Specifically, the vehicle base 58 can be used to return the UGV to the docking station 82, after which the UGV can be transported across obstacles 77, 78 and recovered. In a preferred embodiment, the movement of the macro robotic arm, vehicle base, micro robotic arm, and the grasping device are all under the control of a single control interface grip 102 associated with control system 101.
An example of control system 101 will now be described in further detail with reference to FIG. 5 . The control system 101 includes a robot control interface which senses human control input gestures. For example, these input control gestures can be applied to the interface through the grip 102. The control system responds to these input gestures by generating signals which are used directly or indirectly to control vehicle base 58 and robotic arms 54, 60. In an embodiment of the invention, the control interface grip 102 is coupled via a ball joint 105 to a plurality of articulating arms 107 a, 107 b, 108 a, 108 b, and 109 a, 109 b. Each of articulating arms is mounted at a base end 113, 115, 117 to an interface base 110 and coupled at an opposing end to a housing 106. Each base end 113, 115, 117 pivots on a single defined axis (e.g. 124, 125) in a hinge-like manner. The articulating arms 107 a-109 b are assembled from a plurality of gimbal joints which are operable to permit translational movement of the housing 106 in three orthogonal directions (x, y and z). More particularly, a center portion of the articulating arms 107 a-109 b can comprise a plurality of joints 119 a, 119 b, joints 121 a, 121 b, and joints 123 a, 123 b which permit arm rotation about two transverse axes. At a housing end of the arms joints 126 a, 126 b, 127 a, 127 b, and 128 a, 128 b also permit rotation about two transverse axes. The arrangement of the arms and related joints permit linear translation movement of housing 106 with respect to the x, y, and z orthogonal axes.
In preferred embodiments, while the joints of the articulating arms are arranged to permit such linear translational movement, they are arranged such that the housing 106 is prevented from rotating. In other words, pitch, roll and yaw of the housing 106 is generally precluded by the attachment of the articulating arms 107 a, 107 b, 108 a, 108 b, and 109 a, 109 b to the housing. This type of mechanism is known in the art as a “Delta” mechanism, and the translation-only behavior of the housing has been shown in Clavel, R., “Conception d′un robot parallèle rapide à 4 degrés de liberté,” Ph.D. Thesis, EPFL, Lausanne, Switzerland, 1991.
At least one sensor of a first type is provided for measuring the linear translational movement of the housing 106. Any suitable means may be employed to generate the current x, y and z movement information of the housing 106 with respect to a reference frame defined by the base 110, including both optical and electro-mechanical sensors. In one embodiment the first type sensors are a plurality of angle sensors 116 employed to respectively determine the angular movement of each joint. For example, one group of angle sensors 116 determines the respective angular positions of the base ends 113, 115, 117 with respect to the base 110. Another group of angle sensors 116 measures the angle of the joints 119 a, 121 a, 123 a, while another group measures the angles of the joints 119 b, 121 b, 123 b. Knowledge of the location of each joint 113, 115, 117 directly coupled to the base 110 combined with knowledge of the lengths of the arms 107 a-109 b and the angular positions of the joints 119, 121, 123 is used to compute the location, movement or both of the housing 106 with respect to the base 110, using known trigonometric functions. Determining the movement, position or both of the housing 106 from the angle measurements provided by the joint angle sensors 116 is known in the art (see Clavel 1991). It should be understood that this is only one possible method of measuring the translational movement of the housing 106 and the invention is not limited with respect to the measurement method used.
A pivot member is disposed in the housing 106 and is arranged to pivot with regard to three angular axes about a single pivot point. In other words, the pivot member is configured to roll, pitch and yaw about the pivot point. Any suitable mechanical arrangement can be used to form such a pivot member. However, in a preferred embodiment, the pivot member comprises a ball 105 including a ball surface which is at least partially spherical. The housing 106 includes a socket member for receiving the ball 105. A ball and socket arranged of this kind is described in further detail with respect to FIGS. 6A and 6B . A pin 103 is fixed to ball 105 and extends in a generally radial direction from the ball 105. The pin 103 is securely attached at one end to the grip 102 so that movement of the grip 102 is communicated to the ball 105. It will be appreciated that grip 102 is provided and is attached to the pivot member so that a user, upon grasping the grip 102, can cause the pivot member to rotate within the housing 106.
One or more second type sensors (not shown in FIG. 5 ) are mounted within housing 106 and can detect rotational motion of the ball 105 with regard to as many as three orthogonal axes (i.e., pitch, roll and yaw). The second type sensors can be any type sensor which is suitable to generate output signals which in combination allow the speed, direction or both of rotation of the ball 105 to be determined.
Output signals from the first and second type sensors are advantageously used to generate data regarding movement of the grip 102 with regard to any one of six directions. This data is then used to generate output control signals for a robotic arm 54, 60 or vehicle base 58, as discussed above. More particularly, the control system includes control circuitry which is responsive to a plurality of sensor signals produced by the first and second type sensors for generating output control signals. In some embodiments, this control circuitry can be disposed within the base 110; in other embodiments the control unit may be disposed within shelter 52. The output control signals generated by the control circuitry are configured to control one or more of the robotic arms 54, 60 and vehicle base 58.
Referring now to FIGS. 6A and 6B , one embodiment of a pivot member and related structure is shown in greater detail. The pivot member includes housing 106 and ball 105 mounted therein so as to form a ball joint 201. The housing 106 is shown partially cut away in FIG. 6B to reveal a socket 202 defined within the housing 106 that is configured to accept ball 105. The socket 202 is at least partially spherical in shape. In preferred embodiments the ball 105 has a shape that is at least partially spherical in shape. The pivot member can pivot about a spherical center of the ball 105. In some embodiments, the external surface of the ball 105 is in direct mechanical contact or engagement with an inner surface 204 defined by the socket 202. As can be observed in FIG. 6B , the ball 105 is generally conformal to the inner surface 204 of the socket 202 and preferably fits snugly within the inner surface 204. According to some embodiments, the inner surface 204 forms a conjugate surface with respect to the outer surface of ball 105.
The housing 106 has an opening 206, which in some embodiments is circular; although it will be appreciated that other shapes can be used. Such an arrangement can be useful when desiring to limit movement of the ball 105 to an extent defined by the opening 206. The pin 103 extends through the opening 206 as shown. In some embodiments, the diameter of the opening 206 is advantageously formed somewhat smaller than the diameter of the ball 105 in order to retain the ball 105 inside the socket 202. In other embodiments, the size of the opening 206 can be larger than the diameter of the ball 105, and the ball joint 201 can further include a sealing cup (not shown) fitted around the opening 206 and having an opening that is smaller than the diameter of the ball 105. In some embodiments the sealing cup can be tapered in shape and sealingly connect with the sidewalls of the pin 103 so as to provide a protective cover that prevents debris from entering into the opening 206. The housing opening 206, the opening in the sealing cup, or both are dimensioned to allow the ball 105 to freely rotate within the limits defined by opening 206, the sealing cup, or both. In other words the ball 105 can roll, pitch and yaw within the socket 202.
The inner surface 204 can include a bearing surface (not shown) disposed inside the socket 202 to facilitate movement of the ball 105 in relation to the housing 106. The bearing surface in such embodiments can be detachably connected to the housing 106 or can be an integral part of the housing 106. The bearing surface reduces friction between the ball 105 and the housing 106, which decreases wear and tear of the parts and extends the useful life of the ball joint 201. Preferably, the bearing surface is a low friction lining disposed between the socket 202 and the ball 105. However, other known means for reducing friction, such as bearings or fluids, can also be used to reduce friction in the devices disclosed herein instead of or in addition to the low friction lining.
The first type sensors 116 and second type sensors 208, are electrically connected to control circuitry disposed, for example, within base 110. Such control circuitry advantageously generates movement data for grip 102 by using information concerning rotation of ball 105 within the housing 106 and information regarding the movement of the articulating arms 107 b, 108 b, and 109 b. This information is then used to generate control signals for controlling one or more of the robotic arm(s) 54, 60, and motorized vehicle base 58. A wired or wireless data link can be used for the purpose of communicating motion control command signals to the robotic arms and/or vehicle base. The data link can also be used to communicate haptic information from the robotic arms (as determined from corresponding force sensors on the robotic arm) to the control circuitry. In some embodiments, the control system 101 can have an integrated data transceiver contained therein, including antenna 114. As noted above, wired connections are also possible, and may be preferred in some cases to avoid electromagnetic interference disrupting operations of the robotic arm(s).
Referring now to FIG. 7 , the control system 101 can include an embodiment control unit 304. Control unit 304 can comprise the control circuitry used in the present invention. User input signals generated by movement of the grip 102 can be transmitted, such as by wires, to the control unit 304. The control unit 304 can be disposed within base 110, or at any suitable location in the shelter 52. As shown in FIG. 7 , the control unit 304 comprises a system interface 424, a Central Processing Unit (CPU) 414, a system bus 412, a memory 416 connected to and accessible by other portions of the control unit 304 through system bus 412, and hardware entities 418 connected to system bus 412. System interface 424 allows the control unit 304 to receive data from first and second type sensors 116, 208 and from thumb buttons 118 and trigger 199. The system interface 424 can also be configured to allow control unit 304 to communicate with data transceiver 306, network equipment and other computing devices via a wired, wireless or both communications link or links. At least some of the hardware entities 418 perform actions involving access to and use of memory 416, which may be a random access memory (RAM), a disk drive, and/or a compact disc read only memory (CD-ROM).
As shown in FIG. 7 , the hardware entities 418 can include a disk drive unit 426 or other non-volatile memory (e.g., flash memory or ROM) comprising a computer-readable storage medium 428 on which is stored one or more sets of instructions 420 (e.g., software code). The instructions 420 can be configured to implement one or more of the methodologies, procedures, or functions described herein, such as generation of the output control signals for motion control of the robotic arms and/or motorized vehicle base 58, based upon inputs obtained from the control system 101. The instructions are also configured for sending haptic control signals to the control system 101 based upon data obtained from the robotic arm. The instructions 420 can also reside, completely or at least partially, within the memory 416 and/or within the CPU 414 during execution thereof. The memory 416 and the CPU 414 also can constitute machine-readable media. The term “machine-readable media”, as used herein, refers to a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions 420. The term “machine-readable media”, as used here, also refers to any medium that is capable of storing, encoding or carrying a set of instructions 420 for execution that cause the control unit 304 to perform any one or more of the methodologies of the present disclosure. As evident from the above discussion, the control unit 304 implements one or more method embodiments of the present invention, and in particular enables a user to control movement of the robotic arms 54, 60, and/or the movement of a motorized vehicle base 58.
The control unit 304 can communicate with the robotic arms 54, 60 and/or motorized vehicle base 58 by using wired or wireless connections. A suitable data transceiver can be used for this purpose. For example, if a wireless connection is preferred, a wireless data transceiver 306 can be provided. Notably, the data transceiver 306 may include more or less components than those shown in FIG. 7 . However, the components shown are sufficient to disclose an illustrative embodiment implementing the present invention.
If a wireless connection is used for communications with the robotic arm 60 and/or vehicle base 58, then the data transceiver 306 can comprise an antenna 402 (shown as element 114 in FIG. 1 ) for receiving and transmitting Radio Frequency (RF) signals. A receive/transmit (Rx/Tx) switch 404 selectively couples the antenna 402 to the transmitter circuitry 406 and receiver circuitry 408 in a manner familiar to those skilled in the art. The receiver circuitry 408 decodes the RF signals received from the robotic arm 60 to derive information therefrom; this includes haptic information that can be processed to provide haptic feedback to the user. The receiver circuitry 408 is coupled to control unit 304 via an electrical connection 434. The receiver circuitry 408 provides decoded RF signal information to the control unit 304. The control unit 304 uses the decoded RF signal information in accordance with the function(s) of the invention as described herein. In other embodiments a wired connection is made with one or more of the robotic arms and/or motorized vehicle base, as known in the art. For example, a wired connection may be more suitable for communications associated with control and haptic feedback from macro robotic arm 54. Such a wired connection would be functionally equivalent to a wireless connection shown in FIG. 3 but is easier to implement with respect to the communications circuitry 404, 406, 408.
The macro robotic arm 54 and micro robotic arm 60 are responsive to control signals for moving a grasping device 55, 70 in a plurality of directions. For example, the robotic arms 54, 60 can be capable of moving the respective grasping devices in as many as three orthogonal linear directions, and in a plurality of rotational directions defined by as many as three orthogonal axes of rotation. Robotic arms as described herein are known in the art and therefore will not be described here in detail. However, each of the macro and micro robotic arms 54, 60 can include a plurality of articulated joints 66 a-66 d and 68 a-68 c, each with a corresponding motor or driving device that sets the angular position of that joint as instructed by the control system 101. More specifically, by transmitting appropriate output control signals the control system can direct the controls actuators to control the spatial positioning of the tip of a grasping device 55, 70.
The position of the grasping device 55, 70 is preferably in accordance with the spatial positioning of the housing 106. The angular positioning of the robotic grasping devices 55, 70 is preferably controlled in accordance with the angular positioning of the grip 102. Power used to activate the articulated joints 68 a-68 c can be provided, for example, from the battery packs (not shown) associated with the robotic arm 60. Power used to activate the articulated joints 67 a-67 d can be provided by a power source associated with shelter 52. For example an onboard hydraulic pump (not shown) included with shelter 52 can be used for this purpose. The control unit 304, can obtain power from a battery or alternator associated with the shelter 52. Still, the invention is not limited in this regard and other power sources are also possible. One or both of the macro robotic arm and micro robotic arm can include a microprocessor based control unit and a data transceiver configured to respond to commands from the control system 101. For example if a wireless link is used to communicate with UGV 56, then the UGV can have an onboard wireless data transceiver similar to data transceiver 306, and appropriate control circuitry for responding to control signals.
One or both of the grasping devices 55, 70 can be a force-sensing grasping device, such as the one disclosed in U.S. application Ser. No. 13/049,982, entitled “Robotic Grasping Device With Multi-Force Sensing At Base Of Fingers,” filed on Mar. 17, 2011, the contents of which are incorporated herein by reference. The grasping device can include fingers, the angular positions of which can be controlled to open or close relative to each other to grasp and release objects. The grasping devices are arranged to send haptic signals back to the control unit 304, as measured by, for example, the fingers. The haptic signals are used by the control system to control motive devices in the control system 101 to create a corresponding haptic feedback response for the user. This haptic feedback response can indicate, for example, the magnitude and direction of a force present on the grasping device 55 and/or 70 and/or the forces upon one or both of the robotic arms 54, 60. This could include, for example, a force that is placing a torque on the arms 54, 60. This force can be felt by the user in the form of a corresponding resistance to motion of the grip 102 in the direction of this force, the resistance being proportional to the magnitude of the force impinging on the robotic arm. In this manner a user can sense, for example, the weight of an object, how hard the grasping device is pressing against a surface, and so forth. If, for example, the angular positioning of the fingers of the grasping device 55, 70 is determined by the displacement of the trigger 199, a haptic signal in the form of resistance to depression of the trigger 199 can be generated that is proportional to how hard the fingers are bearing down upon an object being gripped. In this manner a user can determine how firmly an object is being grasped.
One or more actuators, motors or servo motors (not shown) can be disposed in relation to the control system 101 to provide these haptic feedback signals via the grip 102 and trigger 199. For example, actuators or the like in the controller base 110 can apply a force to the articulating arms 107 b, 108 b, and 109 b, which is then experienced in the grip 102. More particularly, such actuators or motors can be disposed in motor housings 112 which are provided as part of base 110. The force generated by the motors is used to provide haptic feedback to the user. Additional actuators for providing haptic feedback to a user can be provided within grip 102 itself. For example, such an actuator can exert a force upon the trigger 199.
All of the apparatus, methods and algorithms disclosed and claimed herein can be made and executed without undue experimentation in light of the present disclosure. While the invention has been described in terms of preferred embodiments, it will be apparent to those of skill in the art that variations may be applied to the apparatus, methods and sequence of steps of the method without departing from the concept, spirit and scope of the invention. More specifically, it will be apparent that certain components may be added to, combined with, or substituted for the components described herein while the same or similar results would be achieved. All such similar substitutes and modifications apparent to those skilled in the art are deemed to be within the spirit, scope and concept of the invention as defined.
Claims (28)
1. A method for controlling a robot system, comprising:
controlling a movement of a macro robotic arm to move a vehicle from a first position to a second different position within a space, the macro robotic arm having (a) a proximal end coupled to a structure for sheltering an operator, (b) a distal end coupled to the vehicle, and (c) a first set of articulating joints and a macro control unit in a first system state responsive to a user manipulation of a control interface grip in a plurality of directions; and
controlling a movement of a micro robotic arm of the vehicle, the micro robotic arm (a) distinct from the macro robotic arm and (b) having a second set of articulating joints and a micro control unit in a second system state responsive to a user manipulation of said control interface grip in said plurality of directions.
2. The method according to claim 1 , further comprising transitioning between said first state and said second state responsive to at least one user input to said control system.
3. The method according to claim 1 , further comprising automatically transitioning between said first and second state responsive to said user manipulation of said control interface grip.
4. The method according to claim 1 , further comprising generating in said first state a first haptic force upon said control interface grip responsive to a detected force exerted upon said macro robotic arm, and generating in said second state a second haptic force upon said control interface grip responsive to a detected force exerted upon said micro robotic arm.
5. The method according to claim 1 , further comprising selecting said macro robotic arm to have a macro range of movement which exceeds a micro range of movement of said micro robotic arm.
6. The method according to claim 5 , further comprising selecting said macro robotic arm to have a macro precision of movement which is less precise than a micro precision of movement of said micro robotic arm.
7. The method according to claim 5 , further comprising using said macro robotic arm to facilitate positioning said micro robotic arm at a location where an object to be manipulated is within said micro range of movement of said micro robotic arm.
8. The method according to claim 7 , further comprising controlling a vehicle base to which said micro robotic arm is attached in a third system state to further facilitate positioning said micro robotic arm at said location.
9. The method according to claim 7 , further comprising transitioning between said first or second state and said third state responsive to at least one user input to said control system.
10. The method according to claim 7 , further comprising automatically transitioning from said first or second state to said third state responsive to said user manipulation of said control interface grip.
11. The method according to claim 8 , wherein said control of said vehicle base is effected responsive to a user manipulation of said control interface grip.
12. The method according to claim 1 , further comprising controlling a movement of a vehicle base to which said micro robotic arm is attached in a third system state responsive to a user manipulation of said control interface grip.
13. The method according to claim 12 , further comprising deploying said vehicle base to a terrain location using said macro robotic arm.
14. The method according to claim 1 , further comprising deploying and recovering a robot vehicle base to which said micro robotic arm is attached using said macro robotic arm.
15. A robot system, comprising:
a control system including a robot control interface having a control interface grip;
a macro robotic arm having (a) a proximal end coupled to a structure for sheltering an operator and (b) a first set of articulating joints responsive in a first control system state to movement of said control interface grip in a plurality of directions; and
a micro robotic arm (a) distinct from the macro robotic arm and (b) having a second set of articulating joints responsive in a second control system state to movement of said control interface grip in said plurality of directions;
wherein the macro robotic arm is used to relocate the micro robotic arm in a space.
16. The robot system according to claim 15 , wherein said control system is configured to transition between said first control system state and said second control system state responsive to at least one user input to said control system.
17. The robot system according to claim 15 , wherein said control system is configured to automatically transition between said first and second control system states responsive to movement of said control interface grip.
18. The robot system according to claim 15 , wherein said control system generates in said first control system state a first haptic force upon said control interface grip responsive to a detected force exerted upon said macro robotic arm, and generates in said second control system state a second haptic force upon said control interface grip responsive to a detected force exerted upon said micro robotic arm.
19. The robot system according to claim 15 , wherein said macro robotic arm has a macro range of movement which exceeds a micro range of movement of said micro robotic arm.
20. The robot system according to claim 19 , wherein said macro robotic arm has a macro precision of movement which is less precise than a micro precision of movement of said micro robotic arm.
21. The robot system according to claim 19 , wherein said micro robotic arm is configured to be supported at a distal end of said macro robotic arm to facilitate positioning said micro robotic arm.
22. The robot system according to claim 21 , further comprising a vehicle base to which said micro robotic arm is mounted, said vehicle base releasable from said distal end and configured for movement independent of said macro robotic arm.
23. The robot system according to claim 22 , wherein said control system is configured to control a motion of said vehicle base responsive to a user manipulation of said control interface grip in a third control system state.
24. The robot system according to claim 23 , wherein said control system is configured to transition between said first or second state and said third state responsive to at least one user input to said control system.
25. The robot system according to claim 23 , wherein said control system is configured to automatically transition from said first or second state to said third state responsive to said user manipulation of said control interface grip.
26. The robot system according to claim 15 , wherein said control system is configured to control a motion of a vehicle base to which said micro robotic arm is attached in a third system state responsive to movement of said control interface grip.
27. The robot system according to claim 26 , wherein said macro robotic arm includes a docking structure from which said vehicle base can be releasably deployed for movement of said vehicle base independent of said macro robotic arm.
28. The robot system according to claim 27 , wherein said docking structure is configured to securely engage said robot vehicle base after said vehicle base has been recovered, following deployment.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/667,048 US9638497B2 (en) | 2011-10-06 | 2015-03-24 | Improvised explosive device defeat system |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/267,357 US8996244B2 (en) | 2011-10-06 | 2011-10-06 | Improvised explosive device defeat system |
US14/667,048 US9638497B2 (en) | 2011-10-06 | 2015-03-24 | Improvised explosive device defeat system |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/267,357 Division US8996244B2 (en) | 2011-10-06 | 2011-10-06 | Improvised explosive device defeat system |
Publications (2)
Publication Number | Publication Date |
---|---|
US20150345914A1 US20150345914A1 (en) | 2015-12-03 |
US9638497B2 true US9638497B2 (en) | 2017-05-02 |
Family
ID=47143277
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/267,357 Active 2033-04-30 US8996244B2 (en) | 2011-10-06 | 2011-10-06 | Improvised explosive device defeat system |
US14/667,048 Active 2032-05-13 US9638497B2 (en) | 2011-10-06 | 2015-03-24 | Improvised explosive device defeat system |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/267,357 Active 2033-04-30 US8996244B2 (en) | 2011-10-06 | 2011-10-06 | Improvised explosive device defeat system |
Country Status (3)
Country | Link |
---|---|
US (2) | US8996244B2 (en) |
EP (1) | EP2764323B1 (en) |
WO (1) | WO2013052419A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180029516A1 (en) * | 2016-08-01 | 2018-02-01 | Toyota Motor Engineering & Manufacturing North America, Inc. | Vehicle docking and control systems for robots |
US20200307966A1 (en) * | 2017-12-21 | 2020-10-01 | Hans Kunz GmbH | Crane controller |
Families Citing this family (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9409297B2 (en) * | 2013-12-24 | 2016-08-09 | Tieman Vehicle Technologies LLC | Remote control button actuation module, system, and method |
US9576414B2 (en) * | 2013-12-24 | 2017-02-21 | Tieman Vehicle Technologies LLC | Remote control button actuation module, system, and method |
US20160082557A1 (en) * | 2014-09-22 | 2016-03-24 | Kuka Systems Corporation North America | Robotic apparatus and process for the installation of collars and nuts onto fasteners |
US9358684B1 (en) * | 2015-02-18 | 2016-06-07 | Merry Electronics Co., Ltd. | Wireless transmission device and robot arm using the same |
CN106032011A (en) * | 2015-03-07 | 2016-10-19 | 南京福尔摩斯智能科技有限公司 | A control type exoskeleton robot for exploder clearing |
US11221196B2 (en) * | 2015-03-30 | 2022-01-11 | Director General, Defence Research & Development Organisation (Drdo) | Vehicle and method for detecting and neutralizing an incendiary object |
US20170039671A1 (en) * | 2015-08-07 | 2017-02-09 | Seoul National University R&Db Foundation | Robotic self-filming system |
JP6778198B2 (en) * | 2015-08-25 | 2020-10-28 | 川崎重工業株式会社 | Remote control robot system |
SE539323C2 (en) | 2015-10-19 | 2017-07-04 | Husqvarna Ab | Improved control of remote demolition robot |
US10865578B2 (en) | 2016-07-15 | 2020-12-15 | Fastbrick Ip Pty Ltd | Boom for material transport |
WO2018009981A1 (en) | 2016-07-15 | 2018-01-18 | Fastbrick Ip Pty Ltd | Brick/block laying machine incorporated in a vehicle |
US10286564B2 (en) | 2017-05-01 | 2019-05-14 | Lincoln Global, Inc. | System for locally generating electricity on a robotic device |
DE102017112931A1 (en) * | 2017-06-13 | 2018-12-13 | Prüftechnik Dieter Busch Aktiengesellschaft | Mobile means of transporting data collectors, data collection system and data collection procedures |
WO2019006511A1 (en) | 2017-07-05 | 2019-01-10 | Fastbrick Ip Pty Ltd | Real time position and orientation tracker |
AU2018317941B2 (en) | 2017-08-17 | 2023-11-09 | Fastbrick Ip Pty Ltd | Laser tracker with improved roll angle measurement |
EP3669242A4 (en) | 2017-08-17 | 2021-07-21 | Fastbrick IP Pty Ltd | Communication system for an interaction system |
TR201714256A2 (en) * | 2017-09-26 | 2019-04-22 | Protarm Muehendislik Ve Ticaret A S | ONE ROBOTIC ARM |
ES2971624T3 (en) | 2017-10-11 | 2024-06-06 | Fastbrick Ip Pty Ltd | Machine for transporting objects |
US10875662B2 (en) * | 2018-04-19 | 2020-12-29 | Aurora Flight Sciences Corporation | Method of robot manipulation in a vibration environment |
US11340649B2 (en) | 2019-02-07 | 2022-05-24 | Blue Eclipse, Llc | Two button remote control actuator |
CN111923059A (en) * | 2020-07-29 | 2020-11-13 | 合肥瞬瞬科技有限公司 | 5G vision intelligence blasting robot |
Citations (123)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3280991A (en) | 1964-04-28 | 1966-10-25 | Programmed & Remote Syst Corp | Position control manipulator |
US3637092A (en) | 1970-04-30 | 1972-01-25 | Gen Electric | Material-handling apparatus |
US4216467A (en) | 1977-12-22 | 1980-08-05 | Westinghouse Electric Corp. | Hand controller |
US4521685A (en) | 1982-03-01 | 1985-06-04 | Lord Corporation | Tactile sensor for an industrial robot or the like |
US4604016A (en) | 1983-08-03 | 1986-08-05 | Joyce Stephen A | Multi-dimensional force-torque hand controller having force feedback |
US4655673A (en) | 1983-05-10 | 1987-04-07 | Graham S. Hawkes | Apparatus providing tactile feedback to operators of remotely controlled manipulators |
US4661032A (en) | 1984-12-20 | 1987-04-28 | Agency Of Industrial Science & Technology, Ministry Of International Trade & Industry | Bilateral master-slave manipulator control device |
US4762006A (en) | 1984-09-29 | 1988-08-09 | Fujitsu Limited | Force-detecting apparatus |
US4791588A (en) | 1984-03-09 | 1988-12-13 | Fujitsu Limited | Movable apparatus driving system |
US4795296A (en) | 1986-11-17 | 1989-01-03 | California Institute Of Technology | Hand-held robot end effector controller having movement and force control |
US4837734A (en) | 1986-02-26 | 1989-06-06 | Hitachi, Ltd. | Method and apparatus for master-slave manipulation supplemented by automatic control based on level of operator skill |
US4842308A (en) | 1988-06-23 | 1989-06-27 | Australux North America Limited | Rotation limiting ball-joint conduit apparatus |
US4853874A (en) | 1986-12-12 | 1989-08-01 | Hitachi, Ltd. | Master-slave manipulators with scaling |
US4860215A (en) | 1987-04-06 | 1989-08-22 | California Institute Of Technology | Method and apparatus for adaptive force and position control of manipulators |
US4893254A (en) | 1988-04-20 | 1990-01-09 | University Of British Columbia | Manipulator arm position sensing |
US4893981A (en) | 1987-03-26 | 1990-01-16 | Kabushiki Kaisha Komatsu Seisakusho | Master/slave type manipulator |
GB2228783A (en) | 1989-03-03 | 1990-09-05 | Atomic Energy Authority Uk | Multi-axis hand controller |
US4975856A (en) | 1986-02-18 | 1990-12-04 | Robotics Research Corporation | Motion controller for redundant or nonredundant linkages |
US5004391A (en) | 1989-08-21 | 1991-04-02 | Rutgers University | Portable dextrous force feedback master for robot telemanipulation |
US5018922A (en) | 1987-03-26 | 1991-05-28 | Kabushiki Kaisha Komatsu Seisakusho | Master/slave type manipulator |
US5092645A (en) | 1987-09-18 | 1992-03-03 | Wacoh Corporation | Robotic gripper having strain sensors formed on a semiconductor substrate |
US5178032A (en) | 1990-10-04 | 1993-01-12 | Comau Spa | Robot wrist |
US5184319A (en) | 1990-02-02 | 1993-02-02 | Kramer James F | Force feedback and textures simulating interface device |
US5193963A (en) | 1990-10-31 | 1993-03-16 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Force reflecting hand controller |
US5231693A (en) | 1991-05-09 | 1993-07-27 | The United States Of America As Represented By The Administrator, National Aeronautics And Space Administration | Telerobot control system |
US5382885A (en) | 1993-08-09 | 1995-01-17 | The University Of British Columbia | Motion scaling tele-operating system with force feedback suitable for microsurgery |
US5413454A (en) | 1993-07-09 | 1995-05-09 | Movsesian; Peter | Mobile robotic arm |
US5430643A (en) | 1992-03-11 | 1995-07-04 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Configuration control of seven degree of freedom arms |
US5451924A (en) | 1993-01-14 | 1995-09-19 | Massachusetts Institute Of Technology | Apparatus for providing sensory substitution of force feedback |
EP0672507A1 (en) | 1994-02-21 | 1995-09-20 | Asea Brown Boveri Ab | Method for controlling the movements of an industrial robot at and near singularities |
WO1995030571A1 (en) | 1994-05-09 | 1995-11-16 | The Trustees Of The University Of Pennsylvania | Adaptive mobility system |
US5508596A (en) | 1993-10-07 | 1996-04-16 | Omax Corporation | Motion control with precomputation |
US5565891A (en) | 1992-03-05 | 1996-10-15 | Armstrong; Brad A. | Six degrees of freedom graphics controller |
US5648897A (en) | 1994-04-22 | 1997-07-15 | Northrop Grumman Corporation | System for controlling a remote unit |
US5694013A (en) | 1996-09-06 | 1997-12-02 | Ford Global Technologies, Inc. | Force feedback haptic interface for a three-dimensional CAD surface |
US5737500A (en) | 1992-03-11 | 1998-04-07 | California Institute Of Technology | Mobile dexterous siren degree of freedom robot arm with real-time control system |
US5792165A (en) | 1993-07-21 | 1998-08-11 | Charles H. Klieman | Endoscopic instrument with detachable end effector |
US5831408A (en) | 1992-12-02 | 1998-11-03 | Cybernet Systems Corporation | Force feedback system |
US6028593A (en) | 1995-12-01 | 2000-02-22 | Immersion Corporation | Method and apparatus for providing simulated physical interactions within computer generated environments |
US6047610A (en) | 1997-04-18 | 2000-04-11 | Stocco; Leo J | Hybrid serial/parallel manipulator |
US6084587A (en) | 1996-08-02 | 2000-07-04 | Sensable Technologies, Inc. | Method and apparatus for generating and interfacing with a haptic virtual reality environment |
US6088017A (en) | 1995-11-30 | 2000-07-11 | Virtual Technologies, Inc. | Tactile feedback man-machine interface device |
US6178775B1 (en) | 1998-10-30 | 2001-01-30 | The Boc Group, Inc. | Method and apparatus for separating air to produce an oxygen product |
US6184868B1 (en) | 1998-09-17 | 2001-02-06 | Immersion Corp. | Haptic feedback control devices |
US6191796B1 (en) | 1998-01-21 | 2001-02-20 | Sensable Technologies, Inc. | Method and apparatus for generating and interfacing with rigid and deformable surfaces in a haptic virtual reality environment |
US20010002098A1 (en) | 1997-11-03 | 2001-05-31 | Douglas Haanpaa | Haptic pointing devices |
US6246390B1 (en) | 1995-01-18 | 2001-06-12 | Immersion Corporation | Multiple degree-of-freedom mechanical interface to a computer system |
US6271833B1 (en) | 1995-09-27 | 2001-08-07 | Immersion Corp. | Low cost force feedback peripheral with button activated feel sensations |
US20010037163A1 (en) | 2000-05-01 | 2001-11-01 | Irobot Corporation | Method and system for remote control of mobile robot |
US6522952B1 (en) | 1999-06-01 | 2003-02-18 | Japan As Represented By Secretary Of Agency Of Industrial Science And Technology | Method and system for controlling cooperative object-transporting robot |
WO2003055061A1 (en) | 2001-12-19 | 2003-07-03 | Analog Devices, Inc. | Differential parametric amplifier with modulated mems capacitors |
US6592315B2 (en) | 2000-05-08 | 2003-07-15 | William Joseph Osborne, Jr. | Self-feeding apparatus with hover mode |
US20030169235A1 (en) | 2002-03-07 | 2003-09-11 | Gron Mikkel Hartmann | Three dimensional track ball systems |
US20030179097A1 (en) * | 2001-02-02 | 2003-09-25 | David Collingwood | Machine tool probe |
US6636161B2 (en) | 1996-11-26 | 2003-10-21 | Immersion Corporation | Isometric haptic feedback interface |
US6705871B1 (en) | 1996-09-06 | 2004-03-16 | Immersion Corporation | Method and apparatus for providing an interface mechanism for a computer simulation |
US6781569B1 (en) | 1999-06-11 | 2004-08-24 | Immersion Corporation | Hand controller |
US6793653B2 (en) | 2001-12-08 | 2004-09-21 | Computer Motion, Inc. | Multifunctional handle for a medical robotic system |
US20040189675A1 (en) | 2002-12-30 | 2004-09-30 | John Pretlove | Augmented reality system and method |
US6801008B1 (en) | 1992-12-02 | 2004-10-05 | Immersion Corporation | Force feedback system and actuator power management |
US20040254771A1 (en) | 2001-06-25 | 2004-12-16 | Robert Riener | Programmable joint simulator with force and motion feedback |
US6857878B1 (en) | 1998-01-26 | 2005-02-22 | Simbionix Ltd. | Endoscopic tutorial system |
US20050087373A1 (en) | 2003-10-28 | 2005-04-28 | Tsutomu Wakitani | Electric vehicle |
US20050154431A1 (en) * | 2003-12-30 | 2005-07-14 | Liposonix, Inc. | Systems and methods for the destruction of adipose tissue |
US20050252329A1 (en) | 2004-05-13 | 2005-11-17 | Jean-Guy Demers | Haptic mechanism |
WO2006016799A1 (en) | 2004-05-27 | 2006-02-16 | Exact Dynamics B.V. | Wheelchair with mechanical arm |
US20060048364A1 (en) | 2004-09-08 | 2006-03-09 | Hui Zhang | Robotic machining with a flexible manipulator |
US20060066574A1 (en) | 2003-05-21 | 2006-03-30 | Munsang Kim | Parallel haptic joystick system |
US20060117258A1 (en) | 2004-11-30 | 2006-06-01 | Raymond Yu | User interface device |
US20060178775A1 (en) | 2005-02-04 | 2006-08-10 | George Zhang | Accelerometer to monitor movement of a tool assembly attached to a robot end effector |
US20070013336A1 (en) | 2005-05-19 | 2007-01-18 | Intuitive Surgical Inc. | Software center and highly configurable robotic systems for surgery and other uses |
US7168748B2 (en) | 2002-09-26 | 2007-01-30 | Barrett Technology, Inc. | Intelligent, self-contained robotic hand |
US20070050139A1 (en) | 2005-04-27 | 2007-03-01 | Sidman Adam D | Handheld platform stabilization system employing distributed rotation sensors |
US7208900B2 (en) | 2001-10-23 | 2007-04-24 | Abb Ab | Industrial robot system |
WO2007051000A2 (en) | 2005-10-27 | 2007-05-03 | Immersion Corporation | System and method for controlling force applied to and manipulation of medical instruments |
US7225404B1 (en) | 1996-04-04 | 2007-05-29 | Massachusetts Institute Of Technology | Method and apparatus for determining forces to be applied to a user through a haptic interface |
FR2898824A1 (en) | 2006-03-27 | 2007-09-28 | Commissariat Energie Atomique | Intelligent interface device for e.g. grasping object, has controls permitting displacement of clamp towards top, bottom, left and right, respectively, and marking unit marking rectangular zone surrounding object in image using input unit |
EP1876505A1 (en) | 2006-07-03 | 2008-01-09 | Force Dimension S.à.r.l | Haptic device gravity compensation |
US20080009971A1 (en) | 2006-07-05 | 2008-01-10 | Samsung Electronics Co., Ltd. | Walking robot and control method thereof |
US20080063400A1 (en) | 2006-05-12 | 2008-03-13 | Irobot Corporation | Method and Device for Controlling a Remote Vehicle |
US7345672B2 (en) | 1992-12-02 | 2008-03-18 | Immersion Corporation | Force feedback system and actuator power management |
US20080161733A1 (en) | 2004-02-05 | 2008-07-03 | Motorika Limited | Methods and Apparatuses for Rehabilitation and Training |
US7411576B2 (en) | 2003-10-30 | 2008-08-12 | Sensable Technologies, Inc. | Force reflecting haptic interface |
US20080266254A1 (en) | 2007-04-24 | 2008-10-30 | Irobot Corporation | Control System for a Remote Vehicle |
WO2008135978A2 (en) | 2007-05-06 | 2008-11-13 | Wave Group Ltd. | A robotic platform |
US7480600B2 (en) | 1993-10-01 | 2009-01-20 | The Massachusetts Institute Of Technology | Force reflecting haptic interface |
US20090074252A1 (en) | 2007-10-26 | 2009-03-19 | Honda Motor Co., Ltd. | Real-time self collision and obstacle avoidance |
US20090182436A1 (en) | 2006-02-24 | 2009-07-16 | Paolo Ferrara | Robot Arm |
US20090234499A1 (en) | 2008-03-13 | 2009-09-17 | Battelle Energy Alliance, Llc | System and method for seamless task-directed autonomy for robots |
US20100023185A1 (en) | 2008-07-28 | 2010-01-28 | Torc Technologies, Llc | Devices and methods for waypoint target generation and mission spooling for mobile ground robots |
US20100041991A1 (en) | 2006-09-25 | 2010-02-18 | Koninklijke Philips Electronics N.V. | Haptic feedback medical scanning methods and systems |
US20100070079A1 (en) | 2008-09-18 | 2010-03-18 | Intouch Technologies, Inc. | Mobile videoconferencing robot system with network adaptive driving |
US20100084513A1 (en) | 2008-09-09 | 2010-04-08 | Aeryon Labs Inc. | Method and system for directing unmanned vehicles |
WO2010040215A1 (en) | 2008-10-06 | 2010-04-15 | Kinova | Portable robotic arm |
US20100092267A1 (en) | 2006-12-19 | 2010-04-15 | Deakin University | Method and apparatus for haptic control |
US20100100256A1 (en) | 2007-03-28 | 2010-04-22 | Jacob Curtis Jurmain | Remote Vehicle Control System and Method |
US20100169815A1 (en) | 2008-12-31 | 2010-07-01 | Intuitive Surgical, Inc. | Visual force feedback in a minimally invasive surgical procedure |
US20100168918A1 (en) | 2008-12-31 | 2010-07-01 | Intuitive Surgical, Inc. | Obtaining force information in a minimally invasive surgical procedure |
WO2010085184A1 (en) | 2009-01-20 | 2010-07-29 | Husqvarna Ab | Control system for a remote control work machine |
US7783384B2 (en) | 2006-05-31 | 2010-08-24 | Kraft Brett W | Ambidextrous robotic master controller |
US20100259614A1 (en) | 2009-04-14 | 2010-10-14 | Honeywell International Inc. | Delay Compensated Feature Target System |
US20110015569A1 (en) | 2008-03-27 | 2011-01-20 | Kirschenman Mark B | Robotic catheter system input device |
US20110046781A1 (en) | 2009-08-21 | 2011-02-24 | Harris Corporation, Corporation Of The State Of Delaware | Coordinated action robotic system and related methods |
US7933667B2 (en) | 2004-06-24 | 2011-04-26 | Abb Ab | Industrial robot system with a portable operator control device |
US20110106339A1 (en) | 2006-07-14 | 2011-05-05 | Emilie Phillips | Autonomous Behaviors for a Remote Vehicle |
US20110144828A1 (en) | 2009-12-11 | 2011-06-16 | The Boeing Company | Unmanned Multi-Purpose Ground Vehicle with Different Levels of Control |
WO2011075093A1 (en) | 2009-12-14 | 2011-06-23 | Gokhan Vargin Gok | A multi axis robot |
US20110155785A1 (en) | 2009-12-24 | 2011-06-30 | Ethicon Endo-Surgery, Inc. | Motor-driven surgical cutting instrument with electric actuator directional control assembly |
US20110255945A1 (en) * | 2007-12-05 | 2011-10-20 | Graham Siddle | Cash handling system |
US20120095619A1 (en) * | 2010-05-11 | 2012-04-19 | Irobot Corporation | Remote Vehicle Missions and Systems for Supporting Remote Vehicle Missions |
US20120150351A1 (en) | 2010-12-09 | 2012-06-14 | Harris Corporation | Ball joint having a passageway for routing a cable therethrough |
US20120185098A1 (en) | 2011-01-19 | 2012-07-19 | Harris Corporation | Telematic interface with directional translation |
US20120184955A1 (en) | 2008-01-16 | 2012-07-19 | Catheter Robotics Inc. | Remotely Controlled Catheter Insertion System with Automatic Control System |
US20120185099A1 (en) | 2011-01-19 | 2012-07-19 | Harris Corporation | Telematic interface with control signal scaling based on force sensor feedback |
US8226072B2 (en) | 2008-04-24 | 2012-07-24 | Toyota Jidosha Kabushiki Kaisha | Power assist apparatus with a controlled brake mechanism for positioning a workpiece and control method thereof |
US20120294696A1 (en) | 2011-05-20 | 2012-11-22 | Harris Corporation | Haptic device for manipulator and vehicle control |
US20120306741A1 (en) | 2011-06-06 | 2012-12-06 | Gupta Kalyan M | System and Method for Enhancing Locative Response Abilities of Autonomous and Semi-Autonomous Agents |
US8373391B1 (en) | 2008-10-02 | 2013-02-12 | Esterline Technologies Corporation | Rechargeable hand-held devices using capacitors, such as supercapacitors |
US20130090194A1 (en) | 2010-06-17 | 2013-04-11 | Commissariat A L'energie Atomique Et Aux Energies Alternatives | Reducing device having a high reduction ratio, robot and haptic interface comprising at least one such reducing device |
US8447440B2 (en) | 2007-05-14 | 2013-05-21 | iRobot Coporation | Autonomous behaviors for a remote vehicle |
US20130328770A1 (en) | 2010-02-23 | 2013-12-12 | Muv Interactive Ltd. | System for projecting content to a display surface having user-controlled size, shape and location/direction and apparatus and methods useful in conjunction therewith |
US20140031983A1 (en) | 2011-03-23 | 2014-01-30 | Sri International | Dexterous telemanipulator system |
US8950286B2 (en) | 2009-10-02 | 2015-02-10 | Commissariat à l'énergie atomique et aux énergies alternatives | Robot or haptic interface structure with parallel arms |
-
2011
- 2011-10-06 US US13/267,357 patent/US8996244B2/en active Active
-
2012
- 2012-10-01 WO PCT/US2012/058303 patent/WO2013052419A1/en active Application Filing
- 2012-10-01 EP EP12781516.5A patent/EP2764323B1/en active Active
-
2015
- 2015-03-24 US US14/667,048 patent/US9638497B2/en active Active
Patent Citations (138)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US3280991A (en) | 1964-04-28 | 1966-10-25 | Programmed & Remote Syst Corp | Position control manipulator |
US3637092A (en) | 1970-04-30 | 1972-01-25 | Gen Electric | Material-handling apparatus |
US4216467A (en) | 1977-12-22 | 1980-08-05 | Westinghouse Electric Corp. | Hand controller |
US4521685A (en) | 1982-03-01 | 1985-06-04 | Lord Corporation | Tactile sensor for an industrial robot or the like |
US4655673A (en) | 1983-05-10 | 1987-04-07 | Graham S. Hawkes | Apparatus providing tactile feedback to operators of remotely controlled manipulators |
US4604016A (en) | 1983-08-03 | 1986-08-05 | Joyce Stephen A | Multi-dimensional force-torque hand controller having force feedback |
US4791588A (en) | 1984-03-09 | 1988-12-13 | Fujitsu Limited | Movable apparatus driving system |
US4862751A (en) | 1984-09-29 | 1989-09-05 | Fujitsu Limited | Force-detecting apparatus |
US4762006A (en) | 1984-09-29 | 1988-08-09 | Fujitsu Limited | Force-detecting apparatus |
US4661032A (en) | 1984-12-20 | 1987-04-28 | Agency Of Industrial Science & Technology, Ministry Of International Trade & Industry | Bilateral master-slave manipulator control device |
US4975856A (en) | 1986-02-18 | 1990-12-04 | Robotics Research Corporation | Motion controller for redundant or nonredundant linkages |
US4837734A (en) | 1986-02-26 | 1989-06-06 | Hitachi, Ltd. | Method and apparatus for master-slave manipulation supplemented by automatic control based on level of operator skill |
US4795296A (en) | 1986-11-17 | 1989-01-03 | California Institute Of Technology | Hand-held robot end effector controller having movement and force control |
US4853874A (en) | 1986-12-12 | 1989-08-01 | Hitachi, Ltd. | Master-slave manipulators with scaling |
US4893981A (en) | 1987-03-26 | 1990-01-16 | Kabushiki Kaisha Komatsu Seisakusho | Master/slave type manipulator |
US5018922A (en) | 1987-03-26 | 1991-05-28 | Kabushiki Kaisha Komatsu Seisakusho | Master/slave type manipulator |
US4860215A (en) | 1987-04-06 | 1989-08-22 | California Institute Of Technology | Method and apparatus for adaptive force and position control of manipulators |
US5092645A (en) | 1987-09-18 | 1992-03-03 | Wacoh Corporation | Robotic gripper having strain sensors formed on a semiconductor substrate |
US4893254A (en) | 1988-04-20 | 1990-01-09 | University Of British Columbia | Manipulator arm position sensing |
US4842308A (en) | 1988-06-23 | 1989-06-27 | Australux North America Limited | Rotation limiting ball-joint conduit apparatus |
GB2228783A (en) | 1989-03-03 | 1990-09-05 | Atomic Energy Authority Uk | Multi-axis hand controller |
US5007300A (en) | 1989-03-03 | 1991-04-16 | United Kingdom Atomic Energy Authority | Multi-axis hand controller |
US5004391A (en) | 1989-08-21 | 1991-04-02 | Rutgers University | Portable dextrous force feedback master for robot telemanipulation |
US5184319A (en) | 1990-02-02 | 1993-02-02 | Kramer James F | Force feedback and textures simulating interface device |
US5178032A (en) | 1990-10-04 | 1993-01-12 | Comau Spa | Robot wrist |
US5193963A (en) | 1990-10-31 | 1993-03-16 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Force reflecting hand controller |
US5231693A (en) | 1991-05-09 | 1993-07-27 | The United States Of America As Represented By The Administrator, National Aeronautics And Space Administration | Telerobot control system |
US5589828A (en) | 1992-03-05 | 1996-12-31 | Armstrong; Brad A. | 6 Degrees of freedom controller with capability of tactile feedback |
US5565891A (en) | 1992-03-05 | 1996-10-15 | Armstrong; Brad A. | Six degrees of freedom graphics controller |
US5737500A (en) | 1992-03-11 | 1998-04-07 | California Institute Of Technology | Mobile dexterous siren degree of freedom robot arm with real-time control system |
US5430643A (en) | 1992-03-11 | 1995-07-04 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Configuration control of seven degree of freedom arms |
US6801008B1 (en) | 1992-12-02 | 2004-10-05 | Immersion Corporation | Force feedback system and actuator power management |
US5831408A (en) | 1992-12-02 | 1998-11-03 | Cybernet Systems Corporation | Force feedback system |
US7345672B2 (en) | 1992-12-02 | 2008-03-18 | Immersion Corporation | Force feedback system and actuator power management |
US6104158A (en) | 1992-12-02 | 2000-08-15 | Immersion Corporation | Force feedback system |
US5451924A (en) | 1993-01-14 | 1995-09-19 | Massachusetts Institute Of Technology | Apparatus for providing sensory substitution of force feedback |
US5619180A (en) | 1993-01-14 | 1997-04-08 | Massachusetts Inst Technology | Apparatus for providing vibrotactile sensory substitution of force feedback |
US5413454A (en) | 1993-07-09 | 1995-05-09 | Movsesian; Peter | Mobile robotic arm |
US5792165A (en) | 1993-07-21 | 1998-08-11 | Charles H. Klieman | Endoscopic instrument with detachable end effector |
US5382885A (en) | 1993-08-09 | 1995-01-17 | The University Of British Columbia | Motion scaling tele-operating system with force feedback suitable for microsurgery |
US7480600B2 (en) | 1993-10-01 | 2009-01-20 | The Massachusetts Institute Of Technology | Force reflecting haptic interface |
US5508596A (en) | 1993-10-07 | 1996-04-16 | Omax Corporation | Motion control with precomputation |
EP0672507A1 (en) | 1994-02-21 | 1995-09-20 | Asea Brown Boveri Ab | Method for controlling the movements of an industrial robot at and near singularities |
US5648897A (en) | 1994-04-22 | 1997-07-15 | Northrop Grumman Corporation | System for controlling a remote unit |
WO1995030571A1 (en) | 1994-05-09 | 1995-11-16 | The Trustees Of The University Of Pennsylvania | Adaptive mobility system |
US6246390B1 (en) | 1995-01-18 | 2001-06-12 | Immersion Corporation | Multiple degree-of-freedom mechanical interface to a computer system |
US6271833B1 (en) | 1995-09-27 | 2001-08-07 | Immersion Corp. | Low cost force feedback peripheral with button activated feel sensations |
US6088017A (en) | 1995-11-30 | 2000-07-11 | Virtual Technologies, Inc. | Tactile feedback man-machine interface device |
US6028593A (en) | 1995-12-01 | 2000-02-22 | Immersion Corporation | Method and apparatus for providing simulated physical interactions within computer generated environments |
US7158112B2 (en) | 1995-12-01 | 2007-01-02 | Immersion Corporation | Interactions between simulated objects with force feedback |
US7225404B1 (en) | 1996-04-04 | 2007-05-29 | Massachusetts Institute Of Technology | Method and apparatus for determining forces to be applied to a user through a haptic interface |
US6084587A (en) | 1996-08-02 | 2000-07-04 | Sensable Technologies, Inc. | Method and apparatus for generating and interfacing with a haptic virtual reality environment |
US6705871B1 (en) | 1996-09-06 | 2004-03-16 | Immersion Corporation | Method and apparatus for providing an interface mechanism for a computer simulation |
US5694013A (en) | 1996-09-06 | 1997-12-02 | Ford Global Technologies, Inc. | Force feedback haptic interface for a three-dimensional CAD surface |
US6636161B2 (en) | 1996-11-26 | 2003-10-21 | Immersion Corporation | Isometric haptic feedback interface |
US6047610A (en) | 1997-04-18 | 2000-04-11 | Stocco; Leo J | Hybrid serial/parallel manipulator |
US20010002098A1 (en) | 1997-11-03 | 2001-05-31 | Douglas Haanpaa | Haptic pointing devices |
US6281651B1 (en) | 1997-11-03 | 2001-08-28 | Immersion Corporation | Haptic pointing devices |
US6191796B1 (en) | 1998-01-21 | 2001-02-20 | Sensable Technologies, Inc. | Method and apparatus for generating and interfacing with rigid and deformable surfaces in a haptic virtual reality environment |
US6857878B1 (en) | 1998-01-26 | 2005-02-22 | Simbionix Ltd. | Endoscopic tutorial system |
US6184868B1 (en) | 1998-09-17 | 2001-02-06 | Immersion Corp. | Haptic feedback control devices |
US6178775B1 (en) | 1998-10-30 | 2001-01-30 | The Boc Group, Inc. | Method and apparatus for separating air to produce an oxygen product |
US6522952B1 (en) | 1999-06-01 | 2003-02-18 | Japan As Represented By Secretary Of Agency Of Industrial Science And Technology | Method and system for controlling cooperative object-transporting robot |
US6781569B1 (en) | 1999-06-11 | 2004-08-24 | Immersion Corporation | Hand controller |
US6535793B2 (en) | 2000-05-01 | 2003-03-18 | Irobot Corporation | Method and system for remote control of mobile robot |
US20010037163A1 (en) | 2000-05-01 | 2001-11-01 | Irobot Corporation | Method and system for remote control of mobile robot |
US6592315B2 (en) | 2000-05-08 | 2003-07-15 | William Joseph Osborne, Jr. | Self-feeding apparatus with hover mode |
US20030179097A1 (en) * | 2001-02-02 | 2003-09-25 | David Collingwood | Machine tool probe |
US20040254771A1 (en) | 2001-06-25 | 2004-12-16 | Robert Riener | Programmable joint simulator with force and motion feedback |
US7208900B2 (en) | 2001-10-23 | 2007-04-24 | Abb Ab | Industrial robot system |
US6793653B2 (en) | 2001-12-08 | 2004-09-21 | Computer Motion, Inc. | Multifunctional handle for a medical robotic system |
WO2003055061A1 (en) | 2001-12-19 | 2003-07-03 | Analog Devices, Inc. | Differential parametric amplifier with modulated mems capacitors |
US20030169235A1 (en) | 2002-03-07 | 2003-09-11 | Gron Mikkel Hartmann | Three dimensional track ball systems |
US7168748B2 (en) | 2002-09-26 | 2007-01-30 | Barrett Technology, Inc. | Intelligent, self-contained robotic hand |
US20040189675A1 (en) | 2002-12-30 | 2004-09-30 | John Pretlove | Augmented reality system and method |
US7714895B2 (en) | 2002-12-30 | 2010-05-11 | Abb Research Ltd. | Interactive and shared augmented reality system and method having local and remote access |
US20060066574A1 (en) | 2003-05-21 | 2006-03-30 | Munsang Kim | Parallel haptic joystick system |
US7138981B2 (en) | 2003-05-21 | 2006-11-21 | Korea Institute Of Science And Technology | Parallel haptic joystick system |
US20050087373A1 (en) | 2003-10-28 | 2005-04-28 | Tsutomu Wakitani | Electric vehicle |
US7411576B2 (en) | 2003-10-30 | 2008-08-12 | Sensable Technologies, Inc. | Force reflecting haptic interface |
US20050154431A1 (en) * | 2003-12-30 | 2005-07-14 | Liposonix, Inc. | Systems and methods for the destruction of adipose tissue |
US20080161733A1 (en) | 2004-02-05 | 2008-07-03 | Motorika Limited | Methods and Apparatuses for Rehabilitation and Training |
US20050252329A1 (en) | 2004-05-13 | 2005-11-17 | Jean-Guy Demers | Haptic mechanism |
WO2006016799A1 (en) | 2004-05-27 | 2006-02-16 | Exact Dynamics B.V. | Wheelchair with mechanical arm |
US20070095582A1 (en) | 2004-05-27 | 2007-05-03 | Exact Dynamics B.V. | Wheelchair with mechanical arm |
US7933667B2 (en) | 2004-06-24 | 2011-04-26 | Abb Ab | Industrial robot system with a portable operator control device |
US20060048364A1 (en) | 2004-09-08 | 2006-03-09 | Hui Zhang | Robotic machining with a flexible manipulator |
US20060117258A1 (en) | 2004-11-30 | 2006-06-01 | Raymond Yu | User interface device |
US20060178775A1 (en) | 2005-02-04 | 2006-08-10 | George Zhang | Accelerometer to monitor movement of a tool assembly attached to a robot end effector |
US20070050139A1 (en) | 2005-04-27 | 2007-03-01 | Sidman Adam D | Handheld platform stabilization system employing distributed rotation sensors |
US20070013336A1 (en) | 2005-05-19 | 2007-01-18 | Intuitive Surgical Inc. | Software center and highly configurable robotic systems for surgery and other uses |
WO2007051000A2 (en) | 2005-10-27 | 2007-05-03 | Immersion Corporation | System and method for controlling force applied to and manipulation of medical instruments |
US20090182436A1 (en) | 2006-02-24 | 2009-07-16 | Paolo Ferrara | Robot Arm |
US20100172733A1 (en) | 2006-03-27 | 2010-07-08 | Commissariat A L'energie Atomique | Intelligent interface device for grasping of an object by a manipulating robot and method of implementing this device |
FR2898824A1 (en) | 2006-03-27 | 2007-09-28 | Commissariat Energie Atomique | Intelligent interface device for e.g. grasping object, has controls permitting displacement of clamp towards top, bottom, left and right, respectively, and marking unit marking rectangular zone surrounding object in image using input unit |
US20080063400A1 (en) | 2006-05-12 | 2008-03-13 | Irobot Corporation | Method and Device for Controlling a Remote Vehicle |
US7783384B2 (en) | 2006-05-31 | 2010-08-24 | Kraft Brett W | Ambidextrous robotic master controller |
US20100019890A1 (en) | 2006-07-03 | 2010-01-28 | Force Dimension S.A.R.L. | Haptic Device Gravity Compensation |
EP1876505A1 (en) | 2006-07-03 | 2008-01-09 | Force Dimension S.à.r.l | Haptic device gravity compensation |
US20080009971A1 (en) | 2006-07-05 | 2008-01-10 | Samsung Electronics Co., Ltd. | Walking robot and control method thereof |
US20110106339A1 (en) | 2006-07-14 | 2011-05-05 | Emilie Phillips | Autonomous Behaviors for a Remote Vehicle |
US20100041991A1 (en) | 2006-09-25 | 2010-02-18 | Koninklijke Philips Electronics N.V. | Haptic feedback medical scanning methods and systems |
US20100092267A1 (en) | 2006-12-19 | 2010-04-15 | Deakin University | Method and apparatus for haptic control |
US20100100256A1 (en) | 2007-03-28 | 2010-04-22 | Jacob Curtis Jurmain | Remote Vehicle Control System and Method |
US20080266254A1 (en) | 2007-04-24 | 2008-10-30 | Irobot Corporation | Control System for a Remote Vehicle |
WO2008135978A2 (en) | 2007-05-06 | 2008-11-13 | Wave Group Ltd. | A robotic platform |
US8447440B2 (en) | 2007-05-14 | 2013-05-21 | iRobot Coporation | Autonomous behaviors for a remote vehicle |
US20090074252A1 (en) | 2007-10-26 | 2009-03-19 | Honda Motor Co., Ltd. | Real-time self collision and obstacle avoidance |
US20110255945A1 (en) * | 2007-12-05 | 2011-10-20 | Graham Siddle | Cash handling system |
US20120184955A1 (en) | 2008-01-16 | 2012-07-19 | Catheter Robotics Inc. | Remotely Controlled Catheter Insertion System with Automatic Control System |
US20090234499A1 (en) | 2008-03-13 | 2009-09-17 | Battelle Energy Alliance, Llc | System and method for seamless task-directed autonomy for robots |
US20110015569A1 (en) | 2008-03-27 | 2011-01-20 | Kirschenman Mark B | Robotic catheter system input device |
US8226072B2 (en) | 2008-04-24 | 2012-07-24 | Toyota Jidosha Kabushiki Kaisha | Power assist apparatus with a controlled brake mechanism for positioning a workpiece and control method thereof |
US20100023185A1 (en) | 2008-07-28 | 2010-01-28 | Torc Technologies, Llc | Devices and methods for waypoint target generation and mission spooling for mobile ground robots |
US20100084513A1 (en) | 2008-09-09 | 2010-04-08 | Aeryon Labs Inc. | Method and system for directing unmanned vehicles |
US20100070079A1 (en) | 2008-09-18 | 2010-03-18 | Intouch Technologies, Inc. | Mobile videoconferencing robot system with network adaptive driving |
US8373391B1 (en) | 2008-10-02 | 2013-02-12 | Esterline Technologies Corporation | Rechargeable hand-held devices using capacitors, such as supercapacitors |
US20110257786A1 (en) | 2008-10-06 | 2011-10-20 | Caron L Ecuyer Louis Joseph | Portable robotic arm |
WO2010040215A1 (en) | 2008-10-06 | 2010-04-15 | Kinova | Portable robotic arm |
US20100168918A1 (en) | 2008-12-31 | 2010-07-01 | Intuitive Surgical, Inc. | Obtaining force information in a minimally invasive surgical procedure |
US20100169815A1 (en) | 2008-12-31 | 2010-07-01 | Intuitive Surgical, Inc. | Visual force feedback in a minimally invasive surgical procedure |
WO2010085184A1 (en) | 2009-01-20 | 2010-07-29 | Husqvarna Ab | Control system for a remote control work machine |
US20100259614A1 (en) | 2009-04-14 | 2010-10-14 | Honeywell International Inc. | Delay Compensated Feature Target System |
US20110046781A1 (en) | 2009-08-21 | 2011-02-24 | Harris Corporation, Corporation Of The State Of Delaware | Coordinated action robotic system and related methods |
US8473101B2 (en) | 2009-08-21 | 2013-06-25 | Harris Corporation | Coordinated action robotic system and related methods |
US8950286B2 (en) | 2009-10-02 | 2015-02-10 | Commissariat à l'énergie atomique et aux énergies alternatives | Robot or haptic interface structure with parallel arms |
US20110144828A1 (en) | 2009-12-11 | 2011-06-16 | The Boeing Company | Unmanned Multi-Purpose Ground Vehicle with Different Levels of Control |
WO2011075093A1 (en) | 2009-12-14 | 2011-06-23 | Gokhan Vargin Gok | A multi axis robot |
US20110155785A1 (en) | 2009-12-24 | 2011-06-30 | Ethicon Endo-Surgery, Inc. | Motor-driven surgical cutting instrument with electric actuator directional control assembly |
US20130328770A1 (en) | 2010-02-23 | 2013-12-12 | Muv Interactive Ltd. | System for projecting content to a display surface having user-controlled size, shape and location/direction and apparatus and methods useful in conjunction therewith |
US20120095619A1 (en) * | 2010-05-11 | 2012-04-19 | Irobot Corporation | Remote Vehicle Missions and Systems for Supporting Remote Vehicle Missions |
US20130090194A1 (en) | 2010-06-17 | 2013-04-11 | Commissariat A L'energie Atomique Et Aux Energies Alternatives | Reducing device having a high reduction ratio, robot and haptic interface comprising at least one such reducing device |
US20120150351A1 (en) | 2010-12-09 | 2012-06-14 | Harris Corporation | Ball joint having a passageway for routing a cable therethrough |
US20120185099A1 (en) | 2011-01-19 | 2012-07-19 | Harris Corporation | Telematic interface with control signal scaling based on force sensor feedback |
US20120185098A1 (en) | 2011-01-19 | 2012-07-19 | Harris Corporation | Telematic interface with directional translation |
US20140031983A1 (en) | 2011-03-23 | 2014-01-30 | Sri International | Dexterous telemanipulator system |
US20120294696A1 (en) | 2011-05-20 | 2012-11-22 | Harris Corporation | Haptic device for manipulator and vehicle control |
US20120306741A1 (en) | 2011-06-06 | 2012-12-06 | Gupta Kalyan M | System and Method for Enhancing Locative Response Abilities of Autonomous and Semi-Autonomous Agents |
Non-Patent Citations (44)
Title |
---|
Alqasemi R et al: "Kinematics, control and redundancy resolution of a 9-DoF wheelchair-mounted robotic arm system for ADL tasks",Mechatronics and Its Applications, 2009. ISMA '09. 6th International Symposium on, IEEE, Piscataway, NJ, USA, Mar. 23, 2009 (Mar. 23, 2009), pp. 1-7. |
Alqasemi, R., et al., "Maximizing Manipulation Capabilities for People with Disabilities Using 9-DoF Wheelchair-Mounted Robotic Arm System", 2007, IEEE. |
Bley F et al: "Supervised navigation and manipulation for impaired wheelchair users", Systems, Man and Cybernetics, 2004 IEEE International Conference on, IEEE, Piscataway, NJ, USA, vol. 3, Oct. 10, 2004 (Oct. 10, 2004), pp. 2790-2796. |
Cheung, Y., et al., "Cooperative Control of a Multi-Arm System Using Semi-Autonomous Telemanipulations and Adaptive Impedance", Adanced Robotis, 2009. ICAR 2009. International Conference on, IEEE, Piscataway, NJ, USA, Jun. 22, 2009, pp. 1-7. |
European Search Report mailed Mar. 14, 2012, Application Serial No. 11009319.2-2316, in the name of Harris Corporation. |
Everett L J et al; "Automatic Singularity Avoidance Using Joint Variations in Robot Task Modification", IEEE Robotics & Automation Magazine, IEEE Service Center, Piscataway, NJ, US, vol. 1, No. 3, Sep. 1, 1994, pp. 13-19, XP011420425. |
HAMID ABDI ; SAEID NAHAVANDI: "Joint velocity redistribution for fault tolerant manipulators", ROBOTICS AUTOMATION AND MECHATRONICS (RAM), 2010 IEEE CONFERENCE ON, IEEE, PISCATAWAY, NJ, USA, 28 June 2010 (2010-06-28), Piscataway, NJ, USA, pages 492 - 497, XP031710198, ISBN: 978-1-4244-6503-3 |
Hamid Abdi et al: "Joint Velocity Redistribution for Fault Tolerant Manipulators", Robotics Automation and Mechatronics (RAM), 2010 IEEE Conference on, IEEE, Piscataway, NJ, USA, Jun. 28, 2010, pp. 492-497, XP031710198. |
Information about Related Patents and Patent Applications, see section 6 of the accompanying Information Disclosure Statement Letter, which concerns Related Patents and Patent Applications (Feb. 12, 2015). |
Information about Related Patents and Patent Applications, see section 6 of the accompanying Information Disclosure Statement Letter, which concerns Related Patents and Patent Applications. |
International Search Report dated Jan. 15, 2013 (Jan. 15, 2013), Application Serial No. PCT/US2012/037751 in the name of Harris Corporation. |
International Search Report dated Oct. 29, 2012; Application Serial No. PCT/US2012/034207 in the name of Harris Corporation. |
International Search Report mailed Jan. 4, 2013, International Application Serial No. PCT/US2012/058303 in the name of Harris Corporation. |
International Search Report mailed Jun. 28, 2012, Application Serial No. PCT/US2012/027475 in the name of Harris Corporation. |
International Search Report mailed May 12, 2014, Applicaiton Serial No. PCT/US2013/069071, in the name of Harris Corporation. |
International Search Report mailed May 2, 2013, International Application No. PCT/US2012/051314, in the name of Harris Corporation. |
International Search Report mailed May 23, 2012; Application Serial No. PCT/US2011/066873 in the name of Harris Corporation. |
Jonghoon Park et al.: "Reconstruction of Inverse Kinematic Solution Subject to Joint Kinematic Limits Using Kinematic Redundancy", Intelligent Robots and Systems '96, IROS 96, Proceedings of the 1996 L EEE/RSJ International Conference on Osaka, Japan, Nov. 4-8, 1996, New York, NY, USA, IEEE, US, vol. 2, 4, Nov. 1996, pp. 425-430, XP010212433. |
JONGHOON PARK, WAN-KYUN CHUNG, YOUNGIL YOUM, MOONSANG KIM: "Reconstruction of inverse kinematic solution subject to joint kinematic limits using kinematic redundancy", INTELLIGENT ROBOTS AND SYSTEMS '96, IROS 96, PROCEEDINGS OF THE 1996 L EEE/RSJ INTERNATIONAL CONFERENCE ON OSAKA, JAPAN 4-8 NOV. 1996, NEW YORK, NY, USA,IEEE, US, vol. 2, 4 November 1996 (1996-11-04) - 8 November 1996 (1996-11-08), US, pages 425 - 430, XP010212433, ISBN: 978-0-7803-3213-3, DOI: 10.1109/IROS.1996.570808 |
L.J. EVERETT ; J.C. COLSON ; B.W. MOORING: "Automatic singularity avoidance using joint variations in robot task modification", IEEE ROBOTICS & AUTOMATION MAGAZINE., IEEE SERVICE CENTER, PISCATAWAY, NJ., US, vol. 1, no. 3, 1 September 1994 (1994-09-01), US, pages 13 - 19, XP011420425, ISSN: 1070-9932, DOI: 10.1109/100.326724 |
Marshall, W.C., et al., "A Testbed for Design of User-Friendly, Multiple-Degree-Of-Freedom, Manual Controllers", Scientific Honeyweller, Honeywell's Corporate. Minneapolis, US Jan. 1, 1993 (Jan. 1, 1993), pp. 78-86. |
Rocco, Ana Catalina Torres, Development and testing of a new C-based algoithm to control a 9-degree-of-freedom wheelchair-mounted-robotic-arm system, University of South Florida, Jun. 1, 2010. |
Rogers, JE., et al., "Bi-directional Gap Closing MEMS Actuator Using Timing and Control Techniquest", IEEE Industrial Electronics, IECON 2006—32nd Annual Conference on, IEEE, Piscataway, NJ USA Nov. 1, 2006, pp. 3469-3154. |
Suzuki, S., et al., "Performance conditioning of time delayed bilaterial teleoperation system by scaling down compensation value of communication disturbance observer", Advanced Motion Control, 2010, 11th IEEE International Conference on, IEEE, Piscataway, NJ, USA, Mar. 12, 2010, pp. 524-529. |
Tas, NR, et al., "Technical Note: Design, fabrication and testing of laterally driven electrostatic motors employing walking motion and mechanical leverage", Journal of Micromechanics & Microengineering, Institute of Physics Publishing, Bristol, GB, vol. 13, No. 1, Jan. 1, 2003. N6-N15. |
Tijsma, et al., "A framework of interface improvements for designing new user interfaces for the MANUS robot arm", 2005, IEEE, 9th International Conference on Rehabilitation Robotics, Jul. 28-Jul. 1, 2005, Chicago, IL, USA. |
Tijsma, H.A. et al., A Framework of Interface Improvements for Designing New User Interfaces for the MANUS Robot Arm, Proceedings of the 2005 IEEE, 2005, 235-240. |
Torres Rocco, A.C., "Development and testing of a new C-based algorithm to control 9-degree-of-freedom wheelchair-mounted-robotic-arm system". Jun. 1, 2010, Univ. of So. Florida. |
Tsumaki Y et al: "Design of a compact 6-DOF haptic interface", Robotics and Automation, 1998. Proceedings. 1998 IEEE International Conference on Leuven, Belgium May 16-20, 1998, New York, NY, USA, IEEE, US, vol. 3, May 16, 1998 (May 16, 1998), pp. 2580-2585 |
Tzafestas, C., et al., "Adaptive impedance control in haptic teleoperation to improve transparency under time-delay", 2008 IEEE International Conference on Robotics and Automation. The Half-Day Workshop on: Towards Autonomous Agriculture of Tomorrow, IEEE-Piscataway, NJ, USA, Piscataway, NJ, USA, May 19, 2008, pp. 212-219. |
U.S. Appl. No. 12/967,267, filed Dec. 14, 2010, Haptic Interface Handle With Force-Indicating Trigger Mechanism. |
U.S. Appl. No. 13/009,019, filed Jan. 19, 2011, Telematic Interface With Directional Translation. |
U.S. Appl. No. 13/009,040, filed Jan. 19, 2011, Telematic Interface With Control Signal Scaling Based on Force Sensor Feedback. |
U.S. Appl. No. 13/053,430, filed Mar. 22, 2011, Manipulator Joint-Limit Handling Algorithm. |
U.S. Appl. No. 13/101,618, filed May 5, 2011, Remote Control Interface. |
U.S. Appl. No. 13/112,366, filed May 20, 2011, Haptic Device for Manipulator and Vehicle Control. |
U.S. Appl. No. 13/211,739, filed Aug. 17, 2011, Haptic Manipulation System for Wheelchairs. |
U.S. Appl. No. 13/267,357, filed Oct. 6, 2011, Improvised Explosive Device Defeat System. |
U.S. Appl. No. 13/627,357, filed Oct. 6, 2011, Improved Explosive Device Defeat System. |
U.S. Appl. No. 13/672,774, filed Nov. 9, 2012, Hybrid Gesture Control Haptic System. |
U.S. Appl. No. 13/761,321, filed Feb. 7, 2013, Systems and Methods for Controlling Movement of Unmanned Vehicles. |
U.S. Appl. No. 14/143,045, filed Dec. 30, 2013, Compact Haptic Interface. |
U.S. Appl. No. 14/494,635, filed Sep. 24, 2014, Telematic Interface With Directional Translation. |
Zarrad, W., et al., "Stability and Transparency Analysis of a Haptic Feedback Controller for Medical Applications", Proceedings of the 46th IEEE Conference on Decision and Control : New Orleans, LA, Dec. 12-14, 2007, IEEE, Piscataway, NJ, USA, Dec. 1, 2007, pp. 5767-5772. |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180029516A1 (en) * | 2016-08-01 | 2018-02-01 | Toyota Motor Engineering & Manufacturing North America, Inc. | Vehicle docking and control systems for robots |
US10562432B2 (en) * | 2016-08-01 | 2020-02-18 | Toyota Motor Engineering & Manufacturing North America, Inc. | Vehicle docking and control systems for robots |
US20200307966A1 (en) * | 2017-12-21 | 2020-10-01 | Hans Kunz GmbH | Crane controller |
US11919750B2 (en) * | 2017-12-21 | 2024-03-05 | Hans Kunz GmbH | Crane controller |
Also Published As
Publication number | Publication date |
---|---|
EP2764323A1 (en) | 2014-08-13 |
US20150345914A1 (en) | 2015-12-03 |
US8996244B2 (en) | 2015-03-31 |
WO2013052419A1 (en) | 2013-04-11 |
EP2764323B1 (en) | 2016-06-22 |
US20130090764A1 (en) | 2013-04-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9638497B2 (en) | Improvised explosive device defeat system | |
US11628072B2 (en) | System and apparatus for robotic device and methods of using thereof | |
US9026250B2 (en) | Haptic manipulation system for wheelchairs | |
EP2709802B1 (en) | Haptic device for manipulator and vehicle control | |
US20130013108A1 (en) | Robotic Agile Lift System With Extremity Control | |
JP6320915B2 (en) | Remote control robot system | |
US11154985B1 (en) | Null space jog control for robotic arm | |
KR20230126711A (en) | Integrated robotic vehicle system and control method | |
US11325246B2 (en) | Grounded SEA actuator | |
US8965620B2 (en) | Systems and methods for controlling movement of unmanned vehicles | |
Tunstel et al. | Recent enhancements to mobile bimanual robotic teleoperation with insight toward improving operator control | |
WO2012149402A2 (en) | Robotic agile lift system with extremity control | |
Ryu et al. | Multi-modal user interface for teleoperation of ROBHAZ-DT2 field robot system | |
Ryu et al. | Wearable haptic-based multi-modal teleoperation of field mobile manipulator for explosive ordnance disposal | |
Kurisu | A study on teleoperation system for a hexapod robot—Development of a prototype platform | |
Kang et al. | ROBHAZ-DT2: Passive double-tracked mobile manipulator for explosive ordnance disposal | |
US20240181656A1 (en) | Robotic end effector |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: HARRIS CORPORATION, FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SUMMER, MATTHEW D.;BOSSCHER, PAUL M.;RUST, JOHN B.;REEL/FRAME:035249/0293 Effective date: 20111005 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |