WO2021138577A1 - Proxy controller suit with optional dual range kinematics - Google Patents
Proxy controller suit with optional dual range kinematics Download PDFInfo
- Publication number
- WO2021138577A1 WO2021138577A1 PCT/US2020/067693 US2020067693W WO2021138577A1 WO 2021138577 A1 WO2021138577 A1 WO 2021138577A1 US 2020067693 W US2020067693 W US 2020067693W WO 2021138577 A1 WO2021138577 A1 WO 2021138577A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- proxy
- suit
- operator
- controller
- proxy controller
- Prior art date
Links
- 230000009977 dual effect Effects 0.000 title description 3
- 230000033001 locomotion Effects 0.000 claims abstract description 205
- 238000005259 measurement Methods 0.000 claims abstract description 7
- 230000002093 peripheral effect Effects 0.000 claims description 19
- 230000000007 visual effect Effects 0.000 claims description 18
- 230000001133 acceleration Effects 0.000 claims description 8
- 238000009877 rendering Methods 0.000 claims description 2
- 238000000034 method Methods 0.000 description 31
- 210000000245 forearm Anatomy 0.000 description 14
- 230000008569 process Effects 0.000 description 12
- 210000002414 leg Anatomy 0.000 description 8
- 238000010586 diagram Methods 0.000 description 5
- 210000003414 extremity Anatomy 0.000 description 5
- 230000006870 function Effects 0.000 description 5
- 230000009471 action Effects 0.000 description 4
- 230000007613 environmental effect Effects 0.000 description 4
- 238000012545 processing Methods 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000007812 deficiency Effects 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 210000002683 foot Anatomy 0.000 description 2
- 230000003993 interaction Effects 0.000 description 2
- 230000008447 perception Effects 0.000 description 2
- 238000012546 transfer Methods 0.000 description 2
- 210000000689 upper leg Anatomy 0.000 description 2
- 210000000707 wrist Anatomy 0.000 description 2
- 241000282412 Homo Species 0.000 description 1
- 239000000853 adhesive Substances 0.000 description 1
- 230000001070 adhesive effect Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 244000309466 calf Species 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000004140 cleaning Methods 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 210000003811 finger Anatomy 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 210000004247 hand Anatomy 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- 238000002955 isolation Methods 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003278 mimic effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003252 repetitive effect Effects 0.000 description 1
- 230000010076 replication Effects 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 210000003813 thumb Anatomy 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/014—Hand-worn input/output arrangements, e.g. data gloves
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J13/00—Controls for manipulators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/033—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor
- G06F3/0346—Pointing devices displaced or positioned by the user, e.g. mice, trackballs, pens or joysticks; Accessories therefor with detection of the device orientation or free movement in a 3D space, e.g. 3D mice, 6-DOF [six degrees of freedom] pointers using gyroscopes, accelerometers or tilt-sensors
Definitions
- the present invention relates generally to the field of proxy controllers and more particularly, but not by way of limitation, to systems and methods for controlling robot proxies using a controller suit worn by an operator.
- Robots have found successful adoption in addressing narrowly-focused problems where the robots are highly specialized to perform specific tasks. The control systems that allow them to do these tasks are typically tailored to those tasks. An example of this narrowly focused application can be seen in the use of robotic systems for automotive manufacturing. These robots perform discrete, repetitive tasks within a static environment.
- a goal of robotics designers is to build more human-like robots to provide assistance across a broader range of tasks.
- Humanoid robots may more easily navigate an open environment with an infrastructure that has been developed for humanoid interaction.
- Many interfaces and controls have been developed for human interaction, e.g., buttons, doorknobs, etc., and these interfaces are designed for engagement with the unique physical features, proportions and abilities of humans.
- building a humanoid machine (with limbs and torso etc.) is not an impossible task, building the control systems necessary for them to autonomously operate successfully in an unrestricted environment is very difficult. Nonetheless, most robot designers have taken the approach of attempting full automation where the machine attempts to operate in the world with limited assistance from a human guide. Such autonomy is very difficult to achieve as the robot lacks critical thinking and requires advanced programming and sensors.
- a significant problem in remotely controlling robot proxies is ensuring that the operator’s movements are accurately tracked and faithfully transmitted to the robot proxy.
- the sensors may experience “drift” in which the tracking system loses registry with one or more of the sensors worn by the operator and thereafter misinterprets the position of the operator or his appendages.
- the incorrect data resulting from sensor drift is particularly problematic when the robot proxy is instructed to match the misinterpreted position of the operator’s sensors.
- the present invention includes a proxy controller suit for enabling an operator to control an actual or virtual proxy in an actual or virtual remote space.
- the proxy controller suit includes a central tracking unit that is configured to provide the position of the operator in a tracked space.
- the proxy controller suit further includes a first motion sensor attached to a first location on a first appendage of the operator at a known first distance from the central tracking unit, wherein the movement of the first motion sensor is determined with reference to the first motion sensor.
- the proxy controller suit also includes a second motion sensor attached to a second location on the first appendage of the operator at a known distance from the first motion sensor, wherein the movement of the second motion sensor is determined as a function of the measurements made by the first motion sensor.
- the present invention includes a proxy controller suit for enabling an operator to control an actual or virtual proxy within an actual or virtual remote space
- the proxy controller suit has a central tracking unit configured to provide the position of the operator in a tracked space, and a plurality of sensor clusters that each corresponds to an appendage or portion of the operator's body.
- Each of the sensor clusters comprises a plurality of motion sensors. The movement of each of the plurality of motion sensors is determined as a function of the measurements of movement made by other motion sensors in the applicable sensor cluster and with reference to the position of the central tracking unit.
- the present invention includes a proxy controller suit for enabling an operator to control a robot proxy in a remote space, where the proxy controller suit has a central tracking unit located on the operator's torso that is configured to provide the position of the operator in a tracked space.
- the proxy controller suit also includes a plurality of sensor clusters that each correspond to an appendage or portion of the operator's body, and wherein each of the sensor clusters has a plurality of motion sensors.
- the proxy controller suit also includes a suit controller that translates output from the plurality of sensors into command signals delivered to the proxy.
- the suit controller is configured to shift between a conventional kinematic model and an inverse kinematic model for controlling the remote proxy.
- the proxy controller suit further includes a proxy controller that is configured to accept the command signals from the suit controller and apply those command signals to the proxy.
- FIG. 1 depicts the control of a robot proxy by an operator using a proxy controller suit.
- FIG. 2 depicts the operator and proxy controller suit of FIG. 1.
- FIG. 3 depicts the central tracking unit of the proxy controller suit.
- FIG. 4 provides a functional diagram of the central tracking unit and its connection to the suit controller system.
- FIG. 5 provides a depiction of the visor of the proxy controller suit.
- FIG. 6 provides a depiction of a hand grip peripheral controller of the proxy controller suit.
- FIG. 7 depicts the networked arrangement of motion sensors, central tacking unit, visor and controllers of the proxy controller suit of FIG. 2.
- FIG. 8 shows the parent-child relationship between the networked motion sensors within the proxy controller suit of FIG. 2.
- FIG. 9 provides a depiction of a bipedal robot proxy.
- FIG. 10 provides a depiction of a wheeled track robot proxy.
- FIG. 11 provides an exemplar method of controlling the movement of a robot proxy with the proxy controller suit.
- FIG. 12 provides an exemplar method of switching to an inverse kinematic control scheme for close-range control of the robot proxy with a target object.
- FIG. 1 shown therein is a depiction of human operator 200 controlling a robot proxy 202 with a proxy controller suit 100 worn by the human operator 200.
- the operator 200 is located in a tracked local space 204 that is spaced apart by a distance from the remote space 206 where the robot proxy 202 is located.
- the proxy controller suit 100 is a wearable system for enabling the operator 200 to control in real time the movements of an actual or virtual (computer-generated) proxy 202.
- the proxy controller suit 100 is connected to the robot proxy 202 through a private, dedicated or public network 208.
- the network 208 may include a combination of wired, wireless, public, and private resources.
- the distance between the human operator 200 and the robot proxy 202 is only limited by the availability, suitability and span of the network 208. This permits the robot proxy 202 to operate in nearby and distant locations and to operate in environments that are dangerous, challenging or inconvenient for the human operator 200.
- the position and movement of the human operator 200 and proxy controller suit 100 within the local space 204 can be determined and updated in real time using resources internal and external to the proxy controller suit 100.
- the position of the proxy controller suit 100 within the tracked local space 204 can be determined using a global positioning satellite (GPS) system 210, an external camera system 212 or local wireless location systems 214 (including Wi-Fi and Bluetooth), or a combination of these systems.
- the local space 204 includes a treadmill 216 or other mobility sensor that permits the operator 200 to walk or otherwise ambulate in various directions while remaining in a substantially fixed location within the local space 204.
- the output from the proxy controller suit 100 is presented to a suit controller 218, which interprets the signals generated by the proxy controller suit 100 and passes the interpreted signals to a proxy controller 220 through the network 208.
- the suit controller 218 and proxy controller 220 are depicted as standalone modules, it will be appreciated that in some embodiments, the functionality of the suit controller 218 and proxy controller 220 are incorporated into one or more computers that are connected to the proxy 202 and proxy controller suit 100 through local, direct or distributed computer networks.
- FIG. 2 shown therein is a close-up view of the proxy controller suit 100 worn by the human operator 100.
- the proxy controller suit 100 has a central tracking unit 102, a plurality of motion sensors 104, a motion sensor mainboard 106, a visor 108 and one or more peripheral controllers 110.
- the central tracking unit 102 and motion sensors 104 are attached to the operator 200 at designated parts of the operator’s body and are used to accurately track the movements of the operator 200.
- the proxy controller suit 100 utilizes a unique coordination between the central tracking unit 102 that accurately tracks and reports the position of the controller suit 100 within the local space 204, and the various motion sensors 104 that accurately track the movement of the arms, legs, body and head of the operator 200 relative to the position of the central tracking unit 102.
- This novel arrangement of sensors minimizes sensor drift by identifying the position and movement of the controller suit 100 in space using the central tracking unit 102, and then identifying the position and movement of discrete parts of the controller suit 100 (e.g., arms, gloves, legs or shoes) as a part of a skeletal network of the motion sensors 104 that are each referenced directly or indirectly back to the central tracking unit 102.
- the motion sensor mainboard 106 collects the motion data from the motion sensors 104 via a wired or wireless connection or any other sufficient means for sending and collecting data.
- the visor 108 displays information about the proxy controller suit 100 and the proxy environment and collects information from the user such as auditory commands, motion data, or any other data provided by the user or the user’s environment.
- FIG. 1 depicts a particular arrangement of the central tracking unit 102, visor 108, motion sensors 104 and motion sensor mainboard 106 within the controller suit 100, it will be appreciated that alternative arrangements and configurations of these components within the controller suit 100 are also contemplated as falling within the scope of the exemplary embodiments.
- the proxy controller suit 100 may have fewer or additional motion sensors 104 and more than one central tracking unit 102.
- the central tracking unit 102 and motion sensor mainboard 106 may be integrated into a single unit, or combined with other elements of the proxy controller suit 100, such as the visor 108 or peripheral controllers 110.
- proxy controller suit 100 is generally connected to one another with wired (as shown) or wireless connections.
- Data which may include sensor data, environmental data, controller commands, or any other suitable data, is sent between these components over the wired or wireless connections.
- the central tracking unit 102 includes a pair of wide angle (fisheye) cameras 112, stereo image capture lenses 114, an IR projector 116 for illumination in low-light conditions and an RGB sensor 118.
- the central tracking unit 102 may also include an integrated motion sensor 104. As described above, the central tracking unit 102 is used to determine the gross position of the proxy controller suit 100 in the tracked local space 204.
- the central tracking unit 102 determines the position of the proxy controller suit 100 within the local space 204 using “inside-out” technology in which the movement and position of the central tracking unit 102 - and hence the proxy controller suit 100 - is determined using one or more of the wide angle cameras 112, stereo image capture lenses 114, light (infrared or visible) projector 116, RGB sensor 118 and motion sensor 104.
- the motion sensor 104 and the left and right wide angle cameras 112 cooperate to track movement of the central tracking unit 102 within the local space 204 by watching changes in the images taken by the cameras 112 and reading the motion detected by the motion sensor 104.
- the RGB sensor 118 and stereo image capture lenses 114 cooperate to measure the depth of field in front of the central tracking unit 102. This information is processed using within the central tracking unit 102, the mainboard 106 or an external computer to determine the absolute position of the central tracking unit 102 within the local space 204.
- the determination of the position of the central tracking unit 102 is augmented or verified with sensors that are external to the proxy controller suit 100, such as the external camera system 212, a connection with the global positioning satellite (GPS) system 210, and the local wireless location system 214.
- the central tracking unit 102 relies primarily or entirely on an “outside-in” location system in which external sensors, such as the external camera system 212, are used to detect and follow the position of the central tracking unit 102 within the local space 204.
- the central tracking unit 102 may be attached to the operator 200 in various different locations, but the central tracking unit 102 is optimally located in a central location such as the head or torso (as depicted) to serve as a tethering point for the distributed motion sensors 104.
- the visor 108 includes a heads-up display (HUD) 120, and an embedded microphone 122, speakers or earpieces 124.
- the visor 108 may optionally include integrated motion sensor 104.
- the visor 108 is generally configured to display to the operator 200 the video and audio output from the proxy 202, while overlaying or alternating visual information about the condition of the proxy 202 (e.g., battery strength, signal strength, operating temperature, CPU loads, and current state and posture of the proxy 202), the position of the operator 200 in the local space 204.
- the visor 108 can also be configured to accept verbal commands from the operator 200 through the microphone 122.
- the visor 108 can be custom built for integration within the proxy controller suit 100, suitable visors are available from Oculus, Microsoft, HTC and other manufacturers.
- the display includes a first person visual representation of the proxy 202, the remote space 206 around the proxy 202, with an overlay of a virtual representation of the position of the peripheral controllers 110 (or other portions of the proxy controller suit 100).
- the operator 200 gains a two or three-dimensional depiction of the remote space surrounding the proxy 202 and the visible movements of the proxy 202 within the remote space, while also providing visual feedback to the operator 200 through a virtualized overlay of the detected position and correlation of the proxy controller suit 100 as the operator 200 moves within the proxy controller suit 100 to control the proxy 202.
- This system also allows the operator 200 to quickly determine if the movements of the proxy controller suit 100 are no longer accurately aligned and registered with the movements of the proxy 202.
- the peripheral controller 110 is configured as a glove to be worn by the operator 200 to detect and measure movement of the operator’s fingers and thumb.
- the peripheral controller 110 includes an integrated motion sensor 104 to monitor the position and movement of the operator’s hand.
- the peripheral controller 110 is configured with buttons, joysticks or other input devices that are specifically designed for controlling aspects of the robot proxy 202 or implements or interfaces used by the robot proxy 202.
- the peripheral controller 110 can be connected through a wired or wireless connection to the central tracking unit 102 or mainboard 106.
- FIGS. 7 and 8 shown therein are hierarchal representations of the distributed motion sensors 104 within the proxy controller suit 100.
- the motion sensors 104 are arranged in an interlinked, skeletal network and functionally tethered to the central tracking unit 102, either directly or through the motion sensor mainboard 106.
- the central tracking unit 102 acts as a reference point against which the motion signals generated by the motion sensors 104 are based.
- the central tracking unit 102 is essential for accurately tracking the gross position of the proxy controller suit 100 in the tracked local space 204, so that the accurate position of the discrete motion sensors 104 within the proxy controller suit 100 can be tracked without sensor drift errors.
- the central tracking unit 102 provides a more accurate tracking system for the motion sensors 104 by providing a common, centralized positional reference point.
- the motion sensors 104 are configured to detect motion (acceleration and/or rotation).
- the motion sensors 104 are inertial-measurement-units (“IMUs”), which measure changes in acceleration and orientation, which can then be extrapolated mathematically into velocities and movement.
- IMUs inertial-measurement-units
- Suitable motion sensors are commercially available and include the Adafruit 9-DOF Absolute Orientation IMU Fusion Breakout BNO055, and custom sensors with similar properties containing the functionality of a gyroscope and accelerometer.
- conventional IMU sensors are not configured to provide distinct position in a tracking space.
- the motion sensors 104 and other components of the control suit 100 are functionally tethered back to the central tracking unit 102. Because the central tracking unit 102 is tracked within the tracking space and has an absolute position, the accurate position of the central tracking unit 102 can be applied to the other motion sensors, thereby eliminating drift. This process eliminates drift without using Kalman filtering and provides more accuracy in controlling the proxy 202. Thus, by tethering the motion sensors 104 back to the central tracking unit 102, it is only necessary to accurately track the absolute position of the central tracking unit 102 to achieve accurate tracking of all of the motion sensors 104.
- the motion sensors 104 are strategically placed at locations on the operator’s body such that each motion sensor 104 tracks the movement of a different portion of the operator’s skeleton and relates that movement back to the referential central tracking unit 102.
- the motion sensor placed on the operator’s right forearm e.g., “MS11” in FIG. 7 measures acceleration and orientation of the right forearm.
- MS11 right forearm
- the control suit 100 leverages the fixed locational relationships created by the operator’s skeleton as the basis for accurately interpreting the data generated by the motion sensors 104, which can be output as movement of a “virtual” skeleton.
- each motion sensor 104 is measured and analyzed as part of a sensor cluster 126 that generally extends from the central tracking unit 102 on the operator’s torso to the motion sensors 104 located on the operator’s extremities.
- the central tracking unit 102 serves as a root node, with all other motion sensors 104 parented to it in the skeletal sensor clusters 126.
- the parent-child structure of the motions sensors 104 to central tracking unit 102 is shown in FIG. 8, with the central tracking unit 102 being the final parent in the parent-child structure for all segmented groups.
- the rotation of any motion sensor 104 relative to the initial calibration position can be determined by comparing the rotation of the child motion sensor to the rotation (or absence of rotation) of the parent motion sensor.
- the movements of the proxy 202 can be determined and set to match this relative rotation. This is accomplished by the same rotation being implemented at the analogous, corresponding points within the proxy 202
- the proxy controller suit 100 has fifteen (15) motion sensors 104 (labeled MSI through MS 15) and a single central tracking unit 102. As explained, the motion sensors 104 are divided into grouped sensor clusters 126 which are tethered back to the central tracking unit 102 through a data connection. Using these sensor clusters 126, a virtual skeletal structure can be approximated. The approximation is created by calibrating the proxy controller suit 100. During a calibration phase, the control suit 100 is worn by the operator 200 and the distances between the central tracking unit 102 and each distributed motion sensor 104 and the distances between each interlinked motion sensor 104 can be estimated or measured.
- the motion sensors 104 may be affixed to the operator’s body using straps, adhesives, magnets, or other attachment mechanisms. In some embodiments, the various motion sensors 104 are integrated into a suit or clothing worn by the operator. In some embodiments, the motion sensors 104 are placed at the following locations on the operator: upper arm/bicep (MS9 and MS 12); forearm (MS8 and MS11); top of hand above wrist (MS7 and MS 10); mid-thigh quadriceps (MS3 and MS6); shin/calf (MSI and MS5); top of foot (MS2 and MS4); head (MS14); upper torso (MS13); and back/waist (MS 15).
- upper arm/bicep MS9 and MS 12
- forearm MS8 and MS11
- top of hand above wrist MS7 and MS 10
- mid-thigh quadriceps MS3 and MS6
- shin/calf MSI and MS5
- top of foot MS2 and MS4
- head MS14
- the proxy controller suit 100 may be configured with five clusters 126 of motion sensors 104: the sensor cluster 126 having motion sensors 104 placed on a left bicep, forearm, and hand; the second sensor cluster 126 having motion sensors 104 on a right bicep, forearm, and hand; the third sensor cluster 126 having motion sensors 104 on a left foot, leg, thigh and on a point in the back; the fourth sensor cluster 126 having motion sensors 104 on the right foot, leg, thigh and on a point in the back; and the fifth sensor cluster 126 having a motion sensor 104 (or visor 108) placed on the head of the operator 200.
- the sensor cluster 126 having motion sensors 104 placed on a left bicep, forearm, and hand
- the second sensor cluster 126 having motion sensors 104 on a right bicep, forearm, and hand
- the third sensor cluster 126 having motion sensors 104 on a left foot, leg, thigh and on a point in the back
- the distance between motion sensors 104 in a segmented group can be estimated or measured and then adjusted for accuracy. For example, during calibration in the first sensor cluster, the distance between the MS 10 motion sensor (right hand) to MS 11 (right forearm) is measured. Next the distance from MS 11 (right forearm) to MS 12 (right bicep) is measured. Finally, the distance from MS12 (right bicep) to MS13 (central tracking unit 102/torso) is measured. In this way, each motion sensor 104 on the right arm sensor cluster 126 is linked back to the central tracking unit 102 so that each part of the right arm is tracked within the tracking space with reference to the central tracking unit 102.
- the left arm sensor cluster 126 (MS7 through MS9), the left leg sensor cluster 126 (MSI through MS3), and the right leg sensor cluster 126 (MS4 through MS6).
- the left and right leg clusters 126 are tethered to the waist motion sensor (MS 15), which is then tethered back to the central tracking unit 102, thereby creating an additional torso sensor cluster 126 (MS15 and MS13).
- the absolute limb position of the proxy 202 will not match the corresponding limb of the operator 200 unless the length of the proxy’s limbs are proportionally scaled to match the operator’s.
- the size of each segment of the virtual skeleton produced by the control suit 100 can be scaled to match the corresponding dimensions of the proxy 202.
- the calibration process allows the proxy controller suit 100 to be used by operators 200 of different sizes with the same increased accuracy.
- the proxy 202 is generally humanoid, with arms, hands, a torso and a head.
- the proxy 202 may include legs (as depicted in FIG. 9) or a wheeled or tracked mobility system (as depicted in FIG. 10).
- the proxy 202 may also include one or more proxy microphones 224 for detecting sounds from the remote space 206 and a proxy speaker 226 configured to output sounds from the proxy 202.
- the proxy microphones 224 are configured to record multi-channel sounds, which are passed to the proxy controller suit 100 in a manner that permits the operator 200 to perceive sounds in a directional format as recorded by the proxy 202.
- the proxy 202 includes a communications module 228 that is configured to exchange data between the proxy 202 and the proxy controller 220.
- the proxy 202 includes one or more proxy cameras 222 that provide a visual depiction of the field surrounding the proxy 202.
- the visual representations presented by the proxy cameras 222 is standard two-dimensional video that is displayed to the operator 200 in a two- dimensional flat or curved representation within the visor 108.
- the two- dimensional representation may be appropriate or adequate in some applications, the lack of depth perception available to the operator 200 may frustrate efforts to accurately control the proxy 202 in the remote three-dimensional space.
- the proxy cameras 222 are configured to provide a three-dimensional representation of the remote space 206 surrounding the proxy 202.
- the proxy cameras 222 can be configured to render a full three-dimensional virtual reality representation of the area surrounding the proxy 202, which can be displayed in a three-dimensional format within the visor 108 of the proxy controller suit 100. Building an accurate three-dimensional model of the space surrounding the proxy 202 and faithfully passing this model back to the proxy controller suit 100 permits a much more accurate representation of the position, size and scale of objects surrounding the robot proxy 202 while addressing the depth perception problems associated with two- dimensional visual fields.
- the proxy cameras 222 are configured to collect data using depth cameras that provide distance information for each pixel in a captured frame in addition to standard RGB color data.
- the proxy controller 220 and suit controller 218 can be configured to process the three-dimensional image data and construct a point cloud using this data. Every pixel within the image is rendered in three dimensional space as a polygon or shape relative to the orientation of the proxy cameras 222 at the corresponding distance away.
- the resulting three dimensional model provides a real-time representation of objects surrounding the proxy 202, with appropriate depth and scale based on the distance between those objects and the proxy 202.
- the objects can be illustrated with using textures and colors obtained from the proxy cameras 222, or through use of custom shaders that use automated image processing algorithms to apply textures to the three-dimensional representations of objects.
- the three dimensional representations can be passed to the visor 108 as simplified, untextured wireframe meshes. This reduces the overall video processing demands of rendering the information and alleviates constraints on communication bandwidth requirements.
- the wire mesh representations presented to the operator 200 through the proxy controller suit 100 provide accurate distance, location, scale and proportions for three dimensional depictions of the objects surrounding the proxy 202, even if the objects are not displayed to the operator 200 with fully rendered images. Thus, this solution has the effect of bringing the real world surrounding the proxy 202 into “virtual reality” for the operator 200 of the proxy controller suit 100.
- Providing the operator 200 with an accurate sense of depth by viewing the remote space 206 environment in three dimensional virtual reality makes it much easier for the operator 200 to accurately manipulate objects in the remote space 206 with the proxy 202, while also providing improved obstacle avoidance for navigating the proxy 202 through the remote space 206.
- the proxy cameras 288 are configured to capture objects within the field of view of the proxy 202.
- the three dimensional representation of the remote space is constructed in real time by recording the immediate field of view from the proxy 202, with no effort to capture images outside or beyond the present field of view from the proxy 202.
- the proxy cameras 288 are configured to construct are more complete three dimensional representation of the remote space 206 surrounding the proxy 202 (including objects outside the immediate field of view of the proxy 202).
- the full three dimensional representation of the remote space 206 can be constructed by capturing three dimensional images can be stored and indexed by location within the remote space 206.
- the stored images can be retrieved for display to the operator 200 or for comparison and refinement based on new images obtained by the proxy 202.
- This method works well for applications in which the proxy 202 includes a single forward facing depth camera 288.
- the operator 200 is shown the portion of the existing three dimensional model that correlates to the proxy’s current field of view.
- a full three dimensional representation of the remote space 206 surrounding the proxy 202 is generated through the use of an array of strategically placed proxy cameras 288 with depth functionality that scan the entire location surrounding the proxy 202, regardless of the immediate orientation or direction of the proxy 202. This permits the operator 200 to perceive a fully immersive virtual representation of the remote space 206 in real-time, including of objects that might be moving or changing outside the forward-facing view of the proxy 202.
- the proxy control suit 100 When the proxy control suit 100 is used in connection with a full three- dimensional representation of the remote space 206, the operator 200 controls the proxy 202 in the same manner the operator 200 would use the proxy controller suit 100 to control a virtual proxy within a fully virtual environment such as a video game.
- the proxy controller suit 100 is well suited for controlling a remote robot proxy 202, the proxy controller suit 100 is equally well suited for controlling a virtual proxy within a video game or other fully virtual environment, which is contemplated as falling within the scope of exemplary embodiments.
- the proxy controller suit 100 communicates through data networks with the proxy 202 through the suit controller 218 and the proxy controller 220, which are each configured to receive, process and send data between proxy controller suit 100 and the proxy 202.
- the suit controller 218 collects information, such as motion information from the motion sensor mainboard 106, as well as any other information which may be useful in tracking the position and movements of the proxy controller suit 100.
- the suit controller 218 may also receive information from the proxy controller 220, sensors, or other suitable inputs.
- the suit controller 218 can send data to components of the proxy controller suit 100 such as the visor 108 or any other suitable components of the proxy controller suit 100.
- the suit controller 218 implements suit controller software to calculate the absolute position of each body segment using central tracking unit 102 position and all motion sensor data.
- the suit controller 218 can be an external general-purpose computing device such as a PC, tablet, smartphone, or any other sufficient computing device.
- the suit controller 218 may also be mounted on the proxy controller suit 100 as a component attached to the user as a microcontroller, system on chip, micro pc or other small computing form factors.
- the suit controller 218 also performs functions such as calculating information, interpreting, converting, cleaning, normalizing and compressing suit data so that it can be sent in an optimal format to the proxy controller 220, where it will be utilized by the proxy controller software to run commands that actuate motors, sensors and/or devices to control the movements and actions of the proxy 202.
- the proxy controller 220 receives motion commands from the suit controller 218.
- the proxy controller 220 processes the commands which are then sent to the proxy 202 as movement or other action commands. Movement commands from the proxy controller 220 to the proxy 202 may include commands to actuate motors and other devices on the proxy 202.
- the proxy controller 220 may also send environmental data (e.g., temperature, gas composition, etc.) and audio and visual data from the proxy 202 back to the suit controller 218.
- environmental data e.g., temperature, gas composition, etc.
- audio and visual data from the proxy 202 back to the suit controller 218.
- the proxy controller suit 100 may also have additional components, such as controllers, additional sensors, displays, connection components, and user feedback devices.
- the suit controller 218 is also used to perform other functions such as sending and receiving data from components of the suit 100.
- Received data may include flex sensor data sent from peripheral controllers 110, audio data sent by the microphone 122 in the visor 108, or any other data which may be sent of a wired or wireless connection.
- Sent data may include feedback data, based on visual or audio data sensed by the proxy 202 and sent to the suit 100 such as visual, auditory, haptic motor, or environmental data.
- the suit controller 218 may also be used to change suit calibration settings or other proxy controller suit 100 settings by a providing a visual interface allowing user to calibrate the proxy controller suit 100 based on the operator’s size and stature, visor display brightness, volume, power settings, haptic feedback sensitivity, TCP/IP network configuration or any other proxy controller suit setting.
- the suit controller 218 may also be used to send movement information to the proxy that is not based on the motion sensors 104 or the central tracking unit 102. For instance the suit controller 218 may be used to send a disconnect signal to the proxy so that the movements of the motion sensors 104 will no longer move the proxy 202, and subsequent or reconnect signals to begin tracked movement based on the motion sensors 104 and central tracking unit 102 again.
- a position reset signal may also be sent the proxy, which may tell the proxy to assume a preprogrammed position, such as the T-position, for calibration.
- Other preset movements may also be set as commands which may be sent via the suit controller such as jump, run, walk, sit, speak, or any other movement or operation command. These commands may be based on inputs given directly to the suit controller 218, other instructions given by the user, such as gestures or auditory signals given through the visor, or any other sufficient means of inputting data and instruction into the suit controller 218.
- the proxy controller 220 is the physical hardware device that runs the proxy control software.
- the proxy controller 220 is connected to various parts of the proxy 202 which may include a machine or robot that it is responsible for controlling, parts such as local motors, servos, speakers, and sensors.
- the proxy controller 220 maintains connectivity to all these components and takes action to enable, actuate or utilize them when it receives commands from the proxy control software. Many of the commands that that the proxy controller 220 performs will have come from the suit controller 218 software because of motion or actions that are being performed by the operator 200 of the proxy controller suit 100.
- the proxy controller 220 is also used to transfer data back to the suit controller 218 to provide feedback to the operator 200 of the proxy controller suit 100.
- the proxy controller 220 may also be used to perform additional functions such as sending and receiving data to and from the proxy 202 and proxy controller suit 100 input and output devices.
- This data may include audio and visual data from input and output devices on the proxy 202 or proxy controller suit 100, such as screens, microphones, speakers, haptic feedback devices, or any other devices which are capable of providing input and output.
- This data may be sent or received by the various peripherals such as cameras, microphones or any other input devices which are connected to the proxy controller 220 and proxy 202.
- the proxy controller 220 may also be used to modify calibration settings, perform scheduled events, or modify other proxy settings.
- the proxy controller 218 may also be used send environmental data, or any other data regarding the remote space 206. It will be understood that the proxy controller 220 may be able to send and receive other types of data and instructions, which the proxy 202 may receive or sense.
- the motion sensors 104 are connected to the motion sensor mainboard 106, which collects and analyzes information from the motion sensors 104.
- the motion sensor data is sent by the motion sensor mainboard 106 to the suit controller 218 for analysis by the suit controller software.
- the tracked position of the central tracking unit 102 is sensed by the central tracking unit 102 (or by an external sensor) and the tracked position data is sent to the suit controller 218.
- the suit controller 218 determines the absolute positions for all tracked segments of the operator’s body. This data is then sent by the suit controller 218 to the proxy controller 220, which processes and translates the movement data from the control suit 100 into directed movements and commands for the proxy 202.
- the control suit 100 and proxy 202 should be placed in a common position to register the position of the control suit 100 with the state and orientation of the proxy 202. For example, it may be desirable following the scaling process to place the proxy 202 in an initial or default position and then instructing the operator to match the default position with the control suit 100.
- a humanoid proxy 202 is placed in a “T- pose” in which the arms of the proxy 202 are extended outward laterally from the torso at a substantially horizontal position. The operator can then match the “T-pose” while wearing the control suit 100 to register the position and orientation of control suit 100 to the proxy 202.
- FIG. 11 shown therein is an exemplar process for using the proxy controller suit 100 to control a proxy 202.
- FIG. 11 describes the process for using a properly calibrated control suit 100 to command the humanoid proxy 202 to raise its right arm.
- the arm is raised from the T-pose position which may be used for calibration and registration as described above.
- the motion sensors 104 on the right hand (MS 10) and right forearm (MSI 1) will sense the acceleration and rotation of the hand and forearm. These measurements will be compared to the lack of movement measured by the shoulder sensor (MS 12) and torso (MS 13 - central tracking unit 102).
- the motion data is sent to the motion sensor mainboard 106 and to the suit controller 218.
- the motion data of each motion sensor 104 is then used to calculate a relative rotation by the suit controller software.
- the relative rotation is the rotation of the child motion sensors compared against the rotation (or lack of rotation) of the parent motion sensor for each motion sensor 104 in the sensor cluster for the right arm.
- the first child motion sensor would be the hand (MS10) and the first parent would be the forearm (MS11).
- the second child would be the forearm (MS11) and the second parent would be the shoulder (MS 12) and so on for each segmented group of motion sensors until the central tracking unit 102 is the parent.
- step 306 the relative position data is packaged for transfer to the proxy controller 108.
- step 308 the relative position data is then sent to the proxy controller 220, via a wired or wireless connection.
- step 310 the data is received by the proxy controller 220 and unpacked for processing.
- step 312 the position data is analyzed by the proxy controller software and the relative rotations are determined at the same relative points on the proxy 202 as the points the motion sensors 104 are placed at on the user.
- the proxy movements are initiated and the proxy 202 will move in the manner desired by the user based on their movements.
- FIG. 12 shown therein is a method for switching from a conventional kinematics model to an inverse kinematics model for controlling the proxy 202.
- the proxy 202 is instructed to follow the movements of the operator 200, with or without scaling to adjust the magnitude of these movements.
- the conventional forward kinematics model may be undesirable for particular movements that require the precise positioning of the proxy’s terminal linkages (e.g., the hand grip of the proxy 202).
- the suit controller 218 and proxy controller 220 can be programmed to automatically or manually switch to an inverse kinematics model to increase the accuracy of the placement of a part of the proxy 202 at the target location.
- the suit controller 218 determines whether the peripheral controller 110 of the proxy controller suit 100 is within a threshold distance of the target object selected by the operator 200 within the representation of the proxy 202 within the remote space 206 displayed in the visor 108. In this way, the operator 200 can use the peripheral controller 110 to select through the visual representation displayed in the visor 108 a particular target object within the remote space (e.g., a lever) and the proxy controller 220 then determines the distance between the target object and the hand of the proxy 202.
- a particular target object within the remote space e.g., a lever
- the proxy controller 220 and suit controller 218 can be configured to automatically place the suit controller in an inverse kinematics mode.
- the process continues as step 402 when the suit controller 218 is manually placed into the inverse kinematics mode if a grip button or other input feature on the peripheral controller 110 is selected.
- the operator 200 moves or maintains the virtual representation of the peripheral controller 110 on the visual representation of the target object within the display of the visor 108.
- the proxy controller 220 then moves the hand of the proxy 202 so that the virtual representation of the hand of the proxy 202 matches the virtual position of the peripheral controller 110 on the target object.
- the proxy controller 220 and/or the suit controller 218 then calculates a proxy movement solution to actually move the hand of the proxy 202 to the target object in the remote space 206 at step 406.
- the proxy movement solution can be calculated to avoid contact with other objects, interferences or restrictions within the remote space 206.
- the movement of the proxy 202 is determined by matching the desired end location of the proxy’s terminal linkage with the placement of the peripheral controller 110, before backing into a movement solution for the intermediate linkages within the proxy 202.
- the proxy controller suit 100, suit controller 218, proxy controller 220 and proxy 202 cooperate to provide a manually or automatically activated inverse kinematics movement model that is particularly well suited for making fine motor movements of the proxy 202 where the placement and position of the end linkage of the proxy 202 is prioritized over the replication of gross motor movements between the operator 200 and the proxy 202.
- proxy control system refers to the collection of the proxy controller suit 100, suit controller 218, proxy controller 220 and proxy 202.
- Methods of the present invention may be implemented by performing or completing manually, automatically, or a combination thereof, selected steps or tasks.
- the term “method” may refer to manners, means, techniques and procedures for accomplishing a given task including, but not limited to, those manners, means, techniques and procedures either known to, or readily developed from known manners, means, techniques and procedures by practitioners of the art to which the invention belongs.
- the term “at least” followed by a number is used herein to denote the start of a range beginning with that number (which may be a ranger having an upper limit or no upper limit, depending on the variable being defined).
- “at least 1” means 1 or more than 1.
- the term “at most” followed by a number is used herein to denote the end of a range ending with that number (which may be a range having 1 or 0 as its lower limit, or a range having no lower limit, depending upon the variable being defined).
- “at most 4” means 4 or less than 4
- “at most 40%” means 40% or less than 40%.
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Robotics (AREA)
- Mechanical Engineering (AREA)
- Manipulator (AREA)
Abstract
Description
Claims
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA3161710A CA3161710A1 (en) | 2019-12-31 | 2020-12-31 | Proxy controller suit with optional dual range kinematics |
EP20910210.2A EP4084936A4 (en) | 2019-12-31 | 2020-12-31 | Proxy controller suit with optional dual range kinematics |
JP2022555734A JP7428436B2 (en) | 2019-12-31 | 2020-12-31 | Proxy controller suit with arbitrary dual range kinematics |
AU2020417826A AU2020417826A1 (en) | 2019-12-31 | 2020-12-31 | Proxy controller suit with optional dual range kinematics |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201962955735P | 2019-12-31 | 2019-12-31 | |
US62/955,735 | 2019-12-31 | ||
US202063022713P | 2020-05-11 | 2020-05-11 | |
US63/022,713 | 2020-05-11 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021138577A1 true WO2021138577A1 (en) | 2021-07-08 |
Family
ID=76547182
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/US2020/067693 WO2021138577A1 (en) | 2019-12-31 | 2020-12-31 | Proxy controller suit with optional dual range kinematics |
Country Status (6)
Country | Link |
---|---|
US (1) | US11422625B2 (en) |
EP (1) | EP4084936A4 (en) |
JP (1) | JP7428436B2 (en) |
AU (1) | AU2020417826A1 (en) |
CA (1) | CA3161710A1 (en) |
WO (1) | WO2021138577A1 (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11461867B2 (en) * | 2019-09-19 | 2022-10-04 | Sanctuary Cognitive Systems Corporation | Visual interface and communications techniques for use with robots |
EP4290347A4 (en) * | 2021-02-02 | 2024-08-07 | Sony Group Corp | Information processing method, information processing device, and program |
CN116998161A (en) * | 2021-02-05 | 2023-11-03 | 阿里·科德 | Motion capture for performing art |
CN114227679B (en) * | 2021-12-17 | 2023-07-25 | 深圳市金大智能创新科技有限公司 | Remote robot control method and system based on digital virtual person driving |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130211594A1 (en) * | 2012-02-15 | 2013-08-15 | Kenneth Dean Stephens, Jr. | Proxy Robots and Remote Environment Simulator for Their Human Handlers |
US20140302462A1 (en) * | 2012-12-03 | 2014-10-09 | Dynamic Motion Group Gmbh | Motion Simulation System Controller and Associated Methods |
US20170129105A1 (en) * | 2012-03-21 | 2017-05-11 | Kenneth Dean Stephens, Jr. | Replicating the Remote Environment of a Proxy Robot |
US20170192496A1 (en) * | 2015-11-25 | 2017-07-06 | Jakob Balslev | Methods and systems of a motion-capture body suit with wearable body-position sensors |
US20180150131A1 (en) | 2016-11-25 | 2018-05-31 | Sensoryx AG | Wearable motion tracking system |
US20190110847A1 (en) * | 2007-06-13 | 2019-04-18 | Intuitive Surgical Operations, Inc. | Medical robotic system with coupled control modes |
US10429924B1 (en) * | 2019-01-30 | 2019-10-01 | Human Mode, LLC | Virtual reality simulation system |
Family Cites Families (25)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
RU2134193C1 (en) | 1997-04-03 | 1999-08-10 | Псковский государственный педагогический институт им.С.М.Кирова | Method of remote control of anthropomorphic walking and copying robot |
US20090325710A1 (en) * | 2008-06-27 | 2009-12-31 | Microsoft Corporation | Dynamic Selection Of Sensitivity Of Tilt Functionality |
GB0901020D0 (en) | 2009-01-21 | 2009-03-04 | Birmingham City University | A motion capture apparatus |
US9205887B2 (en) * | 2010-02-25 | 2015-12-08 | Honda Motor Co., Ltd. | Constrained resolved acceleration control |
JP2011118924A (en) * | 2011-02-17 | 2011-06-16 | Kyokko Denki Kk | Articulated structure, and wearing tool, system and human machine interface using the same |
WO2014114967A1 (en) * | 2013-01-25 | 2014-07-31 | WENNER, Fabian | Self-calibrating motion capture system |
US9064385B2 (en) | 2013-03-15 | 2015-06-23 | Immersion Corporation | Method and apparatus to generate haptic feedback from video content analysis |
JP2017506169A (en) | 2014-02-20 | 2017-03-02 | マーク オレイニク | Method and system for food preparation in a robot cooking kitchen |
US10203762B2 (en) | 2014-03-11 | 2019-02-12 | Magic Leap, Inc. | Methods and systems for creating virtual and augmented reality |
US9665174B2 (en) | 2015-02-20 | 2017-05-30 | Sony Interactive Entertainment Inc. | Magnetic tracking of glove fingertips with peripheral devices |
CA2882968C (en) | 2015-02-23 | 2023-04-25 | Sulfur Heron Cognitive Systems Inc. | Facilitating generation of autonomous control information |
US9643314B2 (en) | 2015-03-04 | 2017-05-09 | The Johns Hopkins University | Robot control, training and collaboration in an immersive virtual reality environment |
WO2016168117A2 (en) | 2015-04-14 | 2016-10-20 | John James Daniels | Wearable electric, multi-sensory, human/machine, human/human interfaces |
US20170108929A1 (en) | 2015-04-28 | 2017-04-20 | Morgan Walker Sinko | System and Method for Full Motion Capture and Haptic Feedback Suite |
US10667697B2 (en) * | 2015-06-14 | 2020-06-02 | Facense Ltd. | Identification of posture-related syncope using head-mounted sensors |
WO2017120669A1 (en) * | 2016-01-12 | 2017-07-20 | Bigmotion Technologies Inc. | Systems and methods for human body motion capture |
JP2017196678A (en) * | 2016-04-25 | 2017-11-02 | 国立大学法人 千葉大学 | Robot motion control device |
US10322506B2 (en) | 2016-05-06 | 2019-06-18 | Kindred Systems Inc. | Systems, devices, articles, and methods for using trained robots |
EP3252714A1 (en) | 2016-06-03 | 2017-12-06 | Univrses AB | Camera selection in positional tracking |
WO2018097223A1 (en) * | 2016-11-24 | 2018-05-31 | 国立大学法人京都大学 | Robot control system, machine control system, robot control method, machine control method, and recording medium |
US10691121B2 (en) * | 2017-01-30 | 2020-06-23 | Kenneth Dean Stephens, Jr. | Biologically controlled proxy robot |
CN107336233B (en) | 2017-06-02 | 2020-10-09 | 南京邮电大学 | Inertial-kinetic-capture-based human-robot virtual-real interaction control system |
JP6678832B2 (en) * | 2017-09-22 | 2020-04-08 | 三菱電機株式会社 | Remote control manipulator system and control device |
US10777006B2 (en) | 2017-10-23 | 2020-09-15 | Sony Interactive Entertainment Inc. | VR body tracking without external sensors |
WO2019203189A1 (en) * | 2018-04-17 | 2019-10-24 | ソニー株式会社 | Program, information processing device, and information processing method |
-
2020
- 2020-12-31 US US17/139,551 patent/US11422625B2/en active Active
- 2020-12-31 EP EP20910210.2A patent/EP4084936A4/en active Pending
- 2020-12-31 AU AU2020417826A patent/AU2020417826A1/en active Pending
- 2020-12-31 WO PCT/US2020/067693 patent/WO2021138577A1/en unknown
- 2020-12-31 CA CA3161710A patent/CA3161710A1/en active Pending
- 2020-12-31 JP JP2022555734A patent/JP7428436B2/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190110847A1 (en) * | 2007-06-13 | 2019-04-18 | Intuitive Surgical Operations, Inc. | Medical robotic system with coupled control modes |
US20130211594A1 (en) * | 2012-02-15 | 2013-08-15 | Kenneth Dean Stephens, Jr. | Proxy Robots and Remote Environment Simulator for Their Human Handlers |
US20170129105A1 (en) * | 2012-03-21 | 2017-05-11 | Kenneth Dean Stephens, Jr. | Replicating the Remote Environment of a Proxy Robot |
US20140302462A1 (en) * | 2012-12-03 | 2014-10-09 | Dynamic Motion Group Gmbh | Motion Simulation System Controller and Associated Methods |
US20170192496A1 (en) * | 2015-11-25 | 2017-07-06 | Jakob Balslev | Methods and systems of a motion-capture body suit with wearable body-position sensors |
US20180150131A1 (en) | 2016-11-25 | 2018-05-31 | Sensoryx AG | Wearable motion tracking system |
US10429924B1 (en) * | 2019-01-30 | 2019-10-01 | Human Mode, LLC | Virtual reality simulation system |
Non-Patent Citations (3)
Title |
---|
KIM DOOYOUNG; KWON JUNGHAN; HAN SEUNGHYUN; PARK YONG-LAE; JO SUNGHO: "Deep Full-Body Motion Network for a Soft Wearable Motion Sensing Suit", IEEE/ASME TRANSACTIONS ON MECHATRONICS, IEEE SERVICE CENTER, PISCATAWAY, NJ., US, vol. 24, no. 1, 1 February 2019 (2019-02-01), US, pages 56 - 66, XP011709102, ISSN: 1083-4435, DOI: 10.1109/TMECH.2018.2874647 * |
REITER ALEXANDER; MULLER ANDREAS; GATTRINGER HUBERT: "On Higher Order Inverse Kinematics Methods in Time-Optimal Trajectory Planning for Kinematically Redundant Manipulators", IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, IEEE SERVICE CENTER, NEW YORK, NY., US, vol. 14, no. 4, 1 April 2018 (2018-04-01), US, pages 1681 - 1690, XP011680333, ISSN: 1551-3203, DOI: 10.1109/TII.2018.2792002 * |
See also references of EP4084936A4 |
Also Published As
Publication number | Publication date |
---|---|
EP4084936A1 (en) | 2022-11-09 |
CA3161710A1 (en) | 2021-07-08 |
AU2020417826A1 (en) | 2022-06-16 |
US20210200311A1 (en) | 2021-07-01 |
EP4084936A4 (en) | 2024-01-10 |
JP2023507241A (en) | 2023-02-21 |
US11422625B2 (en) | 2022-08-23 |
JP7428436B2 (en) | 2024-02-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11422625B2 (en) | Proxy controller suit with optional dual range kinematics | |
US20210205986A1 (en) | Teleoperating Of Robots With Tasks By Mapping To Human Operator Pose | |
US11498216B2 (en) | Remote control manipulator system and control device | |
Liu et al. | High-fidelity grasping in virtual reality using a glove-based system | |
US20190126484A1 (en) | Dynamic Multi-Sensor and Multi-Robot Interface System | |
Fritsche et al. | First-person tele-operation of a humanoid robot | |
CN108564643B (en) | Performance capture system based on UE engine | |
WO2015180497A1 (en) | Motion collection and feedback method and system based on stereoscopic vision | |
WO2011065034A1 (en) | Method for controlling action of robot, and robot system | |
WO2013149586A1 (en) | Wrist-mounting gesture control system and method | |
JPWO2011080882A1 (en) | Action space presentation device, action space presentation method, and program | |
CN110549353B (en) | Force vision device, robot, and computer-readable medium storing force vision program | |
RU187548U1 (en) | VIRTUAL REALITY GLOVE | |
JP6927937B2 (en) | Systems and methods for generating 3D skeletal representations | |
CN113103230A (en) | Human-computer interaction system and method based on remote operation of treatment robot | |
Taunyazov et al. | A novel low-cost 4-DOF wireless human arm motion tracker | |
JP2011200997A (en) | Teaching device and method for robot | |
KR102456872B1 (en) | System and method for tracking hand motion using strong coupling fusion of image sensor and inertial sensor | |
Lambrecht et al. | Markerless gesture-based motion control and programming of industrial robots | |
RU2670649C9 (en) | Method of manufacturing virtual reality gloves (options) | |
JP2007125670A (en) | Expression action conversion system for robot | |
WO2021073733A1 (en) | Method for controlling a device by a human | |
Jin et al. | Human-robot interaction for assisted object grasping by a wearable robotic object manipulation aid for the blind | |
RU2673406C1 (en) | Method of manufacturing virtual reality glove | |
Bai et al. | Kinect-based hand tracking for first-person-perspective robotic arm teleoperation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20910210 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 3161710 Country of ref document: CA |
|
ENP | Entry into the national phase |
Ref document number: 2020417826 Country of ref document: AU Date of ref document: 20201231 Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 2022555734 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: 2020910210 Country of ref document: EP Effective date: 20220801 |