WO2005039836A2 - Verfahren zur einrichtung einer bewegung eines handhabungsgeräts und bildverarbeitung - Google Patents
Verfahren zur einrichtung einer bewegung eines handhabungsgeräts und bildverarbeitung Download PDFInfo
- Publication number
- WO2005039836A2 WO2005039836A2 PCT/EP2004/011863 EP2004011863W WO2005039836A2 WO 2005039836 A2 WO2005039836 A2 WO 2005039836A2 EP 2004011863 W EP2004011863 W EP 2004011863W WO 2005039836 A2 WO2005039836 A2 WO 2005039836A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- movement
- handling device
- image processing
- control
- sequence
- Prior art date
Links
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1694—Programme controls characterised by use of sensors other than normal servo-feedback from position, speed or acceleration sensors, perception control, multi-sensor controlled systems, sensor fusion
- B25J9/1697—Vision controlled systems
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/36—Nc in input of data, input key till input tape
- G05B2219/36412—Fine, autonomous movement of end effector by using camera
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/37—Measurements
- G05B2219/37555—Camera detects orientation, position workpiece, points of workpiece
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/39—Robotics, robotics to robotics hand
- G05B2219/39387—Reflex control, follow movement, track face, work, hand, visual servoing
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/39—Robotics, robotics to robotics hand
- G05B2219/39391—Visual servoing, track end effector with camera image feedback
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/40—Robotics, robotics mapping to robotics vision
- G05B2219/40546—Motion of object
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/40—Robotics, robotics mapping to robotics vision
- G05B2219/40555—Orientation and distance
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/40—Robotics, robotics mapping to robotics vision
- G05B2219/40604—Two camera, global vision camera, end effector neighbourhood vision camera
Definitions
- the invention relates to a method for setting up a movement of a handling device with in particular a plurality of movable axes and a control unit, wherein position, time and speed can be specified for each axis. Freedom of movement around at least three axes is advantageously possible in order to enable free arrangement in space. If only one movement in one plane is required, adjustment options around two axes are sufficient. Depending on the task of the handling device, however, more axes can also be provided, which can be adjusted by corresponding actuators.
- the present invention further relates to a corresponding image processing.
- the handling device can, for example, be a robot, with a robot generally being understood to be a device which can automatically carry out movement and / or work processes.
- the robot has a control system, which issues actuating commands to actuators of the robot so that they execute the movements predefined for them.
- actuating commands to actuators of the robot so that they execute the movements predefined for them.
- the object of the present invention is therefore to propose a simple possibility for setting up the movement of a handling device, with which the movement sequence of the handling device can be flexibly adapted, for example, to the movement of an object to be processed or automatically, i.e. can be changed without external intervention.
- This object is achieved by a method for setting up the movement of a handling device, for example a robot, with at least one by means of a
- control of the handling device or image processing is given an optically recognizable object and a movement sequence related to the object, b) the movement and / or working area of a handling device is recorded with at least one camera,
- the recorded image is evaluated with image processing in such a way that the predetermined object is recognized and its position and / or state of motion is determined in particular relative to the handling device,
- control or the image processing calculates a control command for one or more actuators of the handling device from the position and / or the state of motion of the recognized object and the movement sequence related to the object,
- control system issues a control command in particular to each control element to be moved in accordance with the control command
- an optically recognizable object predefine a specific movement sequence, in particular relative to the object, which is then automatically processed by the control of the handling device, in particular a computer.
- the optically recognizable object is defined by a constellation of optically recognizable features that can be identified by image processing, for example geometric arrangements, specific contrasts and / or other features suitable for recognizing. This makes it possible to visually close the control loop between the in particular moving object and the handling device, ie by means of appropriate image processing, and to provide a to follow the moved object with the handling device without the motion sequence having to be known beforehand and having to be programmed into the control of the handling device.
- a certain movement sequence can thus be specified in an abstract manner.
- the specification of the movement sequence can consist, for example, of tracking a specific object which is moved in a defined or unpredictable (chaotic) manner by the movement of the handling device. It is also possible, for example, to recognize an edge or joint by specifying a specific contrast value and to guide a robot along this joint or edge.
- the control or image processing calculates a control command for one or more actuators of the handling device, so that the abstract movement command by corresponding actuation commands to each actuator actually occurs a movement of the handling device can be implemented.
- the actuating command leads to a movement of the handling device, by means of which either the relative position between the object and the handling device is changed, or the handling device in a constant relative position to the working position. moved object follows. A new relative position, for example due to a movement of the object, is detected again in accordance with the method steps described above and converted into a new control command.
- This type of setting up a movement of a handling device is particularly simple for the user because he does not have to deal with the control program of the handling device or the specification of specific positions to be approached. He can only use the handling device by specifying an object recognizable by image processing and a movement that is abstractly defined in relation to this object. This enables the robot, for example, to automatically track a slot of any length and shape, without having to enter or know the position of this slot. This also leads to a high degree of flexibility in the movement sequence, since the robot can also independently follow new forms of an object, for example an unforeseen deviation in the course of the groove or the like, or an unforeseeable own movement of the object.
- a simple implementation of the method according to the invention provides image processing which, in addition to recognizing the object, also calculates the relative positions and / or movement between the object and the handling device and forwards corresponding information as control commands to the control of the handling device.
- a conventional control for handling devices or robots can then be used, which does not need to be adapted for the use of the method according to the invention. In this case, the control circuit is visually fired by the image processing itself.
- Object itself moves, the location and speed of the object being determined when determining the movement state.
- it makes sense to determine the location and speed of the object relative to the handling device, so that this relative movement can be taken into account particularly easily in the movement sequence to be carried out, which is given abstractly in relation to the object, for example, in its rest coordinate system.
- the object movement is thus determined and overlaid with a known movement of the handling device, or a movement that is determined on the basis of the image processing.
- This also makes it possible for the handling device to carry out work on the moving object, the movement of the object and / or the movement of the handling device not having to be predetermined beforehand.
- the movement sequence of the handling device it is also possible for the movement sequence of the handling device to be predetermined, for example, relative to the object in a control program.
- the method can also be used for simple programming of a handling device for setting up a movement of the handling device or robot, in particular if the handling device is to carry out the same movements again and again.
- the movement sequence is stored in particular with corresponding time information as a result of control commands determined during the execution of the movement. Then the movement of the handling device can take place particularly easily on the basis of this stored sequence of control commands in the desired sequence and at the predetermined time.
- the storage of the control commands in particular in their chronological order, corresponds to the creation of a program for a handling device for controlling its movement, but is much easier to handle than the specification of specific ones Positions or the import of CAD data, on the basis of which the movement is then calculated.
- control command or a sequence of control commands can also depend on the type, position and / or the state of motion of the detected object. This feature can be used, for example, to determine the end of a movement sequence when a certain constellation of optical features reveals a certain object. In addition, this makes it possible, for example in the case of a quality control, to have various handling sequences of a handling device carried out automatically, depending on a known error, in order to make error-dependent corrections.
- the movement of the handling device is checked on the basis of the recorded images.
- this can be used to easily check whether the conditions for carrying out the stored sequence of control commands still exist, for example whether the moving object has been followed correctly. If this is not the case, the movement sequence can be stopped immediately, for example in order to avoid damage to the object.
- Allocate tasks to be performed by the handling device during the movement can be any activity be that can be performed by a handling device-controlled tool. This can include welding, sealing a joint, tracking moving objects or other tasks.
- the tasks can be carried out both during the processing of a stored sequence of control commands as part of a program-controlled handling device movement and during the movement of a handling device based on the currently recognized image data.
- the image can be recorded by a stationary camera or a camera which is moved along with the handling device.
- the stationary camera unit has the entire working and movement area of the handling device in view and . can therefore also handle unforeseen events particularly well capture, e.g. chaotic movements of the object to be tracked.
- the camera unit carried along with the movement of the handling device can be focused on a special work area and offers a higher optical resolution than the stationary camera unit.
- the combination of a stationary and a moving camera is therefore also particularly advantageous, wherein the image processing, for example, allows two or more images to be evaluated simultaneously, in particular also in real time, in order to calculate a control command.
- two, three or more stationary and / or moving cameras can also be provided.
- the method can also be used if objects move or fall out of the field of view of the moving camera in an unexpected way. These objects can then be captured with the stationary camera and the handling device can be guided so that the handling device continues to track this object.
- the method according to the invention for setting up the movement of handling devices considerably simplifies the handling of manipulation devices and their adaptation to specific tasks and activities, because the usually complex programming of a handling device program with one or more predetermined movement sequences is eliminated. This increases the flexibility when using handling devices, such as robots.
- the present invention relates to image processing which is particularly suitable for carrying out the method for setting up a movement of a handling device.
- An image recorded by means of at least one camera is recognized by the image processing, the position of the object is determined spatially and temporally and / or its speed is determined, a relation of the position and / or the speed of the object to the position and / or speed of a handling device is determined and this relation is passed on to the control of the handling device, for example in the form of a control command, in order to track the handling device to the object or to carry out certain tasks or manipulations on the object. This is done in real time if possible and enables the handling device to be controlled on the basis of the visual findings of the image processing.
- the required relationship between the object and the handling device can be formed from the difference between the positions and / or speeds of the object and handling device, in particular in the form of a deviation vector and / or a relative speed vector, which is then fed to the control.
- the difference can be fed directly to the controller, which generates the corresponding control commands from the difference.
- the image processing can convert the differences found into control commands which are fed to the control, which then only generates the specific control commands for the actuators of the handling device.
- the one or more camera (s) can be positioned over the object and tracked when the object moves, the camera movement recorded and this recording converted into movement information for the handling device.
- a movement program for a handling device can be generated in a particularly simple manner by copying an object captured by the image processing to the various positions when it is moved.
- the movement information preferably contains temporal, spatial and / or speed information.
- Fig. 1 shows schematically the implementation of the inventive method for setting up the movement of a handling device with a stationary object
- Fig. 2 shows schematically the implementation of the inventive method for setting up the movement of a handling device for a moving object.
- FIG. 1 shows as a handling device a robot 1 with a plurality of actuators 2 which can be moved about different axes and on which a camera 3 is arranged as a moving sensor.
- a camera 3 is arranged as a moving sensor.
- any tools can also be attached to the robot 1, but these are not shown in FIG. 1.
- the image field 4 of the moving camera 3 is aligned with the object 5.
- a controller 6, which in the example shown is arranged directly on the robot 1, but can also easily be formed separately from it in a computing unit, and / or in an image processing stored in the same or a separate computing unit, there are recognition features for the object 5 and predefine a movement sequence 7 related to the object 5.
- the robot 1 should follow the edge 8 of the object 5 in order to check the edge 8 for defects, for example, or to carry out work on the edge 8 by means of a tool (not shown).
- the image processing of the camera 3 is given features for recognizing the edge, for example a typical contrast curve in the area of the edge 8.
- the camera 3 records the movement or working area of the robot 1 and evaluates the recorded image with the image processing.
- the object 5 is identified, its position is determined relative to the robot 1 and the edge 8 is also recognized, which the handling device is to follow on the basis of the abstractly specified movement sequence 7.
- the controller 6 or the image processing can calculate a control command for the actuators 2 of the robot 1 and issue it accordingly as an actuating command to each actuator 2, so that the handling device 1 follows the edge 8 of the object 5 without the movement sequence having to be permanently programmed in the controller 6 by specifying coordinates.
- the camera 3 takes an image of the object 5 again after each movement of the robot 1 and repeats the above-described method steps. This makes it possible to specify an abstract movement sequence 7, which is related to the object 5 or certain visual features of the object 5 and which the robot 1 automatically follows.
- a stationary camera 9 can also be provided, which has a larger image field 4 than the moving camera 3 and serves to capture the object 5 in an overview.
- the camera 9 is preferably also connected to the image processing of the camera 3 and / or the controller 6 of the robot 1.
- Providing a stationary camera 9 is particularly useful if the object 5 itself is moved, as shown in FIG. 2.
- the direction of movement 10 of the object 5 is indicated in FIG. 2 by arrows.
- the stationary camera 9 serves for a first orientation of the object 5 relative to the robot 1. Because of the larger image field 4 of the stationary camera 9 in comparison to the camera 3 attached to the robot 1, it is easier to find, identify and identify the object 5 to detect unforeseen movement of the object 5, for example slipping on a conveyor belt, quickly and reliably. The precise identification of certain features of the object 5 can then take place with the camera 3 which is also moved.
Landscapes
- Engineering & Computer Science (AREA)
- Robotics (AREA)
- Mechanical Engineering (AREA)
- Manipulator (AREA)
- Numerical Control (AREA)
Abstract
Description
Claims
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP04790671A EP1675709A2 (de) | 2003-10-20 | 2004-10-20 | VERFAHREN ZUR EINRICHTUNG EINER BEWEGUNG EINES HANDHABUNGSGERÄTS UND BILDVERARBEITUNG |
US10/576,129 US20070216332A1 (en) | 2003-10-20 | 2004-10-20 | Method for Effecting the Movement of a Handling Device and Image Processing Device |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
DE10349221.6 | 2003-10-20 | ||
DE10349221 | 2003-10-20 |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2005039836A2 true WO2005039836A2 (de) | 2005-05-06 |
WO2005039836A3 WO2005039836A3 (de) | 2005-11-24 |
Family
ID=34484924
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/EP2004/011863 WO2005039836A2 (de) | 2003-10-20 | 2004-10-20 | Verfahren zur einrichtung einer bewegung eines handhabungsgeräts und bildverarbeitung |
Country Status (3)
Country | Link |
---|---|
US (1) | US20070216332A1 (de) |
EP (1) | EP1675709A2 (de) |
WO (1) | WO2005039836A2 (de) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102007008903A1 (de) * | 2007-02-23 | 2008-08-28 | Abb Technology Ag | Einrichtung zum Steuern eines Roboters |
DE102009058817A1 (de) | 2009-12-18 | 2010-08-05 | Daimler Ag | Anlage und Verfahren zum maßhaltigen Rollfalzen eines Bauteils |
EP2255930A1 (de) * | 2009-05-27 | 2010-12-01 | Leica Geosystems AG | Verfahren und System zum hochpräzisen Positionieren mindestens eines Objekts in eine Endlage im Raum |
EP2602588A1 (de) | 2011-12-06 | 2013-06-12 | Hexagon Technology Center GmbH | Positions- und Ausrichtungsbestimmung in 6-DOF |
WO2016146768A1 (de) * | 2015-03-18 | 2016-09-22 | Kuka Roboter Gmbh | Robotersystem und verfahren zum betrieb eines teleoperativen prozesses |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE10345743A1 (de) * | 2003-10-01 | 2005-05-04 | Kuka Roboter Gmbh | Verfahren und Vorrichtung zum Bestimmen von Position und Orientierung einer Bildempfangseinrichtung |
JP5609760B2 (ja) * | 2011-04-27 | 2014-10-22 | トヨタ自動車株式会社 | ロボット、ロボットの動作方法、及びプログラム |
JP5922932B2 (ja) * | 2012-01-18 | 2016-05-24 | 本田技研工業株式会社 | ロボットティーチング方法 |
DE102015209896B3 (de) * | 2015-05-29 | 2016-08-18 | Kuka Roboter Gmbh | Ermittlung der Roboterachswinkel und Auswahl eines Roboters mit Hilfe einer Kamera |
WO2016203858A1 (ja) * | 2015-06-18 | 2016-12-22 | オリンパス株式会社 | 医療システム |
US9926138B1 (en) * | 2015-09-29 | 2018-03-27 | Amazon Technologies, Inc. | Determination of removal strategies |
JP7467041B2 (ja) | 2018-09-27 | 2024-04-15 | キヤノン株式会社 | 情報処理装置、情報処理方法及びシステム |
JP6898374B2 (ja) | 2019-03-25 | 2021-07-07 | ファナック株式会社 | ロボット装置の動作を調整する動作調整装置およびロボット装置の動作を調整する動作調整方法 |
WO2021028673A1 (en) * | 2019-08-09 | 2021-02-18 | Quantum Leap Technologies Limited | Fabric maintenance sensor system |
US11867630B1 (en) | 2022-08-09 | 2024-01-09 | Glasstech, Inc. | Fixture and method for optical alignment in a system for measuring a surface in contoured glass sheets |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4568816A (en) * | 1983-04-19 | 1986-02-04 | Unimation, Inc. | Method and apparatus for manipulator welding apparatus with improved weld path definition |
US4954762A (en) * | 1989-02-01 | 1990-09-04 | Hitachi, Ltd | Method and apparatus for controlling tracking path of working point of industrial robot |
WO2000045229A1 (en) * | 1999-01-29 | 2000-08-03 | Georgia Tech Research Corporation | Uncalibrated dynamic mechanical system controller |
WO2003064116A2 (en) * | 2002-01-31 | 2003-08-07 | Braintech Canada, Inc. | Method and apparatus for single camera 3d vision guided robotics |
EP1345099A2 (de) * | 2002-03-04 | 2003-09-17 | TECMEDIC GmbH | Verfahren zur Bestimmung der Lage eines Objektes und eines Werkstücks im Raum zur automatischen Montage des Werkstücks am Objekt |
-
2004
- 2004-10-20 WO PCT/EP2004/011863 patent/WO2005039836A2/de not_active Application Discontinuation
- 2004-10-20 US US10/576,129 patent/US20070216332A1/en not_active Abandoned
- 2004-10-20 EP EP04790671A patent/EP1675709A2/de not_active Withdrawn
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4568816A (en) * | 1983-04-19 | 1986-02-04 | Unimation, Inc. | Method and apparatus for manipulator welding apparatus with improved weld path definition |
US4954762A (en) * | 1989-02-01 | 1990-09-04 | Hitachi, Ltd | Method and apparatus for controlling tracking path of working point of industrial robot |
WO2000045229A1 (en) * | 1999-01-29 | 2000-08-03 | Georgia Tech Research Corporation | Uncalibrated dynamic mechanical system controller |
WO2003064116A2 (en) * | 2002-01-31 | 2003-08-07 | Braintech Canada, Inc. | Method and apparatus for single camera 3d vision guided robotics |
EP1345099A2 (de) * | 2002-03-04 | 2003-09-17 | TECMEDIC GmbH | Verfahren zur Bestimmung der Lage eines Objektes und eines Werkstücks im Raum zur automatischen Montage des Werkstücks am Objekt |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE102007008903A1 (de) * | 2007-02-23 | 2008-08-28 | Abb Technology Ag | Einrichtung zum Steuern eines Roboters |
EP2255930A1 (de) * | 2009-05-27 | 2010-12-01 | Leica Geosystems AG | Verfahren und System zum hochpräzisen Positionieren mindestens eines Objekts in eine Endlage im Raum |
WO2010136507A1 (de) * | 2009-05-27 | 2010-12-02 | Leica Geosystems Ag | Verfahren und system zum hochpräzisen positionieren mindestens eines objekts in eine endlage im raum |
AU2010251981B2 (en) * | 2009-05-27 | 2013-08-22 | Leica Geosystems Ag | Method and system for highly precisely positioning at least one object in an end position in space |
US8798794B2 (en) | 2009-05-27 | 2014-08-05 | Leica Geosystems Ag | Method and system for highly precisely positioning at least one object in an end position in space |
DE102009058817A1 (de) | 2009-12-18 | 2010-08-05 | Daimler Ag | Anlage und Verfahren zum maßhaltigen Rollfalzen eines Bauteils |
EP2602588A1 (de) | 2011-12-06 | 2013-06-12 | Hexagon Technology Center GmbH | Positions- und Ausrichtungsbestimmung in 6-DOF |
WO2013083650A1 (en) | 2011-12-06 | 2013-06-13 | Hexagon Technology Center Gmbh | Position and orientation determination in 6-dof |
US9443308B2 (en) | 2011-12-06 | 2016-09-13 | Hexagon Technology Center Gmbh | Position and orientation determination in 6-DOF |
WO2016146768A1 (de) * | 2015-03-18 | 2016-09-22 | Kuka Roboter Gmbh | Robotersystem und verfahren zum betrieb eines teleoperativen prozesses |
Also Published As
Publication number | Publication date |
---|---|
US20070216332A1 (en) | 2007-09-20 |
EP1675709A2 (de) | 2006-07-05 |
WO2005039836A3 (de) | 2005-11-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
DE102018116053B4 (de) | Robotersystem und Roboterlernverfahren | |
DE19930087B4 (de) | Verfahren und Vorrichtung zur Regelung der Vorhalteposition eines Manipulators eines Handhabungsgeräts | |
DE102012104194B4 (de) | Roboter und Punktschweissroboter mit lernender Steuerungsfunktion | |
DE102009034529B4 (de) | Automatisches Führungs- und Erkennungssystem sowie Verfahren für dieses | |
DE102018001026B4 (de) | Robotersystem mit einer lernenden Steuerungsfunktion und lernendes Steuerungsverfahren | |
DE2639774C2 (de) | ||
WO2005039836A2 (de) | Verfahren zur einrichtung einer bewegung eines handhabungsgeräts und bildverarbeitung | |
DE102010023736B4 (de) | Robotersystem mit Problemerkennungsfunktion | |
EP1537009A2 (de) | Verfahren und vorrichtung zur montage mehrerer anbauteile an ein werkst ck | |
DE102014117346B4 (de) | Roboter, Robotersteuerungsverfahren und Robotersteuerungsprogramm zur Werkstückkorrektur | |
DE3317263A1 (de) | Manipulator mit adaptiver geschwindigkeitsgesteuerter bahnbewegung | |
DE102018212531B4 (de) | Artikeltransfervorrichtung | |
DE102008062622A1 (de) | Verfahren und Vorrichtung zur Befehlseingabe in eine Steuerung eines Manipulators | |
DE69837741T2 (de) | Verfahren und system zur steuerung eines roboters | |
DE102015000587A1 (de) | Roboterprogrammiervorrichtung zum Erstellen eines Roboterprogramms zum Aufnehmen eines Bilds eines Werkstücks | |
EP3725472A1 (de) | Verfahren zum ermitteln einer trajektorie eines roboters | |
EP3587044A1 (de) | Verfahren zum greifen von objekten in einem suchbereich, steuereinheit und positioniersystem | |
WO2015055320A1 (de) | Erkennung von gesten eines menschlichen körpers | |
DE102007029398A1 (de) | Verfahren und Vorrichtung zum Programmieren eines Industrieroboters | |
DE102016012227A1 (de) | Verfahren zur automatischen Lagekorrektur eines Roboterarms | |
EP3771952A1 (de) | Verfahren zur automatischen bewegung eines arbeitsgeräts sowie arbeitsgerät | |
DE102015209773B3 (de) | Verfahren zur kontinuierlichen Synchronisation einer Pose eines Manipulators und einer Eingabevorrichtung | |
DE102017005194B3 (de) | Steuern einer Roboteranordnung | |
DE102018124671B4 (de) | Verfahren und Vorrichtung zur Erstellung eines Robotersteuerprogramms | |
DE102005032336B4 (de) | Verfahren zur Beeinflussung einer Steuerung oder zur Steuerung einer Bewegungseinrichtung und Steuerung oder Steuerungskomponente einer Bewegungseinrichtung |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A2 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A2 Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2004790671 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 2004790671 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 10576129 Country of ref document: US Ref document number: 2007216332 Country of ref document: US |
|
WWP | Wipo information: published in national office |
Ref document number: 10576129 Country of ref document: US |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 2004790671 Country of ref document: EP |