US11433537B2 - Automatic path generation device - Google Patents

Automatic path generation device Download PDF

Info

Publication number
US11433537B2
US11433537B2 US16/458,193 US201916458193A US11433537B2 US 11433537 B2 US11433537 B2 US 11433537B2 US 201916458193 A US201916458193 A US 201916458193A US 11433537 B2 US11433537 B2 US 11433537B2
Authority
US
United States
Prior art keywords
motion path
path
robot
motion
generation device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active, expires
Application number
US16/458,193
Other languages
English (en)
Other versions
US20200023516A1 (en
Inventor
Seigo Kato
Yoshiharu Nagatsuka
Yuuki KIMOTO
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fanuc Corp
Original Assignee
Fanuc Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fanuc Corp filed Critical Fanuc Corp
Assigned to FANUC CORPORATION reassignment FANUC CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KATO, SEIGO, KIMOTO, YUUKI, NAGATSUKA, YOSHIHARU
Publication of US20200023516A1 publication Critical patent/US20200023516A1/en
Application granted granted Critical
Publication of US11433537B2 publication Critical patent/US11433537B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1656Programme controls characterised by programming, planning systems for manipulators
    • B25J9/1664Programme controls characterised by programming, planning systems for manipulators characterised by motion, path, trajectory planning
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B25HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
    • B25JMANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
    • B25J9/00Programme-controlled manipulators
    • B25J9/16Programme controls
    • B25J9/1628Programme controls characterised by the control loop
    • B25J9/163Programme controls characterised by the control loop learning, adaptive, model based, rule based expert control
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B19/00Programme-control systems
    • G05B19/02Programme-control systems electric
    • G05B19/42Recording and playback systems, i.e. in which the programme is recorded from a cycle of operations, e.g. the cycle of operations being manually controlled, after which this record is played back on the same machine
    • G05B19/425Teaching successive positions by numerical control, i.e. commands being entered to control the positioning servo of the tool head or end effector
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B2219/00Program-control systems
    • G05B2219/30Nc systems
    • G05B2219/40Robotics, robotics mapping to robotics vision
    • G05B2219/40477Plan path independent from obstacles, then correction for obstacles

Definitions

  • the present disclosure relates to an automatic path generation device.
  • the motion path at a time when a robot performing spot welding or arc welding moves is generated by teaching from a worker.
  • the robot's motion path is automatically generated by a motion planning algorithm such as a rapidly exploring random tree (RRT) and based on given data indicating a hit point position on a workpiece and workpiece or jig shape data given as CAD data (for example, Re-publication of PCT International Publication No. 2017/119088 A1).
  • CAD data for example, Re-publication of PCT International Publication No. 2017/119088 A1
  • the robot's movement path is automatically generated is, for example, a motion path for sequentially moving the robot to the hit point position where welding should be performed while avoiding the workpiece or a jig that hinders the robot's motion in the spot welding.
  • the robot's posture at the previous hit point position, the robot's efficient movement during workpiece or jig avoidance, and the load that is applied to the robot's axis (joint), and the like are not considered.
  • a skilled worker manually teaches the robot a motion path in view of various such factors affecting the robot's efficient motion.
  • the manual teaching is burdensome on the worker's part. Besides, the manual teaching is time-consuming, and thus a problem arises in the form of an increase in cycle time. This problem arises in the other robots including handling robots as well as welding robots.
  • An object of the present disclosure is to provide an automatic path generation device automatically generating an efficient motion path for a robot motion.
  • An automatic path generation device of one embodiment of the present disclosure automatically generates a motion path (hereinafter, referred to as temporary motion path) based on the position of a motion point where a robot performs any motion (such as a hit point position in spot welding) and the shape of an interference object such as a workpiece and a jig and by means of a motion planning algorithm such as RRT.
  • the automatic path generation device has a machine learning device. A skilled worker manually creates a motion path (hereinafter, referred to as actual motion path). The machine learning device learns the correlation between the temporary motion path and the actual motion path based on the actual motion path and the automatically generated motion path.
  • the motion path manually created by the skilled worker is a motion path that the skilled worker creates based on the position of a motion point and the shape of an interference object identical to the position of the motion point and the shape of the interference object that the motion planning algorithm uses in order to generate the motion path.
  • the automatic path generation device estimates the actual motion path from the temporary motion path by using the machine learning device that has learned the correlation between the temporary motion path and the actual motion path.
  • the temporary motion path is automatically generated by the motion planning algorithm and is not ideal as a motion path.
  • the machine learning device learns the difference between the temporary motion path and the actual motion path and automatically estimates the actual motion path from the temporary motion path.
  • the automatic path generation device is capable of automatically deriving an efficient motion path.
  • An aspect of the present disclosure relates to an automatic path generation device generating a motion path of a robot.
  • the automatic path generation device includes a preprocessing unit creating teacher data based on a temporary motion path which is a motion path between a plurality of motion points where the robot moves and which is automatically generated with a motion planning algorithm and an actual motion path which is a motion path between the motion points and which is created by a skilled worker and a motion path learning unit generating a learned model which has learned a difference between the temporary motion path and the actual motion path with teacher data created by the preprocessing unit.
  • the automatic path generation device includes a learning model storage unit storing a learned model which has learned a difference between a temporary motion path which is a motion path between a plurality of motion points where the robot moves and which is automatically generated with a motion planning algorithm and an actual motion path which is a motion path between the motion points and which is created by a skilled worker and a motion path estimation unit estimating an actual motion path of the robot based on a temporary motion path of the robot automatically generated with a motion planning algorithm and a learned model stored in the learning model storage unit.
  • FIG. 1 is a schematic hardware configuration diagram of an automatic path generation device according to one embodiment
  • FIG. 2 is a schematic functional block diagram of an automatic path generation device according to a first embodiment
  • FIG. 3 is a diagram exemplifying a robot motion path
  • FIG. 4 is a schematic functional block diagram of an automatic path generation device according to another modification example.
  • FIG. 5 is a schematic functional block diagram of an automatic path generation device according to another modification example.
  • FIG. 6 is a diagram exemplifying a robot motion path from an actual motion path
  • FIG. 7 is a schematic functional block diagram of an automatic path generation device according to a second embodiment.
  • FIG. 8 is a schematic functional block diagram of an automatic path generation device according to a third embodiment.
  • An automatic path generation device will be described below as a device generating a robot's motion path in spot welding.
  • the position of the robot's motion point is, for example, a hit point position in the spot welding and interference objects include a workpiece and a jig.
  • FIG. 1 is a schematic hardware configuration diagram illustrating the automatic path generation device according to one embodiment.
  • An automatic path generation device 1 is implemented in, for example, a controller controlling the robot.
  • the automatic path generation device 1 is implemented in a personal computer put aside the robot and the robot controller, and a computer such as a cell computer, a host computer, an edge server, and a cloud server connected to the robot controller by means of a wired/wireless network.
  • the automatic path generation device 1 is implemented in the controller controlling the robot.
  • the automatic path generation device 1 of the present embodiment has a function used for general robot control.
  • the automatic path generation device 1 is connected to a robot 2 via an interface 19 and controls the robot 2 .
  • the robot 2 has at least one link (movable portion) and at least one joint.
  • the robot 2 is, for example, a six-axis articulated type robot.
  • the robot 2 includes a tool such as a C gun, an X gun, and a laser for performing welding such as spot welding, arc welding, and laser welding.
  • the robot 2 may grip the workpiece and weld the workpiece by changing the position of the gripped workpiece with respect to the welding tool fixed at a predetermined position.
  • the automatic path generation device 1 includes a machine learning device 100 in which the motion path of the robot 2 during the welding is machine-learned in advance.
  • the automatic path generation device 1 controls the robot 2 and performs the workpiece welding in accordance with the result of estimation of the optimum motion path of the robot 2 output from the machine learning device 100 based on the result of the machine learning.
  • the detailed configuration of the robot 2 is already known, and thus will not be described in detail in the present specification.
  • a CPU 11 of the automatic path generation device 1 is a processor controlling the automatic path generation device 1 as a whole.
  • the CPU 11 reads the system program that is stored in a ROM 12 via a bus 20 .
  • the CPU 11 controls the entire automatic path generation device 1 in accordance with the read system program.
  • a RAM 13 temporarily stores temporary calculation and display data, various data input by a worker via an input device 71 , and the like.
  • a memory, an SSD, or the like backed up by a battery (not illustrated) constitutes a nonvolatile memory 14 . Accordingly, the storage state of the nonvolatile memory 14 is maintained even after the automatic path generation device 1 is turned off.
  • the nonvolatile memory 14 has a setting area in which setting information related to the operation of the automatic path generation device 1 is stored.
  • the nonvolatile memory 14 stores, for example, CAD data or a control program for the robot 2 input from the input device 71 and CAD data or a control program for the robot 2 read from an external storage device (not illustrated).
  • the program and various data stored in the nonvolatile memory 14 may be loaded in the RAM 13 during execution/use.
  • Various system programs including a system program for controlling communication with the machine learning device 100 (described later)
  • a known analysis program are pre-written in the ROM 12 .
  • a display device 70 displays, for example, each data read into the memory and data obtained as a result of program execution. Data and the like output from the machine learning device 100 (described later) are input to the display device 70 via an interface 17 and displayed by the display device 70 .
  • a keyboard, a pointing device, and the like constitute the input device 71 .
  • the input device 71 receives data, a command, and the like based on an operation conducted by the worker and delivers the data, the command, and the like to the CPU 11 via an interface 18 .
  • An interface 21 is an interface for connecting the machine learning device 100 with each unit of the automatic path generation device 1 .
  • the machine learning device 100 includes a processor 101 controlling the machine learning device 100 as a whole and a ROM 102 storing a system program and the like.
  • the machine learning device 100 further includes a RAM 103 for performing temporary storage in each processing related to machine learning and a nonvolatile memory 104 used for a learning model or the like to be stored.
  • the machine learning device 100 observes each piece of information (such as the room temperature that is set via the input device 71 and the state quantity of each motor that is acquired from the robot 2 ) that can be acquired by the automatic path generation device 1 via the interface 21 .
  • Each unit of the automatic path generation device 1 acquires a processing result from the machine learning device 100 via the interface 21 .
  • FIG. 2 is a schematic functional block diagram in the learning mode of the machine learning device 100 of the automatic path generation device 1 according to a first embodiment.
  • the function of each functional block illustrated in FIG. 2 is realized by the CPU 11 of the automatic path generation device 1 illustrated in FIG. 1 and the processor 101 of the machine learning device 100 executing the respective system programs of the CPU 11 and the processor 101 and controlling the operation of each unit of the automatic path generation device 1 and the machine learning device 100 .
  • the automatic path generation device 1 includes a control unit 30 , a preprocessing unit 34 , and a motion path learning unit 38 .
  • the control unit 30 controls the robot 2 .
  • the preprocessing unit 34 creates teacher data used for the machine learning that is executed by the machine learning device 100 .
  • the teacher data is created based on a control program 50 including information indicating a motion path created by teaching from the worker and CAD data 52 including shape information on the workpiece and the jig to be welded.
  • the motion path learning unit 38 learns the motion path of the robot 2 by using the teacher data created by the preprocessing unit 34 .
  • the control unit 30 controls the robot 2 based on the control operation that the worker performs on an operation board (not illustrated), the control program that is stored in the nonvolatile memory 14 or the like, the motion path that is output from the machine learning device 100 , or the like.
  • the control unit 30 has a function for general control required for each unit of the robot 2 to be controlled. In a case where each axis (joint) of the robot 2 is moved, for example, the control unit 30 outputs command data indicating the amount of change in axis angle for each control cycle to the motor that drives the axis.
  • the control unit 30 acquires the motor state quantities (such as the current value, the position, the velocity, the acceleration, and the torque) of each motor of the robot 2 and uses the acquired motor state quantities in controlling the robot 2 .
  • the preprocessing unit 34 is functional means for creating teacher data T used for supervised learning, which is a machine learning method, based on the control program 50 and the CAD data 52 and outputting the teacher data T to the machine learning device 100 .
  • the control program 50 includes information indicating a motion path created by teaching from a skilled worker.
  • the CAD data 52 includes the shape information on the workpiece and the jig to be welded.
  • the preprocessing unit 34 calculates the motion path of the robot 2 (hereinafter, referred to as temporary motion path) by using a known motion planning algorithm (such as an RRT) and based on the spot welding hit point position information extracted from the control program 50 and the workpiece and jig position information included in the CAD data 52 .
  • a known motion planning algorithm such as an RRT
  • the preprocessing unit 34 creates the teacher data and outputs the teacher data to the machine learning device.
  • data related to the calculated motion path is input data and output data is data related to the motion path extracted from the control program 50 and created by teaching from the worker (hereinafter, referred to as actual motion path).
  • Each of the temporary motion path and the actual motion path is, for example, hit point-related information, information related to an inter-hit point midpoint, and time series data on the operation parameters of the robot 2 between the hit point and the midpoint.
  • the motion parameters are, for example, time series data on the motion parameters indicating the velocity, the acceleration, the smoothness, and the like of each axis.
  • Each of the temporary motion path and the actual motion path is hit point-related information, information related to an inter-hit point midpoint, and time series data on the motion parameters of the tool gripped by the robot 2 between the hit point and the midpoint.
  • the motion parameters in this case may be, for example, time series data on the motion parameters indicating, for example, the velocity, the acceleration, and the smoothness of the position of the tool gripped by the robot 2 .
  • FIG. 3 is a diagram illustrating an example of the motion path for moving between the hit point positions of the workpiece.
  • the welding tool is attached to the hand of the robot 2 .
  • the welding tool performs welding on hit point positions P 1 and P 2 of the workpiece under the control of the control unit 30 .
  • FIG. 3 illustrates a motion path in which the welding tool sequentially moves to the hit point position P 1 , a midpoint P 1 - 1 , a midpoint P 1 - 2 , and the hit point position P 2 .
  • the motion path from the hit point position P 1 to the hit point position P 2 is defined by the position of each axis of the robot 2 (or the position of the tool) at each hit point and midpoint position, the velocity, the acceleration, and the smoothness of each axis during the movement from the hit point position P 1 to the midpoint P 1 - 1 , the velocity, the acceleration, and the smoothness of each axis during the movement from the midpoint P 1 - 1 to the midpoint P 1 - 2 , and the velocity, the acceleration, and the smoothness of each axis during the movement from the midpoint P 1 - 2 to the hit point position P 1 - 2 .
  • Such time series data is created with regard to the temporary motion path and the actual motion path.
  • the teacher data T is created by the temporary motion path being used as input data and the actual motion path being used as output data.
  • the preprocessing unit 34 may create the single teacher data T from the entire motion path (temporary and actual motion paths) of the robot 2 .
  • the preprocessing unit 34 may create single teacher data Ti from the partial motion path between two hit point positions Pi and Pj in the entire motion path (temporary and actual motion paths) and create a plurality of teacher data T 1 to Tn (n being a positive integer) respectively corresponding to the partial motion path from the entire motion path of the robot 2 .
  • the preprocessing unit 34 may create teacher data as follows.
  • the preprocessing unit 34 creates teacher data Tj in which the partial motion path between four hit point positions Pi to P 1 in the entire temporary motion path is single input data and the partial motion path between Pj and Pk in the actual motion path is single output data corresponding to the input data.
  • the preprocessing unit 34 creates the plurality of teacher data T 1 to Tn (n being a positive integer) respectively corresponding to the partial motion path from the entire motion path of the robot 2 (In a case where the teacher data is created in this manner, a predetermined fixed value such as 0 may be defined for partial input data that cannot be defined from the temporary motion path at the movement initiation and termination positions of the robot 2 ). As described above, the teacher data may be appropriately set in accordance with how the learning model is defined.
  • the motion path learning unit 38 is functional means for performing supervised learning using the teacher data created by the preprocessing unit 34 and generating (learning) a learned model used for estimating the actual motion path from the temporary motion path.
  • the motion path learning unit 38 may use, for example, a neural network as a learning model.
  • the motion path learning unit 38 performs supervised learning in which data related to the temporary motion path included in the teacher data created by the preprocessing unit 34 is input data and data related to the actual motion path is output data.
  • a neural network provided with the three layers of an input layer, an intermediate layer, and an output layer may be used as the learning model.
  • a so-called deep learning method using a neural network of three or more layers may be used as the learning model. Effective learning and inference are performed as a result.
  • the learning that is conducted by the motion path learning unit 38 is performed in a case where the automatic path generation device 1 functions as a learning mode.
  • the automatic path generation device 1 acquires various control programs 50 and CAD data 52 via an external device such as a USB device (not illustrated) or a wired/wireless network (not illustrated).
  • the automatic path generation device 1 performs learning based on the various acquired control programs 50 and CAD data 52 .
  • the automatic path generation device 1 generates the learned model for estimating the actual motion path from the temporary motion path.
  • the learned model generated by the motion path learning unit 38 is stored in a learning model storage unit 46 provided on the nonvolatile memory 104 .
  • the learned model is used for the estimation of the actual motion path by a motion path estimation unit 40 (described later).
  • the automatic path generation device 1 configured as described above generates, based on the control program 50 and the CAD data 52 , the learned model that has learned the actual motion path (motion path created by teaching from the skilled worker) corresponding to the temporary motion path (motion path automatically created by the motion planning algorithm or the like).
  • the control program 50 includes information indicating the motion path created by teaching from the skilled worker.
  • the CAD data includes the shape information on the workpiece and the jig to be welded.
  • the preprocessing unit 34 creates the teacher data T based on the control program 50 including the motion path created by teaching from the skilled worker and a temporary motion path data 54 including a motion path automatically created by a motion path algorithm or the like and outputs the teacher data T to the machine learning device 100 .
  • the worker creates the temporary motion path data 54 based on the hit point position and the CAD data by using an external device or the like and the created temporary motion path data 54 is read by the automatic path generation device 1 along with the control program 50 .
  • Generated as a result is the learned model that has learned the actual motion path (motion path created by teaching from the skilled worker) corresponding to the temporary motion path (motion path automatically created by the motion planning algorithm or the like).
  • the preprocessing unit 34 creates the teacher data T based on the control program 50 including the motion path created by teaching from the skilled worker and outputs the teacher data T to the machine learning device 100 .
  • the preprocessing unit 34 in this case analyzes the actual motion path extracted from the control program 50 and created by teaching from the skilled worker.
  • the preprocessing unit 34 assumes that a virtual interference object is present at the position of an (preset) inside distance a of the motion path.
  • the preprocessing unit 34 calculates the temporary motion path of the robot 2 by using a known motion planning algorithm (such as an RRT) in view of the assumed virtual interference object.
  • the teacher data T is created in which the temporary motion path obtained as described above and the actual motion path are input data and output data, respectively.
  • the teacher data T is output to the machine learning device 100 .
  • the preprocessing unit 34 may be configured as a machine learner that has learned the temporary motion path corresponding to the actual motion path. In this case, the preprocessing unit 34 estimates the temporary motion path based on the actual motion path extracted from the control program 50 .
  • the preprocessing unit 34 creates teacher data in which the estimated temporary motion path is input data and outputs the teacher data to the machine learning device 100 .
  • FIG. 7 is a schematic functional block diagram in the estimation mode of the machine learning device 100 of the automatic path generation device 1 according to a second embodiment.
  • Each functional block illustrated in FIG. 7 is realized by the CPU 11 of the automatic path generation device 1 illustrated in FIG. 1 and the processor 101 of the machine learning device 100 executing the respective system programs of the CPU 11 and the processor 101 and controlling the operation of each unit of the automatic path generation device 1 and the machine learning device 100 .
  • the automatic path generation device 1 of the present embodiment estimates the actual motion path based on the temporary motion path data 54 .
  • the automatic path generation device 1 controls the robot 2 based on the estimated actual motion path.
  • the control unit 30 in the automatic path generation device 1 according to the present embodiment is similar in function to the control unit 30 in the automatic path generation device 1 according to the first embodiment.
  • the motion path estimation unit 40 performs actual motion path estimation using the learned model stored in the learning model storage unit 46 based on the temporary motion path data 54 acquired via an external device such as a USB device (not illustrated) or a wired/wireless network (not illustrated).
  • the learned model is the (parameter-determined) neural network generated by the supervised learning conducted by the motion path learning unit 38 .
  • the motion path estimation unit 40 estimates (calculates) the actual motion path by inputting the temporary motion path data 54 as input data to the learned model.
  • the actual motion path estimated by the motion path estimation unit 40 is output to the control unit 30 and used for the robot 2 to be controlled.
  • the actual motion path estimated by the motion path estimation unit 40 may be used after, for example, display output to the display device 70 or transmission output to a host computer, a cloud computer, or the like via a wired/wireless network (not illustrated).
  • the automatic path generation device 1 of the present embodiment configured as described above performs motion path learning based on a plurality of teacher data obtained by the robot 2 performing motions of various patterns.
  • the automatic path generation device 1 estimates an efficient motion path of the robot 2 by using a learned model in which a sufficient learning result is obtained.
  • FIG. 8 is a schematic functional block diagram in the estimation mode of the machine learning device 100 of the automatic path generation device 1 according to a third embodiment.
  • the function of each functional block illustrated in FIG. 8 is realized by the CPU 11 of the automatic path generation device 1 illustrated in FIG. 1 and the processor 101 of the machine learning device 100 executing the respective system programs of the CPU 11 and the processor 101 and controlling the operation of each unit of the automatic path generation device 1 and the machine learning device 100 .
  • the control program 50 , the CAD data 52 , the temporary motion path data 54 , and the like are not illustrated in FIG. 8 .
  • the automatic path generation device 1 of the present embodiment includes a simulation unit 42 simulating the motion of the robot 2 .
  • the simulation unit 42 performs the simulation using the actual motion path that the motion path estimation unit 40 estimates based on the temporary motion path data 54 in the estimation mode.
  • interference may occur between the robot 2 or the welding tool and the workpiece or the jig on the estimated actual motion path.
  • a motion path correction unit 44 changes the position of the midpoint of the estimated actual motion path or the like to a position where no interference occurs between the robot 2 or the welding tool and the workpiece or the jig (such as a position away by a predetermined distance set in advance from the object of interference).
  • the changed actual motion path is input to the preprocessing unit 34 .
  • the preprocessing unit 34 creates new teacher data T with the input actual motion path and the temporary motion path used for the actual motion path estimation.
  • the new teacher data T is used for relearning with respect to the learned model.
  • the automatic path generation device 1 of the present embodiment corrects the actual motion path such that no interference occurs in a case where interference occurs on the estimated actual motion path.
  • the automatic path generation device 1 constructs a more appropriate learned model by performing relearning by using the corrected actual motion path.
  • the learning and arithmetic algorithms executed by the machine learning device 100 are not limited to those described above and various algorithms can be adopted.
  • the automatic path generation device 1 and the machine learning device 100 are devices having different CPUs.
  • the machine learning device 100 may be realized by the system program stored in the ROM 12 and the CPU 11 of the automatic path generation device 1 .
  • the automatic path generation device 1 estimates the inter-hit point motion path in spot welding.
  • the automatic path generation device 1 is capable of estimating a motion path at an air cut part in arc welding, laser welding, and the like as well.
  • the automatic path generation device 1 is capable of estimating a motion path related to a welding tool movement between preceding and subsequent welding processes.
  • the position of the motion point is a workpiece gripping position, a pre-workpiece transport position, or the like.
  • the interference object is, for example, another device present on a transport path.
  • the automatic path generation device 1 estimates, for example, a motion path until workpiece gripping and a motion path during the workpiece movement to a movement destination that is subsequent to the workpiece gripping.
  • the automatic path generation device 1 learns an efficient actual motion path created by a skilled worker corresponding to the temporary motion path automatically generated by the motion planning algorithm or the like and estimates an efficient actual motion path from the temporary motion path automatically generated by the motion planning algorithm or the like by using the result of the learning.

Landscapes

  • Engineering & Computer Science (AREA)
  • Robotics (AREA)
  • Mechanical Engineering (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Automation & Control Theory (AREA)
  • Manipulator (AREA)
  • Numerical Control (AREA)
US16/458,193 2018-07-17 2019-07-01 Automatic path generation device Active 2040-02-19 US11433537B2 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
JP2018134515A JP6826076B2 (ja) 2018-07-17 2018-07-17 自動経路生成装置
JP2018-134515 2018-07-17
JPJP2018-134515 2018-07-17

Publications (2)

Publication Number Publication Date
US20200023516A1 US20200023516A1 (en) 2020-01-23
US11433537B2 true US11433537B2 (en) 2022-09-06

Family

ID=69148463

Family Applications (1)

Application Number Title Priority Date Filing Date
US16/458,193 Active 2040-02-19 US11433537B2 (en) 2018-07-17 2019-07-01 Automatic path generation device

Country Status (3)

Country Link
US (1) US11433537B2 (ja)
JP (1) JP6826076B2 (ja)
DE (1) DE102019118637B4 (ja)

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6826076B2 (ja) 2018-07-17 2021-02-03 ファナック株式会社 自動経路生成装置
US11170526B2 (en) * 2019-03-26 2021-11-09 Samsung Electronics Co., Ltd. Method and apparatus for estimating tool trajectories
JP7404797B2 (ja) * 2019-11-15 2023-12-26 セイコーエプソン株式会社 ロボットシステムの制御方法
JP6792184B1 (ja) * 2019-11-27 2020-11-25 株式会社安川電機 シミュレーションシステム、シミュレーション方法、シミュレーションプログラム、ロボットの製造方法、およびロボットシステム
US11851052B2 (en) * 2022-01-31 2023-12-26 Ford Global Technologies, Llc Vehicle path verification
WO2023223571A1 (ja) * 2022-05-18 2023-11-23 三菱電機株式会社 工具経路補正装置、工作機械システム、工具経路の補正方法およびプログラム

Citations (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10291178A (ja) 1997-04-18 1998-11-04 Nippon Telegr & Teleph Corp <Ntt> 作業経路追従式ロボットマニピュレータの教示データ修正方法
US20040254677A1 (en) 2003-06-11 2004-12-16 Torgny Brogardh Method for fine tuning of a robot program
US20090125146A1 (en) * 2005-02-25 2009-05-14 Hui Zhang Method of and Apparatus for Automated Path Learning
JP2012157955A (ja) 2011-02-02 2012-08-23 Sony Corp 移動制御装置及び移動制御方法、並びにコンピューター・プログラム
US20150094855A1 (en) * 2012-05-04 2015-04-02 Leoni Cia Cable Systems Sas Imitation learning method for a multi-axis manipulator
US20150165623A1 (en) 2012-07-13 2015-06-18 Fredrik Kange Method For Programming An Industrial Robot In A Virtual Environment
US20150217445A1 (en) * 2014-02-06 2015-08-06 Kuka Laboratories Gmbh Method For Programming An Industrial Robot and Industrial Robot
US20150231785A1 (en) * 2014-02-17 2015-08-20 Fanuc Corporation Robot system for preventing accidental dropping of conveyed objects
US20150239121A1 (en) 2014-02-27 2015-08-27 Fanuc Corporation Robot simulation device for generating motion path of robot
JP2015160253A (ja) 2014-02-26 2015-09-07 キヤノン株式会社 軌道生成方法、ロボット装置、プログラム及び記録媒体
US20170190052A1 (en) * 2014-06-03 2017-07-06 ArtiMinds Robotics GmbH Method and system for programming a robot
WO2017119088A1 (ja) 2016-01-06 2017-07-13 株式会社日立製作所 ロボットシステムおよび制御方法
JP2018020410A (ja) 2016-08-04 2018-02-08 キヤノン株式会社 レイアウト設定方法、およびレイアウト設定装置
US9993923B2 (en) 2012-03-22 2018-06-12 Toyota Jidosha Kabushiki Kaisha Trajectory generation device, moving object, trajectory generation method
US20180222045A1 (en) * 2017-02-09 2018-08-09 X Development Llc Generating a robot control policy from demonstrations collected via kinesthetic teaching of a robot
EP3718707A1 (en) 2017-11-28 2020-10-07 Kawasaki Jukogyo Kabushiki Kaisha Skill transferring machinery

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP6731289B2 (ja) * 2016-06-22 2020-07-29 プライムアースEvエナジー株式会社 電池の製造方法及び電池
JP6826076B2 (ja) 2018-07-17 2021-02-03 ファナック株式会社 自動経路生成装置

Patent Citations (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH10291178A (ja) 1997-04-18 1998-11-04 Nippon Telegr & Teleph Corp <Ntt> 作業経路追従式ロボットマニピュレータの教示データ修正方法
US20040254677A1 (en) 2003-06-11 2004-12-16 Torgny Brogardh Method for fine tuning of a robot program
US6836702B1 (en) * 2003-06-11 2004-12-28 Abb Ab Method for fine tuning of a robot program
EP1631423B1 (en) 2003-06-11 2009-09-02 Abb Ab A method for fine tuning of a robot program
US20090125146A1 (en) * 2005-02-25 2009-05-14 Hui Zhang Method of and Apparatus for Automated Path Learning
JP2012157955A (ja) 2011-02-02 2012-08-23 Sony Corp 移動制御装置及び移動制御方法、並びにコンピューター・プログラム
US9993923B2 (en) 2012-03-22 2018-06-12 Toyota Jidosha Kabushiki Kaisha Trajectory generation device, moving object, trajectory generation method
US20150094855A1 (en) * 2012-05-04 2015-04-02 Leoni Cia Cable Systems Sas Imitation learning method for a multi-axis manipulator
US20150165623A1 (en) 2012-07-13 2015-06-18 Fredrik Kange Method For Programming An Industrial Robot In A Virtual Environment
US20150217445A1 (en) * 2014-02-06 2015-08-06 Kuka Laboratories Gmbh Method For Programming An Industrial Robot and Industrial Robot
US20150231785A1 (en) * 2014-02-17 2015-08-20 Fanuc Corporation Robot system for preventing accidental dropping of conveyed objects
JP2015160253A (ja) 2014-02-26 2015-09-07 キヤノン株式会社 軌道生成方法、ロボット装置、プログラム及び記録媒体
US20150239121A1 (en) 2014-02-27 2015-08-27 Fanuc Corporation Robot simulation device for generating motion path of robot
JP2015160277A (ja) 2014-02-27 2015-09-07 ファナック株式会社 ロボットの動作経路を生成するロボットシミュレーション装置
US20170190052A1 (en) * 2014-06-03 2017-07-06 ArtiMinds Robotics GmbH Method and system for programming a robot
WO2017119088A1 (ja) 2016-01-06 2017-07-13 株式会社日立製作所 ロボットシステムおよび制御方法
US20180250827A1 (en) 2016-01-06 2018-09-06 Hitachi, Ltd. Robot system, and control method
JP2018020410A (ja) 2016-08-04 2018-02-08 キヤノン株式会社 レイアウト設定方法、およびレイアウト設定装置
US20180036882A1 (en) 2016-08-04 2018-02-08 Canon Kabushiki Kaisha Layout setting method and layout setting apparatus
US20180222045A1 (en) * 2017-02-09 2018-08-09 X Development Llc Generating a robot control policy from demonstrations collected via kinesthetic teaching of a robot
EP3718707A1 (en) 2017-11-28 2020-10-07 Kawasaki Jukogyo Kabushiki Kaisha Skill transferring machinery
JPWO2019107455A1 (ja) 2017-11-28 2020-11-19 川崎重工業株式会社 技能伝承機械装置

Also Published As

Publication number Publication date
JP2020011328A (ja) 2020-01-23
DE102019118637B4 (de) 2022-11-10
US20200023516A1 (en) 2020-01-23
DE102019118637A1 (de) 2020-01-23
CN110722552A (zh) 2020-01-24
JP6826076B2 (ja) 2021-02-03

Similar Documents

Publication Publication Date Title
US11433537B2 (en) Automatic path generation device
CN110198813B (zh) 机器人路径生成装置和机器人系统
CN109002012B (zh) 控制装置以及机器学习装置
US10279476B2 (en) Method and system for programming a robot
US8483878B2 (en) Method and apparatus for predicting interference between target section of robot and peripheral object
US10940585B2 (en) Vibration suppression device
JP6333795B2 (ja) 学習による教示作業の簡易化及び動作性能向上機能を備えたロボットシステム
CN112123341B (zh) 机器人双臂协调运动控制方法、装置和电子设备
CN108687767B (zh) 离线编程装置以及离线编程方法
Rahman Cognitive cyber-physical system (C-CPS) for human-robot collaborative manufacturing
JP7237447B2 (ja) 情報処理方法、プログラム、記録媒体、情報処理装置、ロボットシステム、および物品の製造方法
Agravante et al. Visual servoing for the REEM humanoid robot's upper body
US10507585B2 (en) Robot system that displays speed
Lee et al. A robot teaching framework for a redundant dual arm manipulator with teleoperation from exoskeleton motion data
Rout et al. Optimal trajectory generation of an industrial welding robot with kinematic and dynamic constraints
JP7210201B2 (ja) 情報処理方法、プログラム、記録媒体、情報処理装置、ロボットシステム、物品の製造方法
CN110722552B (zh) 自动路径生成装置
JP2020110884A (ja) ロボット制御装置、ロボット制御方法、及びロボット制御プログラム
WO2022137581A1 (ja) シミュレーション装置およびシミュレーションプログラム
Long et al. Robotic meat cutting
JP2020192621A (ja) ロボットに教示点を教示するための装置、ロボットシステム、及びロボットの教示方法
WO2022079764A1 (ja) 干渉チェック装置
US11921492B2 (en) Transfer between tasks in different domains
WO2022044742A1 (ja) ロボット制御装置、ロボット制御システム、及びロボット制御方法
Cui et al. Automated Folding of a Deformable Thin Object through Robot Manipulators

Legal Events

Date Code Title Description
AS Assignment

Owner name: FANUC CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KATO, SEIGO;NAGATSUKA, YOSHIHARU;KIMOTO, YUUKI;REEL/FRAME:049637/0864

Effective date: 20190418

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

STPP Information on status: patent application and granting procedure in general

Free format text: APPLICATION DISPATCHED FROM PREEXAM, NOT YET DOCKETED

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS

STPP Information on status: patent application and granting procedure in general

Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID

STPP Information on status: patent application and granting procedure in general

Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED

STCF Information on status: patent grant

Free format text: PATENTED CASE