US20220055877A1 - Control augmentation apparatus and method for automated guided vehicles - Google Patents

Control augmentation apparatus and method for automated guided vehicles Download PDF

Info

Publication number
US20220055877A1
US20220055877A1 US17/462,956 US202117462956A US2022055877A1 US 20220055877 A1 US20220055877 A1 US 20220055877A1 US 202117462956 A US202117462956 A US 202117462956A US 2022055877 A1 US2022055877 A1 US 2022055877A1
Authority
US
United States
Prior art keywords
agv
sensor
sensor data
map
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US17/462,956
Inventor
Ryan Christopher GARIEPY
Andrew Dobson
Jesse Tebbs
Robert Dam
Roydyn Clayton
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Rockwell Automation Technologies Inc
Original Assignee
Clearpath Robotics Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Clearpath Robotics Inc filed Critical Clearpath Robotics Inc
Priority to US17/462,956 priority Critical patent/US20220055877A1/en
Publication of US20220055877A1 publication Critical patent/US20220055877A1/en
Assigned to Clearpath Robotics Inc. reassignment Clearpath Robotics Inc. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: GARIEPY, RYAN CHRISTOPHER, TEBBS, JESSE, CLAYTON, ROYDYN, DAM, ROBERT, DOBSON, ANDREW
Assigned to ROCKWELL AUTOMATION TECHNOLOGIES, INC. reassignment ROCKWELL AUTOMATION TECHNOLOGIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ROCKWELL AUTOMATION, INC.
Assigned to ROCKWELL AUTOMATION, INC. reassignment ROCKWELL AUTOMATION, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CLEARPATH ROBOTICS, INC.
Pending legal-status Critical Current

Links

Images

Classifications

    • BPERFORMING OPERATIONS; TRANSPORTING
    • B66HOISTING; LIFTING; HAULING
    • B66FHOISTING, LIFTING, HAULING OR PUSHING, NOT OTHERWISE PROVIDED FOR, e.g. DEVICES WHICH APPLY A LIFTING OR PUSHING FORCE DIRECTLY TO THE SURFACE OF A LOAD
    • B66F9/00Devices for lifting or lowering bulky or heavy goods for loading or unloading purposes
    • B66F9/06Devices for lifting or lowering bulky or heavy goods for loading or unloading purposes movable, with their loads, on wheels or the like, e.g. fork-lift trucks
    • B66F9/063Automatically guided
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B66HOISTING; LIFTING; HAULING
    • B66FHOISTING, LIFTING, HAULING OR PUSHING, NOT OTHERWISE PROVIDED FOR, e.g. DEVICES WHICH APPLY A LIFTING OR PUSHING FORCE DIRECTLY TO THE SURFACE OF A LOAD
    • B66F9/00Devices for lifting or lowering bulky or heavy goods for loading or unloading purposes
    • B66F9/06Devices for lifting or lowering bulky or heavy goods for loading or unloading purposes movable, with their loads, on wheels or the like, e.g. fork-lift trucks
    • B66F9/075Constructional features or details
    • B66F9/0755Position control; Position detectors
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0231Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
    • G05D1/0238Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using obstacle or wall sensors
    • G05D2201/0216

Definitions

  • the specification relates generally to automated guided vehicles, and specifically to a control augmentation apparatus and method for automated guided vehicles.
  • Automated guided vehicles are employed in a variety of settings, including industrial settings such as warehouse facilities (e.g. for retrieving and moving inventory, or for in-process assembly).
  • the deployment of some AGVs requires the accompanying deployment of physical infrastructure in the facility, such as magnetic or coloured tape, reflectors, RFID tags, and the like.
  • Such infrastructure can be detected by the AGVs, which are in turn configured to execute instructions based on which infrastructure elements they detect.
  • Systems such as those described above can be time-consuming and costly to deploy, and are also not suitable in some environments.
  • the guiding infrastructure is subject to wear and tear and accidental damage, which can result in interruptions to AGV operations while the infrastructure is repaired or replaced.
  • the above-mentioned AGVs are typically capable of executing only a limited set of instructions, requiring the system operators to translate complex spatial and process requirements into such limited instruction sets.
  • FIG. 1 depicts an unaugmented automated guided vehicle, according to a non-limiting embodiment
  • FIG. 2 depicts an augmented automated guided vehicle, according to a non-limiting embodiment
  • FIG. 3 depicts certain internal components of the vehicle of FIG. 2 , according to a non-limiting embodiment
  • FIG. 4 depicts a method of control augmentation in the vehicle of FIG. 2 , according to a non-limiting embodiment
  • FIG. 5 depicts an example overhead view of the operation of the AGV of FIG. 2 in a facility
  • FIG. 6 depicts an example of an AGV consisting of a human-operated vehicle equipped with a control module and navigation sensors according to a non-limiting embodiment.
  • FIG. 1 depicts an automated guided vehicle (AGV) 100 according to certain non-limiting embodiments.
  • AGV 100 in the configuration shown in FIG. 1 , is referred to as an unaugmented or legacy AGV.
  • AGV 100 is a tow vehicle in the example of FIG. 1 , configured to attach to a cart (not shown) and tow the cart about a facility in which AGV is deployed.
  • AGV 100 can have any of a wide variety of other form factors, including other land-based configurations (e.g. with payload-bearing surfaces for carrying payloads directly on the AGV), marine configurations and aerial configurations.
  • AGV 100 includes a chassis 104 supporting various other components, including a locomotion assembly enabling AGV 100 to move about the facility (e.g. a building such as a warehouse, a collection of buildings, an outdoor space, an the like).
  • the locomotion assembly in the example of FIG. 1 , includes a pair of wheels 108 powered by a drive mechanism 112 .
  • different numbers of wheels in different positions on chassis 104 can be implemented.
  • other locomotion assemblies can be employed that include propellers or rotors instead of, or in addition to, wheels 108 .
  • Drive mechanism 112 in the example of FIG. 1 , includes an electric motor connected to each wheel 108 —that is, wheels 108 are differentially driven, permitting AGV 100 to steer by rotating each wheel 108 at a different rate.
  • AGV 100 also includes at least one tow pin 116 in the present example, which can be raised and lowered (together or independently, if multiple tow pins 116 are provided) from an upper surface of chassis 104 to attach and release AGV 100 to and from the above-mentioned carts.
  • Tow pin 116 may be omitted in other embodiments (e.g. those in which AGV 100 bears a payload directly on its upper surface), or replaced with other structures (hooks, anchors, forks for AGVs configured to operate as forklifts, and the like) for securing loads or tools.
  • AGV 100 includes a control module 120 .
  • Control module 120 in general, interfaces with drive mechanism 112 to control the direction and speed of movement of AGV 100 .
  • Control module 120 can also control the deployment and retraction of tow pin 116 .
  • Control of the movement of AGV 100 can be exerted by a processor executing a navigation application within control module 120 .
  • control module 120 receives data from a navigation sensor 124 mounted adjacent to drive mechanism 112 .
  • Navigation sensor 124 is illustrated as a 25 magnetic field sensor that detects the position of a magnetic field, such as that emitted by a strip of magnetic tape 128 placed on the floor of the facility.
  • navigation sensor 124 typically reports a position of magnetic tape 128 relative to the center of the sensor or to one side of the sensor.
  • Control module 120 receives the reported position of tape 128 , and controls drive mechanism 112 based on the difference between the reported position and a target position stored at control module 120 .
  • control module 120 is configured to adjust the control of drive mechanism 112 to correct deviations in the position of tape 128 from the center of sensor 124 .
  • control module 120 is configured to guide AGV 100 along the length of tape 128 .
  • AGV 100 can also include a variety of other navigation sensors for detecting associated infrastructure elements and reporting those elements to control module 120 .
  • AGV 100 can be equipped with one or more cameras to detect colored tape (instead of or in addition to magnetic tape 128 ) or other objects (e.g. reflective markers) in the facility.
  • AGV 100 can be equipped with a radio-frequency identification (RFID) reader to detect RFID tags placed throughout the facility. The detection of a given RFID tag can be interpreted by control module 120 as an instruction to perform an action (e.g. stop, wait for a preconfigured time period, and the like).
  • RFID radio-frequency identification
  • AGV 100 can also include a range-finding sensor 132 , such as a LIDAR sensor.
  • Sensor 132 can be implemented as a safety sensor that interrupts the movement of AGV 100 when an obstacle is detected within a threshold distance (e.g. one meter) of AGV 100 .
  • sensor 132 can be connected directly to drive mechanism 112 rather than to control module 120 .
  • sensor 132 can be connected to control module, which can in turn issue any required interrupt signals when obstacles are detected by sensor 132 .
  • sensor 132 can also be employed for navigational purposes extending beyond interrupting operation of AGV 100 in response to obstacle detections.
  • AGV 100 also includes a human-machine interface (HMI) for presenting status information and receiving instructions from an operator.
  • HMI human-machine interface
  • the HMI includes, in the example of FIG. 1 , a display 136 and at least one input device 140 such as a button, switch or the like.
  • the components of the HMI are connected to the processor of control module 120 . In other examples, some or all of the HMI can be omitted.
  • AGV 200 is illustrated according to certain non-limiting embodiments.
  • AGV 200 includes the components mentioned initially herein as being omitted from AGV 100 .
  • AGV 200 is also referred to as an augmented AGV.
  • AGV 200 includes the components of AGV 100 as described above, and also includes an augmentation module 204 coupled to chassis 104 .
  • augmentation module 204 is coupled to chassis 104 at the opposite end of chassis 104 from control module 120 .
  • both control module 120 and augmentation module 204 can be coupled to chassis 104 at any suitable location.
  • augmentation module 204 can, in some embodiments, be installed in a space within chassis 104 rather than being coupled to an outer surface of chassis 104 .
  • augmentation module 204 is configured to intermediate, at least under certain conditions, between the sensors (such as navigation sensor 124 ) of AGV 200 and control module 120 .
  • augmentation module 204 can generate data defining desired movement or other operations of AGV 200 , and convert such control data into simulated sensor data and send the simulated sensor data to control module 120 .
  • Control module 120 controls drive mechanism 112 based on the simulated sensor data (as mentioned earlier) and is thus enabled to guide AGV 200 even in the absence of infrastructure such as tape 128 .
  • AGV 200 can also include additional components, such as a differential drive rotation sensor 208 coupled to drive mechanism 112 and one or more automation interfaces, such as a power data interface 208 for providing augmentation module 204 with data defining a remaining level of power (e.g. battery charge) for AGV 200 . In other embodiments, however, such additional components can be omitted.
  • AGV 200 can include additional sensors, either mounted on chassis 104 or on augmentation module 204 .
  • augmentation module 204 can include auxiliary navigation sensors (e.g. additional LIDAR sensors, cameras, global positioning system (GPS) sensors, and the like).
  • AGV 200 Before a detailed discussion of the operation of AGV 200 , certain internal components of AGV 200 will be described with reference to FIG. 3 . As will now be apparent, various components of AGV 200 , such as display 136 and sensor 208 , are not shown in FIG. 3 for simplicity of illustration. It will also be apparent that other components of AGV not directly related to the present invention are not shown herein, including a power source (e.g. batteries, solar panels, fuel tank or the like) for the above-mentioned modules and drive mechanism, as well as safety systems.
  • a power source e.g. batteries, solar panels, fuel tank or the like
  • control module 120 includes a central processing unit (CPU) 300 , also referred to as a processor 300 , interconnected with a non-transitory computer-readable medium such as a memory 304 .
  • processor 300 and memory 304 are generally comprised of one or more integrated circuits (ICs), and can have a variety of structures, as will now occur to those skilled in the art (for example, more than one CPU can be provided).
  • Memory 304 can be any suitable combination of volatile (e.g. Random Access Memory (“RAM”)) and non-volatile (e.g. read only memory (“ROM”), Electrically Erasable Programmable Read Only Memory (“EEPROM”), flash memory, magnetic computer storage device, or optical disc) memory.
  • RAM Random Access Memory
  • ROM read only memory
  • EEPROM Electrically Erasable Programmable Read Only Memory
  • flash memory magnetic computer storage device, or optical disc
  • Control module 120 also includes a communications interface connected with processor 300 .
  • Communications interface includes any necessary hardware (executing any suitable programming instructions) for connecting control module to a network 310 .
  • the nature of network 310 is not particularly limited—network 310 can be implemented as any suitable one of, or combination of, wired and wireless networks.
  • network 310 is a local wireless 25 network, and thus communications interface includes one or more radio assemblies for sending and receiving data via network 310 .
  • control module 120 communicates with a legacy control server 312 via communications interface 308 and network 310 .
  • Memory 304 stores a plurality of computer readable instructions executable by processor 300 , including a locomotion application 316 .
  • Application 316 contains a set of instructions for use in controlling the operation of AGV 200 .
  • Each instruction when executed by processor 304 , generates output data for controlling drive mechanism 112 .
  • Some instructions accept sensor data as inputs—for example, an instruction for following tape 128 , as will now be apparent to hose skilled in the art, may take the position of tape 128 as reported by sensor 124 and generate output data for drive mechanism 112 based on the deviation of the reported position of tape 128 and a target position of tape 128 .
  • Another example instruction for stopping AGV 200 may accept as input an identifier from an RFID tag, and generate output data to slow and stop operation of drive mechanism 112 .
  • Other instructions will also occur to those skilled in the art (e.g. an instruction defining which portion of a branch in tape 128 to follow).
  • the above-mentioned instructions can be preconfigured in memory 304 .
  • the instructions can also be deployed to memory 304 , and updated in memory 304 , from legacy control server 312 via network 310 .
  • Augmentation module 204 includes a CPU 320 , also referred to as a processor 320 , interconnected with a non-transitory computer-readable medium such as a memory 324 .
  • Processor 320 and memory 324 are generally comprised of one or more integrated circuits (ICs), and can have a variety of structures, as will now occur to those skilled in the art (for example, more than one CPU can be provided).
  • Memory 324 can be any suitable combination of volatile (e.g. Random Access Memory (“RAM”)) and non-volatile (e.g. read only memory (“ROM”), Electrically Erasable Programmable Read Only Memory (“EEPROM”), flash memory, magnetic computer storage device, or optical disc) memory.
  • RAM Random Access Memory
  • ROM read only memory
  • EEPROM Electrically Erasable Programmable Read Only Memory
  • flash memory magnetic computer storage device, or optical disc
  • Augmentation module 204 also includes a communications interface 328 interconnected with processor 320 . Via communications interface 328 and network 310 , augmentation module 204 can send and receive data to and from an augmented control server 330 , as will be discussed below in greater detail. As noted above in connection with interface 308 , the nature of interface 328 can vary based on the nature of network 310 . In the present example, in which network 310 is a local wireless network deployed within the facility, interface 328 includes one or more radio assemblies, and any necessary supporting hardware.
  • augmentation module 204 includes an inter-module communications interface 332 .
  • interface 332 interconnects processor 320 , processor 300 , range sensor 132 and navigation sensor 124 .
  • Sensor 132 may also, as shown in FIG. 3 , be connected directly to drive mechanism 112 .
  • sensor 132 can be implemented as a safety sensor.
  • sensor 132 in addition to providing range-finding data to interface 332 , sensor 132 can send commands (e.g. interrupt commands) directly to drive mechanism 112 to interrupt the operation of AGV 200 when sensor 132 detects an obstacle within a preconfigured threshold distance of AGV 200 .
  • Sensor 334 can be an additional range-finding sensor (e.g. a LIDAR sensor), an image sensor such as a camera, or the like.
  • a LIDAR sensor e.g. a LIDAR sensor
  • sensor 334 can be employed instead of, or in addition to, sensor 132 for such processes.
  • Interface 332 can have a variety of structures. In the present example, interface 332 includes a plurality of electrical tie-ins intermediating between sensors 124 and 132 and control module 120 , and also connecting augmentation module 204 with control module 120 .
  • interface 332 can include optical, magnetic or other emitting devices (not shown) in proximity with range sensor 132 and navigation sensor 124 to cause sensors 132 and 124 to generate the desired sensor data and transmit such data to processor 300 .
  • Memory 324 stores a plurality of computer readable instructions executable by processor 320 , including a control application 336 .
  • Processor 320 via the execution of application 336 , is configured to determine the location of AGV 200 within the facility (more specifically, within a digital map of the facility stored in memory 324 ).
  • Application 336 also configures processor 320 to generate or receive operational commands, such as paths through the above-mentioned map or target locations within the map. Responsive to obtaining such operational commands, application 336 configures processor 320 to generate control data defining movements or other operations (e.g. deployment of tooling) for AGV 200 .
  • Such control data can be employed to generate operational parameters to control drive mechanism 112 to complete the operational command.
  • augmentation module 204 is not directly connected to drive mechanism 112 . Therefore, application 336 also configures processor 320 to convert the control data into simulated sensor data. As will be seen below, the simulated sensor data is sent to control module 120 where processor 300 (via the execution of application 316 ) acts on such simulated sensor data to control drive mechanism 112 according to the above-mentioned instructions. The control data generated by augmentation module 204 is therefore implemented without being sent directly to drive mechanism 112 .
  • a method 400 of control augmentation will be discussed in conjunction with its performance in AGV 200 . More specifically, the blocks of method 400 are performed by augmentation module 204 via the execution of application 336 by processor 320 . As will be seen below, the performance of method 400 may permit AGV 200 to execute to a broader range of operational commands than could be executed through the operation of control module 104 alone.
  • augmentation module 204 is configured to store a map of the facility in which AGV 200 is deployed.
  • the map is a digital representation of the facility, including indications of the locations of obstacles, tooling or other process stations, areas restricted to AGV 200 , and the like.
  • Various conventional techniques can be employed for generating and deploying the map to AGV 200 , for example from augmented control server 330 , and thus the generation and deployment of the map will not be discussed in further detail herein.
  • the map can also, in some embodiments, include indications of the locations of navigation infrastructure physically installed within the facility, such as tape 128 .
  • augmentation module 204 is configured to obtain an operational command.
  • the operational command can be obtained by receipt from server 330 , for example.
  • the operational command can also be retrieved from memory 324 .
  • application 336 can contain an operational command that is to be performed at a certain time of day, or upon start-up of AGV 200 .
  • the operational command can take a variety of forms.
  • the operational command can be a command for AGV 200 to travel to a target location identifier within the map.
  • the operational command can be a command for AGV 200 to follow a predefined path (e.g. a plurality of sequential line segments each having start and end locations) or portion thereof.
  • the operational command can be a command to execute an action, such as stopping, extending or retracting tow pin 116 , and the like.
  • the operational command can be a command to enable or disable the generation of simulated sensor data by augmentation module 204 (that is, a command to switch between override control by augmentation module 204 via control module 120 , and control by control module 120 alone).
  • augmentation module 204 is configured to receive sensor data from the sensors connected to interface 332 .
  • processor 320 receives sensor data from range sensor 132 and navigation sensor 124 .
  • augmentation module 204 is configured to update the localization of AGV 200 at block 420 .
  • Various localization techniques will now occur to those skilled in the art, and a detailed discussion of localization is therefore not necessary.
  • augmentation module 204 determines a current location of AGV 200 within the map stored at block 405 (i.e. within the facility in which AGV 200 operates). Such a determination may be made by comparing some or all of the sensor data (e.g.
  • block 420 can include simply receiving an updated location at augmentation module 204 , for example from server 330 . That is, server 330 can be responsible for vehicle localization (e.g. via detection of a beacon signal emitted by AGV 200 and determination of AGV location based on the signal). In such embodiments, the receipt of sensor data at block 415 may even be omitted.
  • augmentation module 204 can also be configured to update the map stored in memory 324 at block 425 , following receipt of sensor data. For example, scan data from sensor 132 may be employed to update the map to represent a previously unmapped portion of the facility, or to reflect that a structure has been relocated within the facility. In addition, data from sensor 124 can be employed to update the map to reflect the presence of tape 12 .
  • augmentation module 204 is configured to determine, at block 430 , whether to override the “default” control—that is, the control of AGV 200 provided by the execution of application 316 by processor 300 , based on sensor data from sensors 132 and 124 .
  • the determination at block 430 is a determination of whether to generate control data and corresponding simulated sensor data at augmentation module 204 , or whether to allow control of AGV 200 to be handled entirely by control module 120 (i.e. based on true sensor data).
  • the determination at block 430 can take a variety of forms.
  • the nature of the operational command at block 410 can influence the determination at block 430 .
  • the operational command itself can be a command to set AGV 200 to a particular control mode (e.g. override control via augmentation module 204 or legacy control via control module 120 alone).
  • the determination at block 430 can be based on sensor data received at block 415 : if the sensor data indicates that no guiding infrastructure is present (e.g. magnetic tape 128 is not detected), the determination at block 430 is affirmative.
  • sensor data may indicate that an unmapped obstacle is in the vicinity of AGV 200 and drive mechanism 112 is currently under the control of control module 120 alone, the determination may be affirmative. On the other hand, if sensor data indicates that infrastructure such as tape 128 present, the determination may be negative. In a further example, sensor data may indicate the presence of a nearby RFID tag that corresponds to an instruction to switch to default control or to override control.
  • the determination at block 430 can also be based on the map stored at block 405 .
  • the map can include indications of the locations of guiding infrastructure such as tape 128 in the facility. If, following localization at block 420 , the current location of AGV 200 coincides with the location of guiding infrastructure (i.e. AGV 200 is currently located over tape 128 ), the determination at block 430 may be negative.
  • the map can also include data defining virtual guiding infrastructure, such as simulated magnetic tape (represented in the digital map, but not actually present in the facility). Such virtual infrastructure can be distinguished in the map from physical infrastructure, and if the current location of AGV 200 coincides with virtual infrastructure rather than physical infrastructure, the determination at block 430 may be affirmative.
  • processor 320 is configured to generate control data defining movements or other tasks required of AGV 200 to perform the operational command received at block 410 .
  • the control data generated can include one or more path segments, each defined at least by a starting location and a heading (e.g. a steering angle relative to the current orientation of AGV 200 ).
  • the generation of path data can be performed according to any suitable algorithms that will occur to those skilled in the art.
  • the control data is based on the operational command, the map and the current location of AGV 200 as determined at block 420 .
  • augmentation module 204 is configured to convert the control data to simulated sensor data before transmission to control module 120 .
  • augmentation module 204 exerts control over drive mechanism 112 indirectly through control module 120 .
  • control module 120 executes a set of instructions, for example to follow magnetic tape 128 or perform actions upon detection of certain RFID tags. Thus, control module 120 may not be capable of acting on the control data generated by augmentation module 204 .
  • augmentation module 204 therefore converts the control data into simulated sensor inputs that are known to cause control module to control drive mechanism 112 in the same manner as specified by the original control data.
  • memory 324 stores a plurality of conversion definitions each corresponding to a type of control data generated at block 440 . Table 1 shows three example conversion definitions.
  • Each definition corresponds to a type of control data generated by augmentation module 204 (e.g. a heading direction in degrees, a stop action and a tow pin deployment action).
  • Each definition identifies which sensor or combination of sensors for which to generate simulated sensor data, and also indicates how to generate the simulated sensor data.
  • a stop action is converted by generating simulated sensor data that simulates the detection of a particular RFID tag (having the identifier “ 123 ”) by an RFID reader.
  • control data specifying a heading to steer AGV 200 is converted by generated data simulating a position of tape 128 as reported by sensor 124 .
  • the simulated position can be generated by multiplying the heading by a factor “A” to obtain a reported position of tape 128 (e.g. in millimeters) along the length of sensor 124 .
  • conversion definitions depend on the nature of the sensors being simulated.
  • the conversion definitions need not be stored in a table as shown above.
  • conversion definitions can be stored in any suitable flat file or database structure.
  • conversion definitions can be stored as sequences of executable instructions in application 336 .
  • processor 320 is configured to transmit the simulated sensor data to control module 120 via inter-module communications interface 332 .
  • the simulated sensor data thus arrives at processor 300 , and is processed by control module 120 in the same manner as true sensor data to control the operation of AGV 200 (e.g. to control drive mechanism 112 , tow pin 116 and the like).
  • Performance of method 400 then returns to block 410 for the receipt of a further operational command (or the continued processing of a previous operational command).
  • FIG. 5 a simplified overhead view is shown of a facility in which AGV 200 is deployed.
  • the facility includes a route defined by magnetic tape 128 , and also includes a plurality of processing stations 504 , 508 , 512 (e.g. inventory loading or unloading, tooling stations and the like).
  • AGV 200 can first be configured to travel along tape 128 without overriding control by augmentation module 204 .
  • a first RFID tag 516 can represent an instruction to switch from default control to override control (that is, an affirmative determination at block 430 ).
  • AGV 200 is instructed (via an operational command received at block 410 ) to follow a path 520 .
  • the path 520 can be defined in the map stored at block 405 as a route of virtual magnetic tape.
  • augmentation module 204 generates simulated sensor data to control the movement of AGV 200 via control module 120 .
  • the map stored in augmentation module 204 can also include virtual RFID tags, such as virtual tag 524 , which represents an instruction to stop adjacent to process station 504 .
  • subsequent operational commands e.g. from server 330
  • can instruct AGV 200 still under the indirect control of augmentation module 204 —to travel to, and stop at, process station 508 .
  • an operational command indicating a target location along path 520 can be sent by server 330 to augmentation module 204 .
  • AGV 200 will return onto the route defined by tape 128 , at which point control may be returned to control module 120 via a negative determination at block 430 .
  • augmentation module 204 continues to monitor sensor data, and may detect the presence of an obstacle 528 .
  • augmentation module 204 may begin the generation of simulated sensor data to depart from tape 128 and guide AGV 200 around obstacle 528 .
  • a further RFID tag 536 may instruct control module 120 to stop AGV 200 adjacent process station 512 .
  • RFID tag 536 may be the tag simulated by augmentation module to stop AGV 200 at virtual tag 524 and at process station 508 .
  • sensor data received at block 415 can be compared with a map stored in memory 324 and containing data representing the locations of existing guiding infrastructure (e.g. tape 128 ). If the comparison reveals that tape 128 is not present where indicated by map 405 , augmentation module 204 can not only override the default control by control module 120 , but can also notify server 330 that the guiding infrastructure has been damaged or removed in the current location of AGV 200 .
  • existing guiding infrastructure e.g. tape 128
  • the operational commands received at block 410 can be teleoperation commands received from a manually-controlled computing device operated by a human pilot of AGV 200 . That is, AGV 200 can be piloted substantially in real-time as a remote-controlled vehicle through the performance of method 400 .
  • augmentation module 204 may have a direct connection to drive mechanism 112 .
  • the performance of method 400 may be simplified, in that the conversion of control data to simulated sensor data may no longer be necessary.
  • control module 120 may be configured to operate in one of two modes—a legacy operation mode as described above, and a relay mode in which control module 120 accepts control data from augmentation module 204 and relays such data to drive mechanism 112 (i.e. without augmentation module 204 being required to generate simulated sensor data).
  • augmentation module 204 can be configured to automatically determine conversion definitions, rather than have the conversion definitions preconfigured as described above.
  • processors 300 or 320 executing their respective applications may be implemented using pre-programmed hardware or firmware elements (e.g., application specific integrated circuits (ASICs), electrically erasable programmable read-only memories (EEPROMs), etc.), or other related components.
  • ASICs application specific integrated circuits
  • EEPROMs electrically erasable programmable read-only memories
  • an AGV 600 is illustrated according to certain non-limiting embodiments.
  • the AGV 600 is shown as a forklift, though other human-operated material-transport vehicles may also be used, such as tuggers and pallet trucks/jacks.
  • the forklift as depicted is considered as an AGV 600 for the purposes herein based on the inclusion of a control module (not shown in FIG. 6 ) and a navigation sensor 624 .
  • the AGV 600 may comprise a human-operated forklift (or other material-transport vehicle) that has been retrofitted with equipment to convert the forklift to an AGV.
  • the original equipment of the forklift may include the AGV components (e.g. control module, navigation sensors, etc.).
  • the navigation sensor 624 may be a magnetic field sensor for detecting a strip of magnetic tape 628 placed on the floor of the facility.
  • the control module is configured to control a drive mechanism of the forklift to correct deviations in the position of the tape 628 from the center of the sensor 124 .
  • the control module is configured to guide the AGV 600 along the length of the tape 628 .
  • the AGV 600 further includes an augmentation module 704 coupled to the forklift.
  • the augmentation module 704 may be coupled to the AGV 600 in any suitable location.
  • the AGV 600 can include additional components such as a differential drive rotation sensor (e.g. coupled to the drive mechanism of the forklift), and one or more automation interfaces such as a power data interface for providing the augmentation module 704 with data defining a remaining power level (e.g. battery charge) for the AGV 600 .
  • the AGV 600 can include additional sensors, either mounted on the forklift or on the augmentation module 704 .
  • the augmentation module 704 can include auxiliary navigation sensors (e.g. additional LiDAR sensors, cameras, GPS sensors, and the like).

Landscapes

  • Engineering & Computer Science (AREA)
  • Transportation (AREA)
  • Structural Engineering (AREA)
  • Physics & Mathematics (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Automation & Control Theory (AREA)
  • General Physics & Mathematics (AREA)
  • Remote Sensing (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mechanical Engineering (AREA)
  • Geology (AREA)
  • Civil Engineering (AREA)
  • Electromagnetism (AREA)
  • Navigation (AREA)
  • Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)

Abstract

An augmentation module is described for an automated guided vehicle (AGV) deployed in a facility and including a control module for controlling a drive mechanism based on navigational data received from a navigation sensor. The module includes a inter-module communications interface connected to the control module; a memory; and a processor connected to the communications interface and the memory. The processor is configured to: obtain an operational command; generate control data to execute the operational command; convert the control data to simulated sensor data; and send the simulated sensor data to the control module.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. patent application Ser. No. 16/855,788, filed on 22 Apr. 2020, which is a continuation of U.S. patent application Ser. No. 15/429,222, filed on 10 Feb. 2017, issued as U.S. Pat. No. 10,662,045 on 26 May 2020, which claims the benefit of U.S. Provisional Patent Application No. 62/293,948, filed on 11 Feb. 2016. The entire contents of each of U.S. patent application Ser. No. 15/429,222, U.S. patent application Ser. No. 16/855,788, and U.S. Provisional Patent Application No. 62/293,948 is hereby incorporated by reference.
  • FIELD
  • The specification relates generally to automated guided vehicles, and specifically to a control augmentation apparatus and method for automated guided vehicles.
  • BACKGROUND
  • Automated guided vehicles (AGVs) are employed in a variety of settings, including industrial settings such as warehouse facilities (e.g. for retrieving and moving inventory, or for in-process assembly). The deployment of some AGVs requires the accompanying deployment of physical infrastructure in the facility, such as magnetic or coloured tape, reflectors, RFID tags, and the like. Such infrastructure can be detected by the AGVs, which are in turn configured to execute instructions based on which infrastructure elements they detect.
  • Systems such as those described above can be time-consuming and costly to deploy, and are also not suitable in some environments. In addition, the guiding infrastructure is subject to wear and tear and accidental damage, which can result in interruptions to AGV operations while the infrastructure is repaired or replaced. Further, the above-mentioned AGVs are typically capable of executing only a limited set of instructions, requiring the system operators to translate complex spatial and process requirements into such limited instruction sets.
  • BRIEF DESCRIPTIONS OF THE DRAWINGS
  • Embodiments are described with reference to the following figures, in which:
  • FIG. 1 depicts an unaugmented automated guided vehicle, according to a non-limiting embodiment;
  • FIG. 2 depicts an augmented automated guided vehicle, according to a non-limiting embodiment;
  • FIG. 3 depicts certain internal components of the vehicle of FIG. 2, according to a non-limiting embodiment;
  • FIG. 4 depicts a method of control augmentation in the vehicle of FIG. 2, according to a non-limiting embodiment;
  • FIG. 5 depicts an example overhead view of the operation of the AGV of FIG. 2 in a facility; and
  • FIG. 6 depicts an example of an AGV consisting of a human-operated vehicle equipped with a control module and navigation sensors according to a non-limiting embodiment.
  • DETAILED DESCRIPTION OF THE EMBODIMENTS
  • FIG. 1 depicts an automated guided vehicle (AGV) 100 according to certain non-limiting embodiments. AGV 100, in the configuration shown in FIG. 1, is referred to as an unaugmented or legacy AGV. AGV 100 is a tow vehicle in the example of FIG. 1, configured to attach to a cart (not shown) and tow the cart about a facility in which AGV is deployed. However, AGV 100 can have any of a wide variety of other form factors, including other land-based configurations (e.g. with payload-bearing surfaces for carrying payloads directly on the AGV), marine configurations and aerial configurations.
  • AGV 100 includes a chassis 104 supporting various other components, including a locomotion assembly enabling AGV 100 to move about the facility (e.g. a building such as a warehouse, a collection of buildings, an outdoor space, an the like). The locomotion assembly, in the example of FIG. 1, includes a pair of wheels 108 powered by a drive mechanism 112. In other examples, different numbers of wheels in different positions on chassis 104 can be implemented. In addition, for marine or aerial AGVs, other locomotion assemblies can be employed that include propellers or rotors instead of, or in addition to, wheels 108.
  • Drive mechanism 112, in the example of FIG. 1, includes an electric motor connected to each wheel 108—that is, wheels 108 are differentially driven, permitting AGV 100 to steer by rotating each wheel 108 at a different rate. In other examples, various other drive mechanisms can be employed to drive wheels 108. AGV 100 also includes at least one tow pin 116 in the present example, which can be raised and lowered (together or independently, if multiple tow pins 116 are provided) from an upper surface of chassis 104 to attach and release AGV 100 to and from the above-mentioned carts. Tow pin 116 may be omitted in other embodiments (e.g. those in which AGV 100 bears a payload directly on its upper surface), or replaced with other structures (hooks, anchors, forks for AGVs configured to operate as forklifts, and the like) for securing loads or tools.
  • In addition, AGV 100 includes a control module 120. Control module 120, in general, interfaces with drive mechanism 112 to control the direction and speed of movement of AGV 100. Control module 120 can also control the deployment and retraction of tow pin 116.
  • Control of the movement of AGV 100 can be exerted by a processor executing a navigation application within control module 120. In the example of FIG. 1, control module 120 receives data from a navigation sensor 124 mounted adjacent to drive mechanism 112. Navigation sensor 124 is illustrated as a 25 magnetic field sensor that detects the position of a magnetic field, such as that emitted by a strip of magnetic tape 128 placed on the floor of the facility. As will now be apparent to those skilled in the art, navigation sensor 124 typically reports a position of magnetic tape 128 relative to the center of the sensor or to one side of the sensor. Control module 120 receives the reported position of tape 128, and controls drive mechanism 112 based on the difference between the reported position and a target position stored at control module 120. Generally, the target position is centered on sensor 124, and thus control module 120 is configured to adjust the control of drive mechanism 112 to correct deviations in the position of tape 128 from the center of sensor 124. In other words, control module 120 is configured to guide AGV 100 along the length of tape 128.
  • AGV 100 can also include a variety of other navigation sensors for detecting associated infrastructure elements and reporting those elements to control module 120. For example, AGV 100 can be equipped with one or more cameras to detect colored tape (instead of or in addition to magnetic tape 128) or other objects (e.g. reflective markers) in the facility. In further examples, AGV 100 can be equipped with a radio-frequency identification (RFID) reader to detect RFID tags placed throughout the facility. The detection of a given RFID tag can be interpreted by control module 120 as an instruction to perform an action (e.g. stop, wait for a preconfigured time period, and the like).
  • AGV 100 can also include a range-finding sensor 132, such as a LIDAR sensor. Sensor 132 can be implemented as a safety sensor that interrupts the movement of AGV 100 when an obstacle is detected within a threshold distance (e.g. one meter) of AGV 100. When employed as a safety sensor, sensor 132 can be connected directly to drive mechanism 112 rather than to control module 120. In other examples, however, sensor 132 can be connected to control module, which can in turn issue any required interrupt signals when obstacles are detected by sensor 132. As will be discussed in greater detail below, sensor 132 can also be employed for navigational purposes extending beyond interrupting operation of AGV 100 in response to obstacle detections.
  • AGV 100 also includes a human-machine interface (HMI) for presenting status information and receiving instructions from an operator. The HMI includes, in the example of FIG. 1, a display 136 and at least one input device 140 such as a button, switch or the like. The components of the HMI are connected to the processor of control module 120. In other examples, some or all of the HMI can be omitted.
  • Referring now to FIG. 2, an AGV 200 is illustrated according to certain non-limiting embodiments. AGV 200 includes the components mentioned initially herein as being omitted from AGV 100. As a result, AGV 200 is also referred to as an augmented AGV.
  • AGV 200 includes the components of AGV 100 as described above, and also includes an augmentation module 204 coupled to chassis 104. In the present example, augmentation module 204 is coupled to chassis 104 at the opposite end of chassis 104 from control module 120. In other embodiments, however, both control module 120 and augmentation module 204 can be coupled to chassis 104 at any suitable location. For example, augmentation module 204 can, in some embodiments, be installed in a space within chassis 104 rather than being coupled to an outer surface of chassis 104. As will be discussed in greater detail below, augmentation module 204 is configured to intermediate, at least under certain conditions, between the sensors (such as navigation sensor 124) of AGV 200 and control module 120. More specifically, augmentation module 204 can generate data defining desired movement or other operations of AGV 200, and convert such control data into simulated sensor data and send the simulated sensor data to control module 120. Control module 120, in turn, controls drive mechanism 112 based on the simulated sensor data (as mentioned earlier) and is thus enabled to guide AGV 200 even in the absence of infrastructure such as tape 128.
  • AGV 200 can also include additional components, such as a differential drive rotation sensor 208 coupled to drive mechanism 112 and one or more automation interfaces, such as a power data interface 208 for providing augmentation module 204 with data defining a remaining level of power (e.g. battery charge) for AGV 200. In other embodiments, however, such additional components can be omitted. In further embodiments, AGV 200 can include additional sensors, either mounted on chassis 104 or on augmentation module 204. For instance, augmentation module 204 can include auxiliary navigation sensors (e.g. additional LIDAR sensors, cameras, global positioning system (GPS) sensors, and the like).
  • Before a detailed discussion of the operation of AGV 200, certain internal components of AGV 200 will be described with reference to FIG. 3. As will now be apparent, various components of AGV 200, such as display 136 and sensor 208, are not shown in FIG. 3 for simplicity of illustration. It will also be apparent that other components of AGV not directly related to the present invention are not shown herein, including a power source (e.g. batteries, solar panels, fuel tank or the like) for the above-mentioned modules and drive mechanism, as well as safety systems.
  • As shown in FIG. 3, control module 120 includes a central processing unit (CPU) 300, also referred to as a processor 300, interconnected with a non-transitory computer-readable medium such as a memory 304. Processor 300 and memory 304 are generally comprised of one or more integrated circuits (ICs), and can have a variety of structures, as will now occur to those skilled in the art (for example, more than one CPU can be provided). Memory 304 can be any suitable combination of volatile (e.g. Random Access Memory (“RAM”)) and non-volatile (e.g. read only memory (“ROM”), Electrically Erasable Programmable Read Only Memory (“EEPROM”), flash memory, magnetic computer storage device, or optical disc) memory.
  • Control module 120 also includes a communications interface connected with processor 300. Communications interface includes any necessary hardware (executing any suitable programming instructions) for connecting control module to a network 310. The nature of network 310 is not particularly limited—network 310 can be implemented as any suitable one of, or combination of, wired and wireless networks. In the present embodiment, network 310 is a local wireless 25 network, and thus communications interface includes one or more radio assemblies for sending and receiving data via network 310. In particular, control module 120 communicates with a legacy control server 312 via communications interface 308 and network 310.
  • Memory 304 stores a plurality of computer readable instructions executable by processor 300, including a locomotion application 316. Application 316 contains a set of instructions for use in controlling the operation of AGV 200. Each instruction, when executed by processor 304, generates output data for controlling drive mechanism 112. Some instructions accept sensor data as inputs—for example, an instruction for following tape 128, as will now be apparent to hose skilled in the art, may take the position of tape 128 as reported by sensor 124 and generate output data for drive mechanism 112 based on the deviation of the reported position of tape 128 and a target position of tape 128. Another example instruction for stopping AGV 200 may accept as input an identifier from an RFID tag, and generate output data to slow and stop operation of drive mechanism 112. Other instructions will also occur to those skilled in the art (e.g. an instruction defining which portion of a branch in tape 128 to follow).
  • The above-mentioned instructions can be preconfigured in memory 304. The instructions can also be deployed to memory 304, and updated in memory 304, from legacy control server 312 via network 310.
  • Augmentation module 204 includes a CPU 320, also referred to as a processor 320, interconnected with a non-transitory computer-readable medium such as a memory 324. Processor 320 and memory 324 are generally comprised of one or more integrated circuits (ICs), and can have a variety of structures, as will now occur to those skilled in the art (for example, more than one CPU can be provided). Memory 324 can be any suitable combination of volatile (e.g. Random Access Memory (“RAM”)) and non-volatile (e.g. read only memory (“ROM”), Electrically Erasable Programmable Read Only Memory (“EEPROM”), flash memory, magnetic computer storage device, or optical disc) memory.
  • Augmentation module 204 also includes a communications interface 328 interconnected with processor 320. Via communications interface 328 and network 310, augmentation module 204 can send and receive data to and from an augmented control server 330, as will be discussed below in greater detail. As noted above in connection with interface 308, the nature of interface 328 can vary based on the nature of network 310. In the present example, in which network 310 is a local wireless network deployed within the facility, interface 328 includes one or more radio assemblies, and any necessary supporting hardware.
  • Further, augmentation module 204 includes an inter-module communications interface 332. As seen in FIG. 3, interface 332 interconnects processor 320, processor 300, range sensor 132 and navigation sensor 124. Sensor 132 may also, as shown in FIG. 3, be connected directly to drive mechanism 112. As mentioned above, sensor 132 can be implemented as a safety sensor. In such implementations, in addition to providing range-finding data to interface 332, sensor 132 can send commands (e.g. interrupt commands) directly to drive mechanism 112 to interrupt the operation of AGV 200 when sensor 132 detects an obstacle within a preconfigured threshold distance of AGV 200.
  • Other sensors can also be connected to interface 332, including a localization sensor 334 housed in augmentation module 204. Sensor 334 may be an additional range-finding sensor (e.g. a LIDAR sensor), an image sensor such as a camera, or the like. Although the discussion below assumes that data from sensor 132 is employed by augmentation module 204 for mapping and localization processes, in some embodiments sensor 334 can be employed instead of, or in addition to, sensor 132 for such processes. Interface 332 can have a variety of structures. In the present example, interface 332 includes a plurality of electrical tie-ins intermediating between sensors 124 and 132 and control module 120, and also connecting augmentation module 204 with control module 120. In other embodiments, instead of electrical connections to processor 300, interface 332 can include optical, magnetic or other emitting devices (not shown) in proximity with range sensor 132 and navigation sensor 124 to cause sensors 132 and 124 to generate the desired sensor data and transmit such data to processor 300.
  • Memory 324 stores a plurality of computer readable instructions executable by processor 320, including a control application 336. Processor 320, via the execution of application 336, is configured to determine the location of AGV 200 within the facility (more specifically, within a digital map of the facility stored in memory 324). Application 336 also configures processor 320 to generate or receive operational commands, such as paths through the above-mentioned map or target locations within the map. Responsive to obtaining such operational commands, application 336 configures processor 320 to generate control data defining movements or other operations (e.g. deployment of tooling) for AGV 200. Such control data can be employed to generate operational parameters to control drive mechanism 112 to complete the operational command. However, as will now be apparent, augmentation module 204 is not directly connected to drive mechanism 112. Therefore, application 336 also configures processor 320 to convert the control data into simulated sensor data. As will be seen below, the simulated sensor data is sent to control module 120 where processor 300 (via the execution of application 316) acts on such simulated sensor data to control drive mechanism 112 according to the above-mentioned instructions. The control data generated by augmentation module 204 is therefore implemented without being sent directly to drive mechanism 112.
  • Referring now to FIG. 4, a method 400 of control augmentation will be discussed in conjunction with its performance in AGV 200. More specifically, the blocks of method 400 are performed by augmentation module 204 via the execution of application 336 by processor 320. As will be seen below, the performance of method 400 may permit AGV 200 to execute to a broader range of operational commands than could be executed through the operation of control module 104 alone.
  • Beginning at block 405, augmentation module 204 is configured to store a map of the facility in which AGV 200 is deployed. In general, the map is a digital representation of the facility, including indications of the locations of obstacles, tooling or other process stations, areas restricted to AGV 200, and the like. Various conventional techniques can be employed for generating and deploying the map to AGV 200, for example from augmented control server 330, and thus the generation and deployment of the map will not be discussed in further detail herein. The map can also, in some embodiments, include indications of the locations of navigation infrastructure physically installed within the facility, such as tape 128.
  • At block 410, augmentation module 204 is configured to obtain an operational command. The operational command can be obtained by receipt from server 330, for example. The operational command can also be retrieved from memory 324. For example, application 336 can contain an operational command that is to be performed at a certain time of day, or upon start-up of AGV 200.
  • The operational command can take a variety of forms. For example, the operational command can be a command for AGV 200 to travel to a target location identifier within the map. In other examples, the operational command can be a command for AGV 200 to follow a predefined path (e.g. a plurality of sequential line segments each having start and end locations) or portion thereof. In further examples, the operational command can be a command to execute an action, such as stopping, extending or retracting tow pin 116, and the like. In still further examples, the operational command can be a command to enable or disable the generation of simulated sensor data by augmentation module 204 (that is, a command to switch between override control by augmentation module 204 via control module 120, and control by control module 120 alone).
  • At block 415, augmentation module 204 is configured to receive sensor data from the sensors connected to interface 332. In the present example, therefore, processor 320 receives sensor data from range sensor 132 and navigation sensor 124. Based on the received sensor data, augmentation module 204 is configured to update the localization of AGV 200 at block 420. Various localization techniques will now occur to those skilled in the art, and a detailed discussion of localization is therefore not necessary. In general, however, at block 420 augmentation module 204 determines a current location of AGV 200 within the map stored at block 405 (i.e. within the facility in which AGV 200 operates). Such a determination may be made by comparing some or all of the sensor data (e.g. a laser scan of the immediate surroundings of AGV 200 received from sensor 132) with the map. In certain embodiments, block 420 can include simply receiving an updated location at augmentation module 204, for example from server 330. That is, server 330 can be responsible for vehicle localization (e.g. via detection of a beacon signal emitted by AGV 200 and determination of AGV location based on the signal). In such embodiments, the receipt of sensor data at block 415 may even be omitted.
  • In some embodiments, augmentation module 204 can also be configured to update the map stored in memory 324 at block 425, following receipt of sensor data. For example, scan data from sensor 132 may be employed to update the map to represent a previously unmapped portion of the facility, or to reflect that a structure has been relocated within the facility. In addition, data from sensor 124 can be employed to update the map to reflect the presence of tape 12.
  • Having updated the localization of AGV 200, augmentation module 204 is configured to determine, at block 430, whether to override the “default” control—that is, the control of AGV 200 provided by the execution of application 316 by processor 300, based on sensor data from sensors 132 and 124. In other words, the determination at block 430 is a determination of whether to generate control data and corresponding simulated sensor data at augmentation module 204, or whether to allow control of AGV 200 to be handled entirely by control module 120 (i.e. based on true sensor data).
  • The determination at block 430 can take a variety of forms. In some embodiments, the nature of the operational command at block 410 can influence the determination at block 430. For example, the operational command itself can be a command to set AGV 200 to a particular control mode (e.g. override control via augmentation module 204 or legacy control via control module 120 alone). In another example, the determination at block 430 can be based on sensor data received at block 415: if the sensor data indicates that no guiding infrastructure is present (e.g. magnetic tape 128 is not detected), the determination at block 430 is affirmative. In a further example, if the sensor data indicates that an unmapped obstacle is in the vicinity of AGV 200 and drive mechanism 112 is currently under the control of control module 120 alone, the determination may be affirmative. On the other hand, if sensor data indicates that infrastructure such as tape 128 present, the determination may be negative. In a further example, sensor data may indicate the presence of a nearby RFID tag that corresponds to an instruction to switch to default control or to override control.
  • The determination at block 430 can also be based on the map stored at block 405. For example, the map can include indications of the locations of guiding infrastructure such as tape 128 in the facility. If, following localization at block 420, the current location of AGV 200 coincides with the location of guiding infrastructure (i.e. AGV 200 is currently located over tape 128), the determination at block 430 may be negative. In a further example, the map can also include data defining virtual guiding infrastructure, such as simulated magnetic tape (represented in the digital map, but not actually present in the facility). Such virtual infrastructure can be distinguished in the map from physical infrastructure, and if the current location of AGV 200 coincides with virtual infrastructure rather than physical infrastructure, the determination at block 430 may be affirmative.
  • When the determination at block 430 is negative, the sensor data received at block 415 is sent to control module 120. Control module 120, in turn, controls drive mechanism 112 by executing the above-mentioned instructions of application 316. When the determination at block 430 is affirmative, however, augmentation module 204 proceeds to block 440. At block 440, processor 320 is configured to generate control data defining movements or other tasks required of AGV 200 to perform the operational command received at block 410. For example, if the operational command is to travel to a target location within the map, the control data generated can include one or more path segments, each defined at least by a starting location and a heading (e.g. a steering angle relative to the current orientation of AGV 200). The generation of path data can be performed according to any suitable algorithms that will occur to those skilled in the art. The control data is based on the operational command, the map and the current location of AGV 200 as determined at block 420.
  • At block 445, having generated control data, augmentation module 204 is configured to convert the control data to simulated sensor data before transmission to control module 120. As will now be apparent, due to the lack (in the present embodiment) of a direct connection between augmentation module 120 and drive mechanism 112, augmentation module 204 exerts control over drive mechanism 112 indirectly through control module 120. Further, as noted earlier control module 120 executes a set of instructions, for example to follow magnetic tape 128 or perform actions upon detection of certain RFID tags. Thus, control module 120 may not be capable of acting on the control data generated by augmentation module 204.
  • To exert control over drive mechanism 112 through control module 120, augmentation module 204 therefore converts the control data into simulated sensor inputs that are known to cause control module to control drive mechanism 112 in the same manner as specified by the original control data. In the present example, memory 324 stores a plurality of conversion definitions each corresponding to a type of control data generated at block 440. Table 1 shows three example conversion definitions.
  • TABLE 1
    Example Conversion Definitions
    Control Data Type Sensor to Simulate Simulated Sensor Data
    Heading (degrees) Navigation sensor 124 Tape position: A x
    heading (mm)
    Action: Stop RFID Reader Tag ID: 123
    Action: Deploy RFID Reader Tag ID: 456
    Tow Pin
  • In the above example, three conversion definitions are shown. Each definition corresponds to a type of control data generated by augmentation module 204 (e.g. a heading direction in degrees, a stop action and a tow pin deployment action). Each definition identifies which sensor or combination of sensors for which to generate simulated sensor data, and also indicates how to generate the simulated sensor data. For example, a stop action is converted by generating simulated sensor data that simulates the detection of a particular RFID tag (having the identifier “123”) by an RFID reader. As a further example, control data specifying a heading to steer AGV 200 is converted by generated data simulating a position of tape 128 as reported by sensor 124. The simulated position can be generated by multiplying the heading by a factor “A” to obtain a reported position of tape 128 (e.g. in millimeters) along the length of sensor 124.
  • A wide variety of other conversion definitions will also occur to those skilled in the art. It will also now be apparent that the nature of the conversion definitions depend on the nature of the sensors being simulated. The conversion definitions need not be stored in a table as shown above. In other embodiments, conversion definitions can be stored in any suitable flat file or database structure. In further embodiments, conversion definitions can be stored as sequences of executable instructions in application 336.
  • Having generated the simulated sensor data at block 445, at block 450 processor 320 is configured to transmit the simulated sensor data to control module 120 via inter-module communications interface 332. The simulated sensor data thus arrives at processor 300, and is processed by control module 120 in the same manner as true sensor data to control the operation of AGV 200 (e.g. to control drive mechanism 112, tow pin 116 and the like). Performance of method 400 then returns to block 410 for the receipt of a further operational command (or the continued processing of a previous operational command).
  • Turning now to FIG. 5, a simplified overhead view is shown of a facility in which AGV 200 is deployed. As seen in FIG. 5, the facility includes a route defined by magnetic tape 128, and also includes a plurality of processing stations 504, 508, 512 (e.g. inventory loading or unloading, tooling stations and the like). Through the performance of method 400 by augmentation module 204 and the above-mentioned control processes of control module 120, AGV 200 can first be configured to travel along tape 128 without overriding control by augmentation module 204. A first RFID tag 516 can represent an instruction to switch from default control to override control (that is, an affirmative determination at block 430). Following the detection of tag 516, in the present example AGV 200 is instructed (via an operational command received at block 410) to follow a path 520. For example, the path 520 can be defined in the map stored at block 405 as a route of virtual magnetic tape. Along path 520, therefore, augmentation module 204 generates simulated sensor data to control the movement of AGV 200 via control module 120.
  • The map stored in augmentation module 204 can also include virtual RFID tags, such as virtual tag 524, which represents an instruction to stop adjacent to process station 504. In addition, subsequent operational commands (e.g. from server 330) can instruct AGV 200—still under the indirect control of augmentation module 204—to travel to, and stop at, process station 508. In the absence of a virtual tag stored in the map at the location corresponding to the real location of station 508, an operational command indicating a target location along path 520 can be sent by server 330 to augmentation module 204.
  • As AGV 200 continues to travel along path 520, AGV 200 will return onto the route defined by tape 128, at which point control may be returned to control module 120 via a negative determination at block 430. However, augmentation module 204 continues to monitor sensor data, and may detect the presence of an obstacle 528. As a result, augmentation module 204 may begin the generation of simulated sensor data to depart from tape 128 and guide AGV 200 around obstacle 528. Finally, a further RFID tag 536 may instruct control module 120 to stop AGV 200 adjacent process station 512. As will now be apparent to those skilled in the art, RFID tag 536 may be the tag simulated by augmentation module to stop AGV 200 at virtual tag 524 and at process station 508.
  • Various additional processes can be performed by augmentation module 204. For example, in some embodiments, sensor data received at block 415 can be compared with a map stored in memory 324 and containing data representing the locations of existing guiding infrastructure (e.g. tape 128). If the comparison reveals that tape 128 is not present where indicated by map 405, augmentation module 204 can not only override the default control by control module 120, but can also notify server 330 that the guiding infrastructure has been damaged or removed in the current location of AGV 200.
  • In further variations, the operational commands received at block 410 can be teleoperation commands received from a manually-controlled computing device operated by a human pilot of AGV 200. That is, AGV 200 can be piloted substantially in real-time as a remote-controlled vehicle through the performance of method 400.
  • In other variations, augmentation module 204 may have a direct connection to drive mechanism 112. In such embodiments, the performance of method 400 may be simplified, in that the conversion of control data to simulated sensor data may no longer be necessary. In further embodiments, control module 120 may be configured to operate in one of two modes—a legacy operation mode as described above, and a relay mode in which control module 120 accepts control data from augmentation module 204 and relays such data to drive mechanism 112 (i.e. without augmentation module 204 being required to generate simulated sensor data).
  • In further embodiments, augmentation module 204 can be configured to automatically determine conversion definitions, rather than have the conversion definitions preconfigured as described above.
  • Those skilled in the art will appreciate that in some embodiments, the functionality of processors 300 or 320 executing their respective applications may be implemented using pre-programmed hardware or firmware elements (e.g., application specific integrated circuits (ASICs), electrically erasable programmable read-only memories (EEPROMs), etc.), or other related components.
  • Referring now to FIG. 6, an AGV 600 is illustrated according to certain non-limiting embodiments. The AGV 600 is shown as a forklift, though other human-operated material-transport vehicles may also be used, such as tuggers and pallet trucks/jacks.
  • The forklift as depicted is considered as an AGV 600 for the purposes herein based on the inclusion of a control module (not shown in FIG. 6) and a navigation sensor 624. According to some embodiments, the AGV 600 may comprise a human-operated forklift (or other material-transport vehicle) that has been retrofitted with equipment to convert the forklift to an AGV. In some embodiments, the original equipment of the forklift may include the AGV components (e.g. control module, navigation sensors, etc.).
  • According to some embodiments, the navigation sensor 624 may be a magnetic field sensor for detecting a strip of magnetic tape 628 placed on the floor of the facility. As previously described, for example in reference to the AGV 100 in FIG. 1, the control module is configured to control a drive mechanism of the forklift to correct deviations in the position of the tape 628 from the center of the sensor 124. In other words, the control module is configured to guide the AGV 600 along the length of the tape 628.
  • The AGV 600 further includes an augmentation module 704 coupled to the forklift. The augmentation module 704 may be coupled to the AGV 600 in any suitable location. According to some embodiments, the AGV 600 can include additional components such as a differential drive rotation sensor (e.g. coupled to the drive mechanism of the forklift), and one or more automation interfaces such as a power data interface for providing the augmentation module 704 with data defining a remaining power level (e.g. battery charge) for the AGV 600. According to some embodiments, the AGV 600 can include additional sensors, either mounted on the forklift or on the augmentation module 704. For instance, the augmentation module 704 can include auxiliary navigation sensors (e.g. additional LiDAR sensors, cameras, GPS sensors, and the like).
  • The scope of the claims should not be limited by the embodiments set forth in the above examples, but should be given the broadest interpretation consistent with the description as a whole.

Claims (21)

1.-20. (canceled)
21. An augmentation system for an automated guided vehicle (AGV) deployed in a facility, the system comprising:
a memory storing a map comprising a digital representation of the facility, wherein at least a portion of the map is generated based on first sensor data;
one or more sensors, wherein at least one sensor of the one or more sensors generates second sensor data, the at least one sensor being a different type than a first sensor that generated the first sensor data; and
a processor coupled to the memory and each of the one or more sensors, the processor configured for:
obtaining an operational command for the AGV; and
operating the AGV to execute the operational command based on the map, wherein during execution of the operational command:
determining, with reference to the map, a current location of the AGV within the facility;
receiving, from the one or more sensors, the second sensor data in respect of an area surrounding the current location;
analyzing the second sensor data to determine one or more features in the surrounding area;
comparing the one or more features to information included in the map in respect of the surrounding area;
based on the comparison, determining a discrepancy between the second sensor data and information included in the map; and
updating the map based on the second sensor data.
22. The system of claim 21, wherein the map comprises an indication of a location of a guiding infrastructure element in the facility, and determining the current location comprises:
detecting, using the one or more sensors, the guiding infrastructure element in proximity of the current location; and
localizing the AGV with reference to the position of the guiding infrastructure element in the map.
23. The system of claim 22, wherein the guiding infrastructure element comprises at least one of magnetic tape, coloured tape, a radio-frequency identification (RFID) tags, or a reflective marker.
24. The system of claim 21, wherein the one or more features correspond to an unmapped portion of the facility, and the map is updated to include the unmapped portion.
25. The system of claim 21, wherein the one or more features correspond to a structure in the facility, and the discrepancy corresponds to a change in position of the structure, and wherein the map is updated to include the changed position of the structure.
26. The system of claim 21, wherein the one or more sensors are selected from a group comprising: a camera, a magnetic field sensor, a radio-frequency identification (RFID) reader and a range-finding sensor.
27. The system of claim 21, wherein the operational command comprises at least one of:
a command for the AGV to travel to a target location identified within the map;
a command for the AGV to follow a predefined path; or
a command to execute an action.
28. The system of claim 21, wherein the operational command is received from a pilot-operated computing device.
29. The system of claim 21, wherein the processor is further configured for:
generating control data defining one or more operations for the AGV to execute the operational command;
converting the control data to simulated sensor data, wherein the simulated sensor data corresponds to the second sensor data that would have been regenerated by the at least one sensor in the presence of guidance infrastructure; and
transmitting the simulated sensor data to a control module of the AGV, wherein the control module controls a drive mechanism of the AGV based on the simulated sensor data.
30. The system of claim 21, wherein:
the at least one sensor comprises a first navigation sensor and a second navigation sensor, the first navigation sensor being a same type as the first sensor, the second navigation sensor being a different type than the first sensor; and
the processor is further configured for:
generating control data defining one or more operations for the AGV to execute the operational command, the control data being generated based on the portion of the map generated based on the first sensor data;
converting the control data to simulated sensor data, wherein the simulated sensor data corresponds to sensor data that would have been generated by the first navigation sensor in the presence of guidance infrastructure; and
transmitting the simulated sensor data to a control module of the AGV, wherein the control module controls a drive mechanism of the AGV based on the simulated sensor data.
31. A method for navigating an automated guided vehicle (AGV) deployed in a facility, the AGV being coupled to an augmentation system including a processor coupled to one or more sensors and a memory, the method comprises:
obtaining, using the processor, an operational command for the AGV; and
operating, using the processor, the AGV to execute the operational command, wherein during execution of the operational command:
determining, with reference to a map of the facility, a current location of the AGV within the facility, wherein the map comprises a digital representation of the facility, and at least a portion of the map is generated based on first sensor data;
receiving, from the one or more sensors, second sensor data generated by at least one sensor of the one or more sensors, the second sensor data being in respect of an area surrounding the current location, the at least one sensor being a different type than a first sensor that generated the first sensor data;
analyzing the second sensor data to determine one or more features in the surrounding area;
comparing the one or more features to information included in the map in respect of the surrounding area;
based on the comparison, determining a discrepancy between the second sensor data and information included in the map; and
updating the map to based on the second sensor data.
32. The method of claim 31, wherein the map comprises an indication of a location of a guiding infrastructure element in the facility, and determining the current location further comprises:
detecting, using the one or more sensors, the guiding infrastructure element in proximity of the current location; and
localizing the AGV with reference to the position of the guiding infrastructure element in the map.
33. The method of claim 32, wherein the guiding infrastructure element comprises at least one of magnetic tape, coloured tape, a radio-frequency identification (RFID) tags, or a reflective marker.
34. The method of claim 31, wherein the one or more features correspond to an unmapped portion of the facility, and the map is updated to include the unmapped portion.
35. The method of claim 31, wherein the one or more features correspond to a structure in the facility, and the discrepancy corresponds to a change in position of the structure, and wherein the map is updated to include the changed position of the structure.
36. The method of claim 31, wherein the one or more sensors are selected from a group comprising: a camera, a magnetic field sensor, a radio-frequency identification (RFID) reader and a range-finding sensor.
37. The method of claim 31, wherein the operational command comprises at least one of:
a command for the AGV to travel to a target location identified within the map;
a command for the AGV to follow a predefined path; or
a command to execute an action.
38. The method of claim 31, wherein the operational command is received from a pilot-operated computing device.
39. The method of claim 31, further comprising:
generating control data defining one or more operations for the AGV to execute the operational command;
converting the control data to simulated sensor data, wherein the simulated sensor data corresponds to the second sensor data that would have been regenerated by the at least one sensor in the presence of guidance infrastructure; and
transmitting the simulated sensor data to a control module of the AGV, wherein the control module controls a drive mechanism of the AGV based on the simulated sensor data.
40. The method of claim 31, wherein the at least one sensor comprises a first navigation sensor and a second navigation sensor, the first navigation sensor being a same type as the first sensor, the second navigation sensor being a different type than the first sensor; and
the method further comprises:
generating control data defining one or more operations for the AGV to execute the operational command, the control data being generated based on the portion of the map generated based on the first sensor data;
converting the control data to simulated sensor data, wherein the simulated sensor data corresponds to sensor data that would have been generated by the first navigation sensor in the presence of guidance infrastructure; and
transmitting the simulated sensor data to a control module of the AGV, wherein the control module controls a drive mechanism of the AGV based on the simulated sensor data.
US17/462,956 2016-02-11 2021-08-31 Control augmentation apparatus and method for automated guided vehicles Pending US20220055877A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/462,956 US20220055877A1 (en) 2016-02-11 2021-08-31 Control augmentation apparatus and method for automated guided vehicles

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US201662293948P 2016-02-11 2016-02-11
US15/429,222 US10662045B2 (en) 2016-02-11 2017-02-10 Control augmentation apparatus and method for automated guided vehicles
US16/855,788 US11167964B2 (en) 2016-02-11 2020-04-22 Control augmentation apparatus and method for automated guided vehicles
US17/462,956 US20220055877A1 (en) 2016-02-11 2021-08-31 Control augmentation apparatus and method for automated guided vehicles

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/855,788 Continuation US11167964B2 (en) 2016-02-11 2020-04-22 Control augmentation apparatus and method for automated guided vehicles

Publications (1)

Publication Number Publication Date
US20220055877A1 true US20220055877A1 (en) 2022-02-24

Family

ID=59559518

Family Applications (3)

Application Number Title Priority Date Filing Date
US15/429,222 Active 2037-11-17 US10662045B2 (en) 2016-02-11 2017-02-10 Control augmentation apparatus and method for automated guided vehicles
US16/855,788 Active US11167964B2 (en) 2016-02-11 2020-04-22 Control augmentation apparatus and method for automated guided vehicles
US17/462,956 Pending US20220055877A1 (en) 2016-02-11 2021-08-31 Control augmentation apparatus and method for automated guided vehicles

Family Applications Before (2)

Application Number Title Priority Date Filing Date
US15/429,222 Active 2037-11-17 US10662045B2 (en) 2016-02-11 2017-02-10 Control augmentation apparatus and method for automated guided vehicles
US16/855,788 Active US11167964B2 (en) 2016-02-11 2020-04-22 Control augmentation apparatus and method for automated guided vehicles

Country Status (1)

Country Link
US (3) US10662045B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI823703B (en) * 2022-12-06 2023-11-21 中華汽車工業股份有限公司 Auxiliary positioning method for unmanned vehicles

Families Citing this family (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
PL3167342T3 (en) * 2015-09-22 2018-11-30 Bluebotics Sa Virtual line-following and retrofit method for autonomous vehicles
WO2018039337A1 (en) 2016-08-23 2018-03-01 Canvas Technology, Inc. Autonomous cart for manufacturing and warehouse applications
NL2017784B1 (en) * 2016-11-14 2018-05-25 Ravas Europe B V Pallet truck
US11237545B2 (en) * 2017-01-09 2022-02-01 Debashis Das System and method for converting manual industrial machines into automatic industrial machines
US10585440B1 (en) 2017-01-23 2020-03-10 Clearpath Robotics Inc. Systems and methods for using human-operated material-transport vehicles with fleet-management systems
US11760221B2 (en) 2017-06-27 2023-09-19 A9.Com, Inc. Charging systems and methods for autonomous carts
US10793369B2 (en) 2017-07-12 2020-10-06 A9.Com, Inc. Conveyor system for autonomous robot
JP6725063B2 (en) * 2017-11-28 2020-07-15 Jfeスチール株式会社 Equipment management system
CN109982013A (en) * 2017-12-28 2019-07-05 沈阳新松机器人自动化股份有限公司 A kind of AGV colour band navigation sensor circuit
US11305936B2 (en) * 2018-09-28 2022-04-19 Autoguide, LLC Commercial logistic facility, configurable modular robotic autonomous guided vehicle, and method therefor
US10994865B2 (en) 2018-10-22 2021-05-04 Goodrich Corporation Autonomous system for air cargo end-to-end operations
CN109911819B (en) * 2019-03-22 2021-06-22 杭州昱透实业有限公司 Balance weight type AGV stacker
CN110040194A (en) * 2019-05-20 2019-07-23 山东洛杰斯特物流科技有限公司 Steering wheel and omnidirectional's AGV fork truck based on ring gear transmission
DE102019123659A1 (en) 2019-09-04 2021-03-04 Sick Ag Method for creating a map, method for determining a pose of a vehicle, mapping devices and localization devices
CN110789606A (en) * 2019-10-16 2020-02-14 合肥搬易通科技发展有限公司 Automatic guiding method for controlling driving direction and position based on all-wheel speed detection and forklift system thereof
DE102020206276A1 (en) * 2020-05-19 2021-11-25 Continental Teves Ag & Co. Ohg Driverless transport vehicle with improved ease of maintenance
US11433920B2 (en) 2020-05-29 2022-09-06 Robert Bosch Gmbh Map-based prediction and mitigation of performance limitations for autonomous vehicles
US11312393B2 (en) * 2020-05-29 2022-04-26 Robert Bosch Gmbh Artificially falsifying sensor data to initiate a safety action for an autonomous vehicle
CN112633804A (en) * 2020-12-25 2021-04-09 南京紫津融畅信息科技服务有限公司 AGV-based vault management method and system
US20220274817A1 (en) * 2021-02-26 2022-09-01 Hyster-Yale Group, Inc. Structural integrated sensor
WO2023154422A1 (en) * 2022-02-09 2023-08-17 Cyngn, Inc. Modular sensor system for automated guided vehicles

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009031884A (en) * 2007-07-25 2009-02-12 Toyota Motor Corp Autonomous mobile body, map information creation method in autonomous mobile body and moving route specification method in autonomous mobile body
US20100063680A1 (en) * 2008-09-11 2010-03-11 Jonathan Louis Tolstedt Leader-follower semi-autonomous vehicle with operator on side
EP2385435A1 (en) * 2010-04-22 2011-11-09 Danaher Motion Särö AB A method and a system for gathering data
US20130054129A1 (en) * 2011-08-26 2013-02-28 INRO Technologies Limited Method and apparatus for using unique landmarks to locate industrial vehicles at start-up
US20140188324A1 (en) * 2011-08-29 2014-07-03 Crown Equipment Corporation Vehicular navigation control interface
US20170351261A1 (en) * 2015-11-04 2017-12-07 Zoox, Inc. Sensor-Based Object-Detection Optimization For Autonomous Vehicles
US20180190046A1 (en) * 2015-11-04 2018-07-05 Zoox, Inc. Calibration for autonomous vehicle operation

Family Cites Families (53)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5023790A (en) 1989-02-17 1991-06-11 Whs Robotics Automatic guided vehicle system
US20080183599A1 (en) 1998-11-18 2008-07-31 Visible Inventory, Inc. Inventory control and communication system
JP3484104B2 (en) 1998-12-25 2004-01-06 平田機工株式会社 Automatic warehouse and automatic warehouse management method
JP2001121461A (en) 1999-10-26 2001-05-08 Denso Corp Robot system
KR100640105B1 (en) 2001-04-19 2006-10-30 무라타 기카이 가부시키가이샤 Automated guided vehicle, automated guided vehicle system and wafer conveyance method
SE526913C2 (en) 2003-01-02 2005-11-15 Arnex Navigation Systems Ab Procedure in the form of intelligent functions for vehicles and automatic loading machines regarding mapping of terrain and material volumes, obstacle detection and control of vehicles and work tools
WO2005098729A2 (en) 2004-03-27 2005-10-20 Harvey Koselka Autonomous personal service robot
US7962192B2 (en) 2005-09-30 2011-06-14 Restoration Robotics, Inc. Systems and methods for aligning a tool with a desired location or object
US9436184B2 (en) 2006-06-09 2016-09-06 Amazon Technologies, Inc. Method and system for transporting inventory items
US7873469B2 (en) 2006-06-19 2011-01-18 Kiva Systems, Inc. System and method for managing mobile drive units
US7693757B2 (en) 2006-09-21 2010-04-06 International Business Machines Corporation System and method for performing inventory using a mobile inventory robot
JP4576445B2 (en) * 2007-04-12 2010-11-10 パナソニック株式会社 Autonomous mobile device and program for autonomous mobile device
JP5047709B2 (en) 2007-07-04 2012-10-10 株式会社日立製作所 Moving device, system, moving method, and moving program
US8190295B1 (en) * 2008-05-14 2012-05-29 Sandia Corporation Apparatus and method for modifying the operation of a robotic vehicle in a real environment, to emulate the operation of the robotic vehicle operating in a mixed reality environment
TWM348676U (en) 2008-07-22 2009-01-11 Iner Aec Executive Yuan Environmental survey robot
US7972102B2 (en) 2008-07-24 2011-07-05 Marine Terminals Corporation Automated marine container terminal and system
US8909466B2 (en) 2008-08-01 2014-12-09 Environmental Systems Research Institute, Inc. System and method for hybrid off-board navigation
JP5141507B2 (en) 2008-08-25 2013-02-13 村田機械株式会社 Autonomous mobile device
US8126642B2 (en) 2008-10-24 2012-02-28 Gray & Company, Inc. Control and systems for autonomously driven vehicles
TWI615337B (en) 2009-04-10 2018-02-21 辛波提克有限責任公司 Automated case unit storage system and method for handling case units that are configured for being arrayed into a palletized load of case units for shipping to or from a storage facility
FI20095713A (en) 2009-06-24 2010-12-25 Sandvik Mining & Constr Oy Determination of driving route for arranging automatic control of a moving mining machine
GB2494081B (en) 2010-05-20 2015-11-11 Irobot Corp Mobile human interface robot
US8930019B2 (en) 2010-12-30 2015-01-06 Irobot Corporation Mobile human interface robot
US8718837B2 (en) 2011-01-28 2014-05-06 Intouch Technologies Interfacing with a mobile telepresence robot
US20130086215A1 (en) 2011-05-13 2013-04-04 HNTB Holdings, Ltd. Managing large datasets obtained through a survey-data-acquisition process
US8886359B2 (en) 2011-05-17 2014-11-11 Fanuc Corporation Robot and spot welding robot with learning control function
US8583361B2 (en) 2011-08-24 2013-11-12 Modular Mining Systems, Inc. Guided maneuvering of a mining vehicle to a target destination
DE102012003690A1 (en) 2012-02-23 2013-08-29 Kuka Roboter Gmbh Mobile robot
US9463574B2 (en) 2012-03-01 2016-10-11 Irobot Corporation Mobile inspection robot
US20140040431A1 (en) 2012-08-06 2014-02-06 General Electric Company Systems and methods for an opc ua server
US9125779B2 (en) 2012-09-28 2015-09-08 Elwha Llc Automated systems, devices, and methods for transporting and supporting patients
US9804576B2 (en) 2013-02-27 2017-10-31 Rockwell Automation Technologies, Inc. Recognition-based industrial automation control with position and derivative decision reference
US8965561B2 (en) 2013-03-15 2015-02-24 Cybernet Systems Corporation Automated warehousing using robotic forklifts
US9141107B2 (en) 2013-04-10 2015-09-22 Google Inc. Mapping active and inactive construction zones for autonomous driving
JP6117368B2 (en) 2013-10-11 2017-04-19 株式会社日立製作所 Transport vehicle control device and transport vehicle control method
US9452531B2 (en) 2014-02-04 2016-09-27 Microsoft Technology Licensing, Llc Controlling a robot in the presence of a moving object
US9465388B1 (en) * 2014-03-03 2016-10-11 Google Inc. Remote assistance for an autonomous vehicle in low confidence situations
US9720410B2 (en) * 2014-03-03 2017-08-01 Waymo Llc Remote assistance for autonomous vehicles in predetermined situations
US9486917B2 (en) 2014-04-30 2016-11-08 The Boeing Company Mobile automated assembly tool for aircraft structures
US9996976B2 (en) 2014-05-05 2018-06-12 Avigilon Fortress Corporation System and method for real-time overlay of map features onto a video feed
US9280153B1 (en) 2014-09-19 2016-03-08 Amazon Technologies, Inc. Inventory holder load detection and/or stabilization
US9824592B2 (en) 2014-09-22 2017-11-21 Vinveli Unmanned Systems, Inc. Method and apparatus for ensuring the operation and integrity of a three-dimensional integrated logistical system
WO2016061255A1 (en) 2014-10-14 2016-04-21 Harvest Automation, Inc. Storage material handling system
US9487356B1 (en) 2015-03-02 2016-11-08 Amazon Technologies, Inc. Managing low-frequency inventory items in a fulfillment center
US9649766B2 (en) 2015-03-17 2017-05-16 Amazon Technologies, Inc. Systems and methods to facilitate human/robot interaction
EP3985353A1 (en) * 2015-05-06 2022-04-20 Crown Equipment Corporation Industrial vehicle with tag reader
US9682481B2 (en) 2015-10-26 2017-06-20 X Development Llc Communication of information regarding a robot using an optical identifier
US9632502B1 (en) * 2015-11-04 2017-04-25 Zoox, Inc. Machine-learning systems and techniques to optimize teleoperation and/or planner decisions
CN114671025A (en) 2016-04-29 2022-06-28 美国联合包裹服务公司 Method and system for loading, picking up and delivering packages from vehicles by unmanned aerial vehicle
KR102565501B1 (en) 2016-08-01 2023-08-11 삼성전자주식회사 A robotic cleaner, a refrigerator, a system of delivery of a container and a method of delivering and retrieving of a container of the refrigerator using the robotic cleaner
US10317119B2 (en) 2016-08-25 2019-06-11 Amazon Technologiess, Inc. Transportable climate-controlled units for fulfillment of perishable goods
US11097422B2 (en) 2017-02-07 2021-08-24 Veo Robotics, Inc. Safety-rated multi-cell workspace mapping and monitoring
CA3070625A1 (en) 2017-07-28 2019-01-31 Nuro, Inc. System and mechanism for upselling products on autonomous vehicles

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2009031884A (en) * 2007-07-25 2009-02-12 Toyota Motor Corp Autonomous mobile body, map information creation method in autonomous mobile body and moving route specification method in autonomous mobile body
US20100063680A1 (en) * 2008-09-11 2010-03-11 Jonathan Louis Tolstedt Leader-follower semi-autonomous vehicle with operator on side
EP2385435A1 (en) * 2010-04-22 2011-11-09 Danaher Motion Särö AB A method and a system for gathering data
US20130054129A1 (en) * 2011-08-26 2013-02-28 INRO Technologies Limited Method and apparatus for using unique landmarks to locate industrial vehicles at start-up
US20140188324A1 (en) * 2011-08-29 2014-07-03 Crown Equipment Corporation Vehicular navigation control interface
US20170351261A1 (en) * 2015-11-04 2017-12-07 Zoox, Inc. Sensor-Based Object-Detection Optimization For Autonomous Vehicles
US20180190046A1 (en) * 2015-11-04 2018-07-05 Zoox, Inc. Calibration for autonomous vehicle operation

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Machine translation of JP-2009031884-A (Year: 2009) *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI823703B (en) * 2022-12-06 2023-11-21 中華汽車工業股份有限公司 Auxiliary positioning method for unmanned vehicles

Also Published As

Publication number Publication date
US20200346907A1 (en) 2020-11-05
US20170233231A1 (en) 2017-08-17
US10662045B2 (en) 2020-05-26
US11167964B2 (en) 2021-11-09

Similar Documents

Publication Publication Date Title
US11167964B2 (en) Control augmentation apparatus and method for automated guided vehicles
US11892853B2 (en) Vehicle guidance systems and associated methods of use at logistics yards and other locations
US20220357174A1 (en) Stand-alone self-driving material-transport vehicle
US20120123614A1 (en) Method and apparatus for virtualizing industrial vehicles to automate task execution in a physical environment
US10725471B2 (en) Virtual line-following and retrofit method for autonomous vehicles
JP2019168942A (en) Moving body, management device, and moving body system
CA2961938A1 (en) Systems and methods for moving pallets via unmanned motorized unit-guided forklifts
US20200264616A1 (en) Location estimation system and mobile body comprising location estimation system
US20220371199A1 (en) System and method for connection of service lines to trailer fronts by automated trucks
KR102028346B1 (en) Following cart
US20200363212A1 (en) Mobile body, location estimation device, and computer program
US20200165083A1 (en) Transporting device, transporting system, and shelf transporting method
CN110884488A (en) Auxiliary positioning system for automatic driving engineering vehicle and using method thereof
JP2022518012A (en) Autonomous broadcasting system for self-driving cars
CN109081272A (en) It is a kind of to mix the unmanned transhipment fork truck guided and method with vision based on laser
US11537140B2 (en) Mobile body, location estimation device, and computer program
JPWO2019054209A1 (en) Map making system and map making device
WO2019059307A1 (en) Moving body and moving body system
JP2019175137A (en) Mobile body and mobile body system
JP2019179497A (en) Moving body and moving body system
US20210080960A1 (en) Methods and systems for shifting objects
KR102446517B1 (en) Auto guided vehicle capable of autonomous driving in indoor and outdoor environments
WO2019069921A1 (en) Mobile body
US20230195125A1 (en) Method for controlling an autonomous robotic tool
KR20240054014A (en) Control method and apparatus for smart factory

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: CLEARPATH ROBOTICS INC., CANADA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:GARIEPY, RYAN CHRISTOPHER;DOBSON, ANDREW;TEBBS, JESSE;AND OTHERS;SIGNING DATES FROM 20160218 TO 20160301;REEL/FRAME:059168/0001

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER