CA1277392C - Target and control system for positioning an automatically guided vehicle - Google Patents

Target and control system for positioning an automatically guided vehicle

Info

Publication number
CA1277392C
CA1277392C CA000520829A CA520829A CA1277392C CA 1277392 C CA1277392 C CA 1277392C CA 000520829 A CA000520829 A CA 000520829A CA 520829 A CA520829 A CA 520829A CA 1277392 C CA1277392 C CA 1277392C
Authority
CA
Canada
Prior art keywords
images
light source
reflector elements
target member
camera
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CA000520829A
Other languages
French (fr)
Inventor
Harry B. Hammill, Iii
Frank J. Lukowski, Jr.
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Caterpillar Inc
Original Assignee
Caterpillar Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US06/788,989 external-priority patent/US4684247A/en
Priority claimed from US06/789,280 external-priority patent/US4678329A/en
Application filed by Caterpillar Inc filed Critical Caterpillar Inc
Application granted granted Critical
Publication of CA1277392C publication Critical patent/CA1277392C/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0212Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory
    • G05D1/0225Control of position or course in two dimensions specially adapted to land vehicles with means for defining a desired trajectory involving docking at a fixed facility, e.g. base station or loading bay
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B66HOISTING; LIFTING; HAULING
    • B66FHOISTING, LIFTING, HAULING OR PUSHING, NOT OTHERWISE PROVIDED FOR, e.g. DEVICES WHICH APPLY A LIFTING OR PUSHING FORCE DIRECTLY TO THE SURFACE OF A LOAD
    • B66F9/00Devices for lifting or lowering bulky or heavy goods for loading or unloading purposes
    • B66F9/06Devices for lifting or lowering bulky or heavy goods for loading or unloading purposes movable, with their loads, on wheels or the like, e.g. fork-lift trucks
    • B66F9/075Constructional features or details
    • B66F9/0755Position control; Position detectors
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/02Systems using the reflection of electromagnetic waves other than radio waves
    • G01S17/06Systems determining position data of a target
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/87Combinations of systems using electromagnetic waves other than radio waves
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S17/00Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
    • G01S17/88Lidar systems specially adapted for specific applications
    • G01S17/89Lidar systems specially adapted for specific applications for mapping or imaging
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course or altitude of land, water, air, or space vehicles, e.g. automatic pilot
    • G05D1/02Control of position or course in two dimensions
    • G05D1/021Control of position or course in two dimensions specially adapted to land vehicles
    • G05D1/0231Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means
    • G05D1/0234Control of position or course in two dimensions specially adapted to land vehicles using optical position detecting means using optical markers or beacons
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B2219/00Program-control systems
    • G05B2219/30Nc systems
    • G05B2219/37Measurements
    • G05B2219/37555Camera detects orientation, position workpiece, points of workpiece
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B2219/00Program-control systems
    • G05B2219/30Nc systems
    • G05B2219/45Nc applications
    • G05B2219/45049Forklift

Abstract

Abstract of the Disclosure Target and Control System for Positioning an Automatically Guided Vehicle A target member mounted on a stationary object, such as a pallet, includes at least three reflector elements. An identifying means, such as a high intensity light source, and an imaging sensor are carried by another, movable object, such as a forklift truck. The reflector elements are so configured as to form images of the identifying means with said images defining a plane oriented other than normal to align from the identifying means to that plane, the images also defining a circle that does not include the identifying means. The target member may be in the form of a vertically oriented planar support member on which are mounted a pair of convex mirrors and a concave mirror. The images of the identifying means in the mirrors are detected by an imaging sensor, such as a television camera, and the directions of each of the images at the camera are used to determine all six degrees of positional information of the sensor with respect to the target member. That information may be used to guide the forklift truck into position relative to a pallet. The target support member includes retroreflector elements to aid in locating a pallet.
Circuit means are provided to flash the light source and in response to the light returned by the retroreflectors to define an area on the image plane of the camera where reflections from the mirrors may be found. Ambient light images are then temporarily recorded, the light source again flashed, and the ambient images are electronically removed so that only those images due to reflections in the mirrors are finally recorded.

Description

~,%~39~
Description Target and Control System for Po5_ tioning an Automatically ~uided Ve icle Background of the Invention This invention relates to method and apparatus for determining the position and orientation of a taryet with respect to a sensor. Both the target and the sensing unit may be affixed to movable objects, although generally the target will be skationary, and the sensing unit will be attached to a self-propelled vehicle.
There are many prior art devices which attempt to determine the location of a vehicle with respect to a target. In these devices, the vehicle carries a sensing unit, the output of which controls some function of the vehicle, such as its forward motion, ~ steering control or the vertical location of the ;~ forks. The target might be the pallet itself or the - frame or rack on which ~he pallet is supported. Some of the prior art devices employ specialized marks whose dimensions are known; others utilize special light projectors and sense the light reflected from the target for positioning a vehicle, or a component of the vehicle, such as the Eorks of a lift truck.
When using a plurality of marks located on the target, it is difficult, and becomes close to impossible, to determine accurately all six degrees of freedom, that is, horizontal position right and left, horizontal position forward and back, vertical position, roll, pitch and yaw. These prior ar-t systems do not give adequate positional information, especially when the vehicle is positioned directly in front of the target, to guide the vehicle accurately.

;

~ ~7'739%
68~97-88 Summary of the Inventi~on This invention is directed to a target member for use in a positioning system for providing at least three identiflable images positioned with respect to a sensor carried by a self-propelled vehicle. These three images are located so as to provide an unambiguous frame of reference thereby to allow the determination of all six positional and orientatlonal degrees of freedom from a single observation of the target by the sensor.
By translating or rotating the plane of the images from what has traditionally been normal viewing by the sensor to a plane which is essentially parallel to a line from the sensor to the target, positlonal ambiguities are resolved and accurate information regarding the relative location of the vehicle to the target is obtainable, even when the vehicle is located directly in front of the $arget.
In one aspect of the present invention, a target member for use in a positionlng system is provided. The target member has a support member. At least three reflector elements are mounted on the support member, each of the reflector elements being so configured as to form an image of a light source located ~7ithin a predetermined field of view with respect to the support member. The images define a plane oriented other than normal to a line from the light source to the plane. The images also define a circle that does not include the light source.

'7'73.~

In another aspect of khe present inventlon, an apparatus for usa in a docking system in which a movin~ vehicle is ~o be aligned with and closed to a dock position is provided. A target member is mounted at the dock position. The target member includes a generally ~lat support member and at least three non-planar refle~ting surfaces supported on the member in predetermined alignment. At least two of the rePlectlng surfaces both are either concave or convex and the thlrd surface is opposite to the two. The reflections from a common radiatlon source are perceived at known spacing and same sense in the two surfaces and at known spacing and inverted sense wlth respect to the others in the third surface.
In yet another aspect of the present invention, an apparatus for use in determining the location of a vehicle within a predetermined field of view with respect to an object is provided. A light source is carried by a vehicle. A target member is mounted on the object. The target member includes a support member and at least three reflector elements mounted on the support member ln a predetermined alignment. The reflector elements together form images o~ the identifying device in an image plane oriented other than normal to a line from the identifying device ~o the plane when the identifying device is located within a predeterminad ~ield of view wlth respect to the support member. A camera is carried by the vehicle for sensing ~ ;~7739~

the directlons of the identifylng device from each of the refleeting elements.
In another aspect of the present invention, an apparatus for use in determining the relative location of a vehicle with respect to a target member is provided. A light source and a camera is moun~ed on the vehicle. The target memher includes at least three reflector elements, each selected to form an image o~
the light source, and all the images together forms an image plane oriented other than normal to a line from the light source to the target member. A video RAM device records those images on the image plane of the camera due to ambient light. Thereafterr a timing logic circuit flashes the light source. A selection circuit compares the images due to ambient light ~tored in the recordlng means and the images resulting both from ambient light and from the light source. A device records those images on the ; image plane of the camera due only to reflections of the light source.
In yet another aspect of the present invention, a method for gathering information for determining the relative location of a target member with respect to a vehicle wherein the vehicle is provided with a light source and a camera, and wherein the target member is provided with at least three reflector elements, each selected to form an image of the light source, and all the imagas together forming an image plane oriented other than normal to a . ~

~.~77392 68297-8~3 line from the light source to the target member is provlded. The method includes the steps of: a) recording the images formed on the image plane of the camera due to ambient light; b) flashing the light source; c) comparing the images recorded ln step a) with the images resulting both from ambient light and from the light source thereby to isolate those reflections of the light source in the reflector elements; d) recording those imageY of the reflections due to the light source; and e) identifyiny the position of each such reflection of the light source from the reflector elements on the image plane of the camera and with respect to each other.
Advantages of the invention will be apparent from the following description, the accompanying drawingæ and the appended claims.
Brief Description of the Drawinas Fig. 1 is a perspective view showing a rack capable of ~ supporting a plurality of pallets, and an automatically gulded -~ vehicle carrying a light source and a camera for sensing the reflections of the light source in a re~lector member attached to a selected pallet;

~ ~7739,~

Fig. 2 iS a perspective view showing the camera and light sollrce located below and to the right of a line passing perpendicular to and through the center of the target or reflector member;
Fig. 3A is a view showing the reflec-tions of the light source in the reElector elements. Fig. 3~ is a view showing the reflections from the mirrors on the target as they appear on the image plane oE the camera;
Fig. 4 is a perspecti.ve view showing the camera and light source mounted on the forklift vehicle;
Fig. 5 illustrates a target or reflector member with attached retroreE:Lector members, spherical reflector elements, and a bar code;
~ ig. 6 is a plan view showing various locations of an automatically guided vehicle, such as a fork lift, with respect to a pallet;
Figs. 7A-7D represent the reflections of the light source in the re~lector elements at the various locations of the vehicle with respect to the pallet, as shown in Fiy. 6;
Figs. 8A-8D represents the images of the reflections sho~n in Figs. 7A-7D as they appear on the image plane of the vehicle carried camera;
Figs. 9A-9D represent the electrical signals on the image plane of the sensor at a single location of the vehicle. Fig. 9A shows the signals due to the images of the retroreflectors as a result of the ~irst flash of the light source. Fig. 9B shows the signals resulting from ambient light. Fig. 9C shows the signals when the light source is flashed a second time. Fig. 9D represents the electrical signals that remain after processing;
Fig. 10 is a vector diagram illustrating the directional relationship between a first image point P
and the final image point P'. The illustrated vector is located at the nodal point of the sensor lens;

~7 Fig. 11 is a block diagram of the video processing circuit used to identify and locate images of the retroreflectors and other reflector elements; and Figsu 12A-12C are timing diagrams showing the various signals that occur during various times during the operation of this invention.

Description of the Preferred Embodiment Referring now to the drawings now which illustated pre~erred embodiment of this invention, and particularly to Fig. 1, a storage rack 10 is shown supporting an object such as a pallet 15. The pallet 15 is provided with a target member 20.
A wooden pallet 15 is shown, and the target member 20 is illustrated as being a separate component attached to the palle-t. It should be understood, however, that any type of pallet structure may be used, and the target member 20 may either be a separa~e unit or it may be formed integral with the pallet itself~
A vehicle 30, such as a forklift truck, carries the identification means 35 (Figs. 2 and 4), such as a high intensity light source, and an imaging sensing means 40, which is preferably a miniature TV
camera, suGh as a Sony~XC-37 CCD camera. The light source and camera are preferably mounted together as a unit 45, with the light source 35 immediately adjacent and above the camera lens ~Fig. 4). In the preferred embodiment, the vertical distance separating the light source and camera lens is approximately one inch~
The target member 20 is shown in more detail in Fig. 5, and it includes a generally flat support member 50, three reflector elements 52l 53, and 54, and three retroreflector members 62, 63, and 64. A bar code 71 for uniquely identifying the pallet may also be printed on or attached to the support member.

~T~ m~ ~

~8--The light source and camera unit 45 is preferably aligned with the direction of travel of the vehicle. It is possible, however, rotatably to mount the camera on the vehicle so that it ma~ scan through a large field of view, both horizontally and vertically.
If this were done, however, the camera would be associated with a drive uni-t and position indicating device so that the proper corrections would be considered when calculating the relative location oE
the target.
The pre~erred embodiment of the invention, as illustrated, employs two convex reflector elements 52 and 54, and one concave reflector element 53. Each o~
the reflector elements is spherical in shape, and all are horizontally arranged on ~he support member 50.
Both reflectors elements 52 and 54 have the same radius of curvature, and the radii of curvatures of all of -the elements and their diameters are selected to provide a reasonable field of view A (Fig. 6) such that a reflection of the identifying means 3S will be viewable by the camera as long as the vehicle is within the field of view. In the embodiment described, it is preferred to have a minimum field of view of + 10 from ~he mirror platen normal. Typical mirrors may be ; 25 approximately 1.5 inches in diameter and have a radius of curvature of 3 inches or greater. Mirrors 52 and 54 may be type 380 convex mirrors, and mirror 53 may be type 100 concave mirror, both manufac-tured by ~OLYN.
It should be emphasized that the identifying means 35, while preferably a brilliant light source, could also be any means that could be detected by ~ensing means 40. A brilliant xenon flash lamp has been found effective for this purpose.
Referring to Fig. 2, the light/camera unit 45 is shown positioned below and to the right of the center line 60 passing through the target 20, but ~27~g~

within the field oE view ~. Under these conditions, the reflectlon of the identifying means or light source 35 appear as images PA, P0, PB in mirrors 52, 53 and 54, respectively, as shown in Fig. 3. Since the mirrors are curved surfaces in the embodiment shown, the images PA and PB will appear toward the lower right portion of mirrors 52 and 54, and the image P0 will be toward the upper left in mirror 53.
Since the unit 45 is also facing essentially parallel to the center line 60, the images PAI, P0', PB' of the identifyin~ means would be grouped on the image plane 70 of the camera towards its upper left hand corner, as shown in Fig. 3B. (It will be assumed for the following illustration that the images formed on the image plane are not inverted or reversed.) The absolute location of the reflections, the spacing between the reflections, and the relative position of all the reflections will provide information su~ficient to determine from a single observation the location of the vehicle with respect to the pallet and the orientation or rotation of the ; pallet. ~s the location o~ the vehicle changes, the observed position o;E the identifying means or reflections on the image plane of the camera will also change, as will be explained.
It will be noted in Fig. 2 that the images PA, P0 and PB o~ the identifying means 35 in the reflec-tor elements 52-54 will define a plane 80, and these images will also define a circle 82. In this invention, the plane 80 is not normal ~r perpendicular to the center line 60 of the target 20; it is in fact essentially parallel to the upper surface of the pallet. Further, the circle 82 will not include the ; lens of the camera 40. These conditions are necessary ~7739~

if the reElections are to provide an unambiguous result when they are analyzed to determine the locatio~ of the vehicle relative to the target.
Referring now to Figs. 9A-9D, these figures represent the images appearing on the image plane of the camera 40 durincJ one sequence of operations necessary to gather positional in~ormation.
The preferred method of this invention provides for flashing the light source 35 and recording the positions of the reElections ~rom the retroreflector members 62, 63r and 64. The circuit for accomplishing this is shown in Fig. 11. These reflections are identified in Fig. 9A as reElections 162, 163, and 164, respectively. These reflections are easily identified because they each occupy a plurality of pixels on the image plane 70 of the camera since they are physically large components of the target 20 and since the retroreflectors return a large percentage of the light emitted ky the identifying means 35 back toward the source. For this reason, the effective sensitivity of the camera is reduced at this stage oE
the operation so that only the reflec~ions of the ~; retroreflectors are likely to be found at the image plane. Also, because of the high intensity of the reflected lightt there may be some blooming of the image. The positions of each oE the retroreflector images is recorded in memory means.
Microprocessor means 310 performs a calculation by reference to the positions of the 30 retroreflector images 162-164, and an area 200 is defined in which the reflections from the reflector elements 52-54 are likely to be ~oundO This defined area 200 may be located anywhere on the image plane of the camera and will vary in area in proportion to the separation of the vehicle from the target. In other ' . , 73~;~

words, the closer the vehicle is to the target, the more widely separated will be the images, and the defined area will consequently be larger.
Once the area 200 has been defined on the image plane, the effective sensitivity of the camera is increased, and all of the images within the defined area are recorded in a recording means (Fig. 9B).
These images are those resulting from ambient light and may include such re1ections 165 as overhead lights, specular reflections from l~etal objects within the field of view, and other light: sources.
The nex-t step is to red~ce the effective sensitivity of the camera and again flash the light source 35. As shown in Fig. 9C, this time the image plane will contain the image of the retroreflectors 162-164, the ambient reflections 165, and also reflections PA, PO and P~ of the light source in each of the reflector elements 52-54. All of the images within the defined area 200 are recorded.
Any images from the retroreflectors that bloom into the defined area are removed from memory, and the images recorded in Fig~ 9B are effectively subtracted from those in Fig. 9C, and what remains are images PA', PO' and PB', reflections of the light source 25 ln the reflector elements 52,54, as shown in Fig. 9D.
The center of each of these images PA', PO' and PB' are calculated and the video signals ~ from the camera image plane are evaluated in accordance ; with the procedure later defined.
As shown in ~ig. 5, the retroreflector members 62,64 surround the reflector elements 52-54. It should be apparent, however, that this physical arrangement is not absoluteLy necessary. The position of each retroreflector should be known so that an area in which the reflector elements are positioned can be deined.

` `~

, ~ , ' ' ' : ' , ' ~ z773~æ
~12-Also, the position of the retroreflector elements further define a second area 210 in which the image 170 of the bar code 71 may be found, and at some appropriate time during the analysis of the image, the bar code may be read to conEirm that the proper target is being approached.
It should also be apparent to those skilled in the art that the reflector elements 52-54 do not all have to be spherical in shape. A11 that is necessary is that the images of the identifying means carried by the vehicle be viewable by the camera. This means that one or more of the reflector elements could be a retroreflector. Using spherical mirrors, however, reduces the cost of the target and also provides relatively smaller images, images whose position can therefore be determined with a high degree of accuracy.
~ t is assumed in this description that the mirrors are evenly spaced and are horizontally aligned, and that the center line of the target, that is, the ~O center mirror, is the desired final position of the light/camera unit 45. It should be recognized, however, that any orientation of the mirrors and any position of the camera unit with respect to the target would be acceptable and would not depart from this invention. A11 that the control circuit would need is information regarding the final desired position of each of the reflections on the image plane. Because of convention, and for purposes of explanation, it will be assumed that the desired final position will be on the center line with the images equally spaced on the image plane and that the images are in a horizontal line.
The area between the retroreflectors is provided with a dark background in order to minimize random noise and false data.

, '~:

~Z773~;~

Referring now to Figs. 6, 7A-7D and 8A-8D, it is assumed that the light/camera uni-t 45 is in the same horizontal plane as the target, and the vehicle 30 is positioned to the right oE the center line 60 in location 1. The reflections of the light source are shown in Fig. 7A, and the images of the reflections on the image plane 70 of the camera 40 are shown in Fig.
8A and would be located in the left center of the image plane~ The images are close together and unequally spaced. The location of the images on the image plane and t~eir relative spacing are all important to the calculations for determining the vehicle's ~elative location with respect to the target.
Assuming the vehicle moves to location 2 in Fig. 6, the images of the reflections on the image plane will move apart (due to the closer proximity to the target) and they will also move toward the right side of the image plane 70, as shown in Fig. 8B (due to the change in the direction of the vehicle).
When the vehicle arrives at location 3, it will be seen in Fig. 8C that the image of the reflections on the image plane move toward the center and each reflection moves away from each other reflection. In this case, the cen-ter reflection is ~- 25 still closer to the left hand reflection because the vehicle is still not located on the center line of the target.
Finally, when the vehicle reaches location 4 in Fig. 6, the reflections of the light source will be centered in the mirrors, as shown in Fig. 7D, and the images on the image plane will be in the position shown in Fig. 8D.
Referring now to the block diagram of Fig. 11, both the light source 35 and the camera ~0 are connected to electronic control and processing circuits :

, ~Z~'7392 300. At the heart of this circuit is a microprocessor system 310 which provides control signals to control the flow of data to the remainder of the circuit. A
timing logic circuitr shown generally at 320, responds -to instructions from the microprocessor system 310 to control the light source or flash 35, the video information from the camera 40, and the way that information is processed and stored in the remainder of the circuit.
A random access memory or video RAM circuit 330 provides the means for recording the images on the image plane o~ the video camera 40.
A multiplexer 340, operating under control of the microprocessor 310 and timing logic circuit 320, transfers video information into the video RAM 330 through a serial-to-parallel converter 350, and out of the video RAM 330 through a parallel-to-serial converter 360~
A digital-to-analog (D/A) converter 370 responds to a digital signal from the microprocessor to establish a threshold level for the output of the video - camera unit, and that threshold level determines what ~ video information passes ~rom the camera 40 through a ;~ comparator circuit 380 into a selection circuit 390, which circuit includes a first AND gate 392, a second AND gate 394, and an exclusive OR gate 396~
Referring now to the timing diagram of Fig.
12A, a typical output of the camera unit 40 on the RS170A video line 301 is illustrated. In this diayram, each in~erval between vertical blanking pulses 302 represents one-half frame. For example, the interval designated 401 includes all oE the odd numbered lines of one frame or screen, while the interval 402 represents all of the even numbered lines. The intervals 401 and 402 together comprise one complete frame.

, ~ ' . ' ' ~ .~ ' .

~7739~

A power up signal is provided by the microprocessor system on line 316 to the timing logic 320 when the system is Eirst turned on to synchronize the microprocessor system with the timing logic S circuit. In turn, the timing logic circuit sends a reset pulse on line 321 to the video camera unit to initialize this device.
'rhe camera 40 provides a video output signal to the comparator circuit 380, and part of this output is a pulse 302 representing the vertical blanking interval. During each vertical blanking interval, the timing logic circuit provides a vertical sync pulse back to the microprocessor on line 324.
The microprocessor system 310, operating under a software program, controls the sequencing of operation o~ the entire system. In normal operation, the microprocessor establishes an initial threshold level for the camera by sending a digital value on the microprocessor bus 315 to the D/A converter 370. This threshold level, shown in Fig. 12A, as the dashed line 372, limits those signals that may pass through the comparator circuit 3~0. This initial level is set high enough that all reflections, except from the retroreflectors within the field o~ view oE the camera, will be ignored and will not be passed on to the selection circuit 390.
Ne~t, the microprocessor 310 sends a Elash enable signal 311 on line 312 to timing logic circuit 320. This signal e~tends through the vertical blanking interval 302. At the beginning of the vertical blanking interval, the microprocessor generates a strobe signal on line 313, and the timing log:ic circuit 320 in response thereto sends a flash trigger pulse 322 on line 323 at the beginning of interval 401 to the light source or strobe 35.

~Z'~73~;2 AS previously explained, the light source 35 is a high intensity xenon strobe which floods the area - in Eront of the camera with a short duration pulse of high intensity light. The threshold level 372 during intervals 401 and 402 is set high enough that only the video signals exceediny the predetermined threshold value, such as those representing the re~lections 162-164 returned by the retroreflectors 62-64, will be allowed to pass through the comparat.or 380. Although the duration of the flash may be measured in microseconds, the light energy of the re~lections therefrom will be retained on the camera image plane for one complete frame.
The video camera unit 40 provides an output on line 326 from the lnternal camera clock, basically a 3.58 MHz series of pulses, to the timing logic circuit 320. These clock pulses are converted by the timing logic circuit into pixel clock pulses on line 327, with each pixel clock pulse representing a single pixel as it appears on the camera image plane. As shown in FigsO 12A-12C, pixel clock pulses are provided to the serial-to-parallel converter 350 and the parallel-to-serial converter 360 for two intervals after a start pulse.
The camera, a Sony XC-37 CC~ Camera, has an image plane providing an array of 384 x 491 pixels.
Half of those pixels will be interrogated during the first interval and the other half during the second interval. Thus, each pixel on the camera image plane is separately and uniquely identified and it can be determined whether or not the output from each pixel exceeds the predetermined threshold level.
The output of the comparator 380 is applied on line 381 to selection circuit 39~, and to both ~he AN~
gate 392 and exclusive OR gate 396. During intervals ~J2773~2 401 and 402, the exclusive OR enable line 318 is low and therefore -the video information from the comparator 380 will be passed directly on line 391 into the serial-to-parallel converter 350.
Since the target 20 typically includes three retroreflectors in the preferred embodiment, it is expected that only three intense returns, or images 162, 163 and 164, will exceed the threshold 372, and therefore these images will be processed through the serial-to-parallel converter 350 and sent on the video R~ data bus 355, through the multiplexer 340 under control of signals provided by timing circuit 320 on bus 328, and into the video RAM 330 via bus 335 where they will be s~ored or recorded in electronic Eorm.
In a preferred embodiment of this invention, because the images of the retroreflectors are so large, it is possible to speed up the process by scaling the pixel clock and limiting the video data stored in the RAM 330 by storing every fourth pixel of every fourth line of each frame and still detect the presence of the retroreflectors.
These stored or recorded images of the retroreflectors are scanned by the microprocessor under software control so that their locations may be determined, and subsequent scanning or interrogation of the video R~ may then be limited to ~he area bounded by the retroreflectors thereby to speed up the identification and location of the images of the light source in the reflector elements~
Referring now to Fig. 12B, the threshold level is set by the microprocessor to that shown at 373 so that low level images 165 and 166, which exceed the threshold level during intervals 403 and 40~, ~ay pass through the comparator 380. The selection circuit 3g0 remains inactive at khis time, so those images will be ~Z'7'739;~

recorded in the video RAM 330 in accordance with the process described above. Thusl the images on the image plane of the camera due to ambient light are temporarily recorded for later reference.
Fig. 12C illustrates the next step in the process. The microprocessor 310 increases the threshold level slightly to that shown at 374 rrhis will permit most of the images due to ambient light to pass through the comparator 3~0, but will eliminate some that might have been marginal, such as image 166.
This will also tend to eliminate noise from the camera itself and slight changes in size of the image due to camera movement.
The flash is once again enabled and it is triggered a second time at the beginning of interval 405. This time, those signals representing images from ambient light 165, the images of the retroreflectors 162-164, and also the images PO, PA and PB of the light source in the reflector elements will be passed 20 by the comparator 380 into the selection circuit 390.
An exclusive 01~ enable signal on line 318 from the microprocessor 310 will now be present, however~ and the selection circuit 390 wiLl now be active and will prevent those signals representing ambient light images ~rom being passed through to the video RAM 330~
The selection circuit 390 provides the means for comparing the images due to ambient light temporarily stored in the recording means or memory 330 and the images resulting from both ambient light and from the light source, and for thereafter recording only those imayes due to reflections of the light source in the video RAM 330.
When the signal level on the exclusive OR
enable line 318 is raised (Fig. 12C at the beginning of 35 frame 405), AND gate 394 is enabled, the data 35~

previously recorded during .intervals 403 and 404 ~Fig.
lZB) will be taken from the video RAM 330 through the multiplexer 340 into the parallel-to-serial converter 360 via bus 345, the output of which on line 361 is applied a pixel at a time as the other input to the AND
gate 394. The parallel-to-serial converter 360 receives a latch signal on line 329 ~rom the timing circuit to enable the converter 360 to accept a block oE data from the video RAM ancl then to read that data out to the selection circuit 90 under direction of the pixel clock signal on line 327. Thus, for each pixel that contains information relating to ambient light, there will be an output from AND gate 394 to the exclusive OR gate 396.
All video signals on line 381 are applied both to AND gate 392 and to exclusive OR gate 396. Any such signal that does not have a coun-terpart in the video RAM, and thus an output on line 361, will be allowed to pass through AND gate 392 and thereafter be stored in the video RAM 330, replacing any information previously stored therein. On the other hand, any signal (image) ~; which previously appeared in the video RAM that does not have a counterpart on line 381 will be ignored.
: Thus, an image, such as represented by 166, which appeared and was allowed to pass into the video RAM
during frames 403 and 404, but not during rames 405 and 406 because the threshold level had been increased slightly, will be ignored.
At the end of frame 404 therefore, only those ~ 30 images appearing on the image plane of the camera which : are due to reflections of the light source in both the retroreflectors and the reflector elements will be recorded in video RAM 330.
Since the location of the retroreflector images is known and was determined during intervals 401 and 402, the location of the reflec~ions of the light 12~739X:

source in the reflector elements with respect to each other and with respect to the image p:Lane of the camera can now be accurately determined by analyzing the data stored in RAM 330 and the location of the vehicle in S relation to the target calculated.
The technique for calculating the relative positions of the vehicle and l:arget will now be dis-cussed.
One of the target points, the center one, PO, is chosen as a target reference point or origin.
The other two pointsl, PA and PB, have 3-dimensional vector offsets from PO. These oEfset vectors are identified as a and b in the target coordi-nate system.
TARGET POINTS AS VIEWED BY SENSOR. Consider an axis system fixed with respect to the camera. The specific axis system chosen is the right-handed system shown in Fig. 2, with x representing horizontal, positive to the right, y representing horizontal, positive along ~orward lens axis; and z representing vertical, positive up. The first nodal point of lens, i.e., the center of the lens of camera 40 for id~alized "thin" lens is considered the origin.
In this system, the target reference point PO will be at some vector location, R.
If the pallet is not rotated, then points P~
and PB would be at vector locations R + a and R ~ b, respectively. In general, however, there will be some rotation, i.e., a and b will be rotated. There are, therefore, these forms for vectors in the two axis systems:

~2773~Z

Vector Vector Target Locations Locations Point in in Label_ Pallet Axes Sensor Axes PO (origin) R
PA a R
PB b R -~

where and are related to the original vectors by a rotation matrix M:
a = Ma B = Mb In short, the image poin-ts will correspond to three points which, from the viewpoint of the sensor, are at locations R, R -~, R ~. The vectors R,~ and ~, and the rotation matrix M, are initially unknown.

DIRECTION VECTORS, u, v, w. In general, the direction of any single source or target point can be established with a camera system, but not distance.
The direction can be defined by a vector Erom image point to lens center (thin lens~ or second nodal point (thick lens)~ See Fig. 9.
For the target points, let the direction vectors corresponding to PO, PA~ and PB be called u, v, w, respectively. These vectors could be chosen as unit vectors (generally convenient for analysis~. A more convenient and natural normalization is to scale these vectors so that the lens-axis or y-component equals focal length. The x and z compo-nents are then simply the horizontal (~) and vertical (n) components of location in the focal plane. (With due regard for signs and "lmage reversal.") '`

~77;~9Z

-22~

B~SELINE VECTOR EQUATIO~S. In terms of known direction vectors u, v, w, the basic vector equations - become R = ~Ou (1) R ~ a = ~Av (2) R + ~ ~ ~BW (3) where ~0~ ~A~ ~B are (unknown) scalars propor-tional to distance.
Equations (1) through (3) are underdetermined.
There are 12 unknowns (three components each for R, a, ~ vectors, plus the three scalar ~'s), and nine scalar equations. The l'missing" three equations come from scalar invariance relationships.

SCALAR INVARIANCE RE~ATIONSHIPS: THREE SCALAR
EQUATIONS. Although a and ~ are unknown as vectors, partial information comes from scalar invariants of (rigid-body) rotation Specifically, if a= Ma; ~= Mb, M = rotation matrix, then : a ~ a = a 2 = a2 (known) (4) 2 = b2 (known) (5) aD~ = a-b (known) (6) These equations correspond to the physical properties that rotation does not change vector lengths, and for rigid-body rotation the angle sub-tended by any two vectors remains unchanged.

.' ' ; .

~;Z7739Z

ALGORIT~M OVE~VIEW: REQUIREMENrrS, PHILOSOPH~, PROBLEMS. Equations (1) through (6) collectively form a fully determined baseline e~uation set. They give 12 nonlinear (scalar) equations in 12 (scalar) un-knowns, if evaluation of unknown rotation matrix M istemporarily regarded as a "later step."
The algorithmic steps are based upon mathe-matical analysis that successively reduces the dimen-sionality of the problem. The ultimate step becomes that o~ solving a single nonlinear equa-tion in one unlcnown, after which the reduction steps are re-traced to give explicit values of the other variables. At the end oE the retrace, R, ~, and 3 are known.
A separate procedure is then developed to establish, from computed ~ and ~ and preset a and b vectors, the rota~ion matrix M. From M, the evaluation of "standard" rotation angles -- pitch, roll, and yaw -- is then straightEorward. Evaluation of explicit pitch, roll, and yaw is convenient for sensitivity studies and design analysis. For the operational system, these angles may not be explicitly required.
Although the obvious goal of algorithmic development is to achieve some method of solving the large set of simultaneous nonlinear equations, there is not necessarily uniqueness in the approach. To the extent possible, the exact approach chosen and docu-mented here was aimed at (a) minimizing sensitivity to effects of computer roundoff errors and (b~ providing For efficient, that is fast, computation.
There are two distinct solutions to a given problem. An algorithm requirement, therefore, is the capability to evaluate both solutions and to select the one that is "correct. ~! Problems of "correctness"
are addressed later.

~2'77392 The sequence of steps to reduce the baseline equations, leading to evaluation of the primary vectors R,~, and ~will now be described.

EQUA'rIOi~S (l) THROUGH (3~: CHANGE IW ~ PARA-MEirERSo In Equations (1), (2), and (3)~ this change of variables for the ~'s:

~0: (no change) ~A = ~o (1 +~) ~B = ~o (1 + ~) gives revised forms 15 R = ~ ou (1') R + ~ = ~o (1 +~) v (2') ~ + B = ~o (1 +~) w- (3') This form allows separation of variables as discussed below, and also eventually allows isolation o "small"
rom "large" quantities.

ELIMINATION OF R. Equation (1') can be substituted into Equations (2') and (3') to give two equations for ~ and ~ with ~0 as a parameter:
~/~0 = (1 + ~) v - u = (v - u) + ~ v (7) ~/~0 = (1 + ~) w - u = (w - u) + ~ w ~8) , ~27739~2 -~5-These equations are characterized by two forms of isola~ion of "large" from "small" variables.
First, the R vector, which is generally large relative to the ~ and 3 vectors, has been eliminated. Second, the parameter ~0, which is generally large relative to the , ~ parameters, has been partially isolated.
This isolation will become complete in the next two steps.

RE~UCTION TO THREE SCALAR EQUATIONS. The vector Equations (7) and (8) can be converted to three scalar equations with vector dot-product forms propor-tional to N2, ~2,~
0 = (V -- U) ~ 2V- (V -- U) + ' V (9) 0 = (W - U) + 2~iW~ (W - U) + ~i ~ W (10) = ( V -- U)' (W -- U) ~ V'(W -- U) + ~iW' (V -- U) +~i V'W (ll) Note that ~2, ~2, and ~-~ are known (Equa-tions 4, 5 and 6)~ Also, all of the vector dot prod-ucts on the right-hand sides o these equations are constants that depend only upon known u, v, w vectors.
These equations, therefore, have the form ~ /~o Al + A2 ~ A3 (9') ~ /~0 = sl + B2~ ~ B3 (10~) 1 2 3 4 (11 ) with ~2, ~2,~.~, Al, ..~ C4 as known constants, and with three scalar unknowns ~0, and ~.

ELIMINATION OF~o : REDUCTION TO TWO SCALAR
EQUATIONS. A consequence of the change from (~0, ~A~B) to (~0, ~ ~ iS the way in which ~0 is isolated in Equations (9) through (11). Elimination ~'27739;2 of ~0 is now a tri.vial algebraic step, giving two equations in scalars and ~ only. The obvious forms are ~A1 + A2E + A3E ~ ~ ~B1 2 3 (~'~)'~A1 + A2 E + A3E ] = ~ CC1 ~ C2 + C3~ + C4~' These are simply two 2-dimensional polynomials in E
and ~, expressible in form D2 + D~E + D4~ + D5~ = (12) E2 + E3E + E4 ~-~ E5~ = (13) where D and E coefficients are combinations of (pre-vious) known constanks.

REDUCTIOW TO EQUATION IN ONE V~RIABLE/E.
Equations (12) and (13) are near the end oE the reduc-tion process. They could be solved simultaneously, using (for example) a 2-dimensional version of Newton successive approximation. That approach is not pre-ferred, largely because of practical problems in determining two distinct solut.ions (a requirement previously reEerred toj, and because convergence would probably be slower than for alternatives discussed below.
Equa-tions (12) and (13) could also be combined to give a purely l-dimensional equation in E--a fourth-order polynomial equation of form :E'1 + P2 E + P E + P4 + P5E = (14) This approach is analytically attractive. In its :~ final form, the polynomial computations would be ~2773~2 simple and easily adapted to finding multiple solutions. Convergence would generally be fast relative to a 2-variable approach.
The prac-tical disadvantage of the Equation (14) approach is the complexity, lengthy software code and computation time involved in evaluating the P
coefficients prior to solving for roots.
The preEerred approach is a l-dimensional approach of a somewhat different form.
First, note that Equa-tion (13) can be solved for ~ as an expLicit function of ~:

~g (~ (El + E2 E + E3 ~ ) (15) E4 + E

Equatlon (12) can then be written as if a function only of F o f () = Dl + D2 + D3 + D4~ + D5~ = O (16) with the understanding that ~ is always evaluated (from Equation (15)) as a function of ~
An ~-solution, therefore, is a value of ~ for which Equation ~16) holds. The corresponding iS
~5 then given by Equation (15), once an ~ solution is identified.
This approach, discussed below, has the convergence rate of the "pure" l-dimensional approach (4th-order polynomial), but with a lesser amount of subsidiary computation.

FINAL EQUATIONS FOR ~ AND~ : GENERAL PROPER-TIES AND PROBLEMS. The final approach involves finding the root, or roots, for ) = Dl ~ D2 ~ + D3 + D~ ~ + D5 (S = (16~ repeated) " ~,7739~

with ~ as an explicit function of ~:

= -(EI + E2 + E3 ~ ) (15, repeated) E4 -~ E5 E
For the class of problems and target-point geometries that occur in applications of the type described here, there are normally two distinct real roots to Equation (16), and two complex roots~ Physi-cal solutions correspond to the two real roots.
OE the two physical solutions, one is"usually" identifiable as not valid. For some of the retroreflector target-point geometries (vs. mirror configurations), and in presence oE mosaic quantization and/or other sources of errors, resolution of correct versus incorrec-t solution is not necessarily reliable.
This problem is strictly a data error problem, not an algorithmic problem. Empirical studies indicate that this type of problem does not occur for the mirror configuration and realistic distance/angle combina-tions.
A potential problem exists wit~ Equation ~15), in that for some value of E (say, ~O) a zero ; denominator occurs. Theoretically (iOe., in absence of roundoff errors)~ this case must imply that the numerator is also zero and that a definite (finite) limiting value exists.
In software implementation, this zero/zero problem is explicity treated only for the following special case analytically established as physically possible:
.
= O
El = E~ = 0 ~Z 7739Z

For this case, Equation (15) is replaced with 2 + E3 ~ (15, special form) No ins-tances of zero~denominator problems have occurred in several thousands of software execu-tion of the algorithms. That fact does not, of course, guarantee that such a problern will never occur. Of course, additional protective code could be added.
Without such code, however: (a) the likelihood of a problem is small~ and (b) at worst, an abort of a pallet load or unload operation would occur.

THE "SECOND ROOT" PRO~LEM. Consider a situa tion in which (a) a real root, ~1~ has been evaluated for Equation (16) t and (b~ this root is judged to be the wrong root. For this case, the second root is established as the root of:
g() - o (17) g(~ = f () i:E ~ 1 (17a) = f (~l) if 1 (17b) Equation (17b) is the limiting form of (17a) as approaches 1-This formulation is equivalent to dividing out an (-1) factor from a pure polynomial form.
The pair of equations (15 and 16) yield double roots for and ~. Procedures for establishing both ~ and ~ pairs and selecting the physically valid values are required.
THE RETRACE STEP. A series o-f steps of variable elimination, leading eventually to the evalua tion of the and ~ parameters, has been descrlbed.

~;27~73~2 Retracing is the process of then working backwards, with known ~ and ~, to get the three required baseline vectors R, ~ and ~.

RETRACE To ~0O The ~0 parameter is a required intermediate variable. It is obtained from Equation (9'), rewritten in the explicit form ~0 = ~ Al -2 - (18) It could also be evaluated from Equation (10'), or (provided ~ O ~ from Equation (11'), since ~ and ~
have, in p~inciple, been evaluated to make these three equations compatible.

RETRACE FOR R, ~, AND~. With~ , ~, and ~0 established, then:

R = ~Ou (19) ~ = ~ Cv-u + ~ v ~ (20) = ~0 [w-u + ~ w ] (21) The steps for determining rotation matrix and the pitch, roll, yaw angles will now be described4 THE ROTATION PROBLEM: THE M l~ATRIX. Once ~
and ~ are evaluated, a new algorithmic problem arises in solving for the unknown rotation matrix M that satisfies a Ma (22) ~ = Mb. (23) ~27739~

M is a 3 x 3 matrix, hence with nine elements. These nine elements are not independent, however. The constraint that M be a rotation matrix implies that only three degrees of freedom exist. These three degrees of freedom can be identified with pitch, roll, and yaw angles, but that identification is neither required nor useful in the s-teps to solve for M.

MATRIX FORMULATI0~ N0. 1 T0 GIVE M. The two Equations (22) and (23) can be supplemented with a third linearly independent equation, that must be true for a rigid body rotation:
= M (a ~1 b) (24) where ~ means vector cross product.
Now write the separate Equations (22) through (24) as a single matrix equation ~ (25) ( ~ M~a b a ~ b ) (25') Q = M P

where, for example, P has vector a as its first column, vector b as second column, etcO
The matrix P is nonsingular (provided only that the pallet vectors a and b are not collinear), so an immedia~e formal solution results:

M = Qp 1 In this forml solution would require computa-tion of the inverse of 3 x 3 P matrix, Eollowed by a matrix~times-matrix product.

:, MATRIX FOl~MUI.ArION NO. 2 ~ro GIVE ~. The actually implemented algorithm for calculation of the M matrix avoids the computational effort of matrix inversion. The concept is to use not the ~ and 3 S vectors (and ~ ~ 3 ) directly, but to use equally valid combinations that are orthonormal~ For a matrix with orthonormal columns, inversion is not explicitly required - the inverse is simply the transpose.
,~ ~ ~
Three base vectors ~, ~, and r will be chosen by combining and norlnalizing the ~ and B vectors.
Exactly the same (hence compatible) combinations and normalization will be applied for the a, b vectors, giving â, b, c. Vectors ~, ~, and y will form an orthonormal set (orthogonal unit vectors); â, b, c will also form an orthonormal set; and the same M-matrix rotation property will hold. A form analogous to Equation (25) is then ~ ) t27) ( ~ ~ y ) = M ( a b c (27') Q = MP
but with the distinction that the inverse of P is simply the transpose of P(_ P'). Therefore .~ ~
M = QPI (28) Algorithmic steps to get the orthonormal vectors are discussed below.
ESTABLISHING â,âO The first vectors, ~ and a, are trivial:

= unit (~) a = unit (a) (29) ~Z773~Z
-33~

where un.it "( )" means normalized to unit length.

ESTABLISHING ~b- For ~, we want a vector orthogonal to ~ (hence, orthogonal to ~), that retains the information in ~. lrhe simplest construct, and the one implemented, is to first form the linear combination ~ c ~ (30) with the constant c chosen as C = ~J ~ ~ (31) u Then ~' will be orthogonal to ~ and ~ The second step is then simply to nor~alize ~' to a unit vector:
~ = unit (~') (32) For b, the same steps are followed using a and : 20 b:
b' = b - ca ~33) b = unit (b') (34) Note that (a~b)/a2 is the same as (~ 2 from scalar invariance properties, so the "c' in Equation (30) is the same as in Equation (33).

ESTABLISHING y,c~ For ~r a unit vector orthogonal to ~ and ~ is given by the vector cross 30 product y ~ (35) Similarly, c = a ~ b (36) ~2'7739Z

Since these are (au-tomatically) unit vectors, explicit numerical normalization steps are not required.
Equation (28) is the form used for software evaluation of M. It requires a matrix-times-matrix ; 5 multiplication, no explicit matrix inversion.

PITCH~ ROLL, AND YA~ ANGLES FROM M. The matrix M can be considered to be the product of three canonic matrices associated with pure pitch, roll, and yaw. The convention chosen for the order of multipli-cation is M ~ (vector) = P R Y ~ (vector)l i.e., the yaw matrix Y is first applied -to the vector, then the roll matrix R, then the pitch matrix P.
The canonic matrices are /cy -sy o \
20 Y =~sy cy O ~ (38) /Cr -St R =¦0 1 0 (39 :~ St c~
/1 0 0 ~
P =1 cp -sp (40) ~O sp cp where cy = cos ~yaw anglé), sy a sin (yaw angle), Cr = cos (roll angle), etc., are used as abbrevia-tions Eor trigonometric functions of angles. The product is then ~ r Y crsy ~sr \
M = P R Y = ¦ cps -spsrc cpcy+s s s -s cr ~ (41) ~ p y cp r y p y p r y Cpcr J

.
~' ' Given the values of the elements (Mij) of the M matrix, pitch, roll, and yaw angles are then evalu-atable as inverse trigonometric functions:
pitch angle 2 tan~l . ~-M23~ (42) roll angle = slnl (-M13) (43) yaw angle = tan~l (-M12 ) (44) Mll While the method herein described and the form of apparatus for carrying this method into effect constitute preferred embodiments of this invention, ik is to be understood that the invention is not limited to this precise method and form of apparatus, and that changes may be made in either without departing from the scope of the invention, which is defined in the appended claims.

:` 25 :

;~

~ 35 ,.,

Claims (24)

1. A target member for use in a positioning system comprising:
a support member; and at least three reflector elements mounted on said support member, each of said reflector elements being so configured as to form an image of an identifying means located within a predetermined field of view with respect to said support member, said images defining a plane oriented other than normal to a line from the identifying means to said plane, said images also defining a circle that does not include the identifying means.
2. The target member of claim 1 wherein said reflector elements include at least two different radii of curvature, the radius of curvature of each element being selected to provide a reflection of an identifying means located within a predetermined field of view with respect to said reflector elements.
3. The target member of claim 1 wherein said reflector elements comprise at least three non-planar reflector elements attached to said support member in a predetermined alignment, two of said reflector elements having essentially the same radius of curvature, and the third reflector element having a different radius of curvature.
4. The target member of claim 1 including at least one retroreflector member and at least one non-planar reflector element.
5. The target member of claim 1 wherein each of said reflector elements have essentially the same radius of curvature and wherein at least one of said reflector elements is longitutionally spaced from the other said reflector elements.
6. Apparatus for use in a docking system in which a moving vehicle is to he aligned with and closed to a dock position, a target member to be mounted at the dock position, said target member including a generally flat support member and at least three non-planar reflecting surfaces supported on said member in predetermined alignment, at least two of said reflecting surfaces both being either concave or convex and the third surface being opposite to the two, whereby reflections from a common radiation source are perceived at known spacing and same sense in the two surfaces and at known spacing and inverted sense with respect to the others in the third surface.
7. A target member for use in a positioning system including:
a generally flat support member;
at least three reflector elements mounted on said support member in a predetermined alignment, said reflector elements together forming images of a remotely positioned identifying means in an image plane oriented other than normal to the plane of the support member;
retroreflector members mounted on said support member and surrounding said reflector elements; and coded means on said support member for uniquely identifying the target member.
8. The target member of claim 7 wherein at least two of said reflector elements are either convex or concave and have the same radius of curvature.
9. The target member of claim 7 wherein said reflector elements are mounted with the center of each element on a horizontal line, and wherein two of said reflector elements are convex and the other concave.
10. The target member of claim 7 wherein said coded means is a bar code printed on said support member above said reflector elements.
11. Apparatus for use in determining the location of a vehicle within a predetermined field of view with respect to an object, said apparatus comprising:
identifying means carried by a vehicle;
a target member mounted on said object, said target member including a support member and at least three reflector elements mounted on said support member in a predetermined alignment, said reflector elements together forming images of said identifying means in an image plane oriented other than normal to a line from said identifying means to said plane when said identifying means is located within a predetermined field of view with respect to said support member; and sensing means carried by said vehicle for sensing the directions of the reflections of said identifying means from each of said reflecting elements.
12. The apparatus of claim 11 wherein said reflector elements are spherical in shape.
13. The apparatus of claim 11 wherein at least two of said reflector elements are non-planar.
14. The apparatus of claim 11 wherein at least two of said reflector elements have the same radius of curvature.
15. The apparatus of claim 11 wherein said identifying means includes a high intensity light source.
16. The apparatus of claim 11 wherein said sensing means includes an imaging camera.
17. The apparatus of claim 11 wherein said target member is stationary.
18. The apparatus of claim 11 wherein said target member is attached to a pallet and wherein said vehicle is a self-propelled forklift truck.
19. A method for determining the relative position of a target member from a vehicle comprising the steps of:
forming on the target member at least three images of an identifiying means associated with the vehicle, said images defining a plane oriented other than normal to a line between the identifying means and said plane, said images also defining a circle that does not include the identifying means; and determining the directions from the vehicle of each of the images of the identifying means.
20. Apparatus for use in determining the relative location of a vehicle with respect to a target member, including:

a light source and a camera mounted on the vehicle, said target member including at least three reflector elements, each selected to form an image of the light source, and all the images together forming an image plane oriented other than normal to a line from the light source to the target member;
means for recording those images on the image plane of said camera due to ambient light;
means for thereafter flashing the light source;
means for comparing said images due to ambient light stored in said recording means and the images resulting both from ambient light and from the light source; and means for recording those images on the image plane of said camera due only to reflections of said light source.
21. The apparatus of claim 20 wherein said target member further includes:
retroreflective members mounted on said support member and in a predetermined relation to said reflector elements;
means for initially flashing the light source;
means for sensing the reflections of the light source from the retroreflector members as they appear on the image plane of the camera; and means responsive to the images of the retroreflectors for defining an area on the image plane of the camera where reflections from the reflector elements would be located.
22. A method for gathering information for determining the relative location of a target member with respect to a vehicle wherein the vehicle is provided with a light source and a camera, and wherein the target member is provided with at least three reflector elements, each selected to form an image of the light source, and all the images together forming an image plane oriented other than normal to a line from the light source to the target member, the method comprising the steps of:
a) recording the images formed on the image plane of the camera due to ambient light;
b) flashing the light source;
c) comparing the images recorded in step a) with the images resulting both from ambient light and from the light source thereby to isolate those reflections of the light source in the reflector elements;
d) recording those images of the reflections due to the light source; and e) identifying the position of each such reflection of the light source from the reflector elements on the image plane of the camera and with respect to each other.
23. A method for gathering information for determining the relative location of a target member with respect to a vehicle, and wherein the vehicle is provided with a light source and a camera wherein the target member is provided with at least three reflector elements, each selected to form an image of the light source and all the images together forming an image plane oriented other than normal to a line from the light source to the target member, and retroreflective members in a predetermined relation to the reflector elements, the method comprising the steps of:
a) flashing the light source;
b) recording the positions of the reflections of the light source from the retroreflector members as viewed by the camera;

c) defining an area within the image plane of the camera where reflections from the reflector elements would be located by reference to the location of the light reflected by the retroreflector members;
d) recording the images from the target member within the defined area due to ambient light;
e) flashing the light source a second time;
f) recording the images within the defined area resulting from both ambient light and from the light source;
g) comparing the images from step d) with those from step f) to isolate those reflections from the light source; and h) determining the position of each reflection of the light source from the reflector elements within the camera and with respect to each other.
24. The method of claim 23 wherein the images recorded in step f) must be brighter than a predetermined level, and wherein the record of all reflections from the target member from step d) are compared with those reflections from the target member brighter than the predetermined level from step f) in order to identify those reflections of the light source reflected only by the reflector elements.
CA000520829A 1985-10-18 1986-10-17 Target and control system for positioning an automatically guided vehicle Expired - Fee Related CA1277392C (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US789,280 1985-10-18
US06/788,989 US4684247A (en) 1985-10-18 1985-10-18 Target member for use in a positioning system
US788,989 1985-10-18
US06/789,280 US4678329A (en) 1985-10-18 1985-10-18 Automatically guided vehicle control system

Publications (1)

Publication Number Publication Date
CA1277392C true CA1277392C (en) 1990-12-04

Family

ID=27120859

Family Applications (1)

Application Number Title Priority Date Filing Date
CA000520829A Expired - Fee Related CA1277392C (en) 1985-10-18 1986-10-17 Target and control system for positioning an automatically guided vehicle

Country Status (4)

Country Link
EP (1) EP0243493A1 (en)
KR (1) KR880700512A (en)
CA (1) CA1277392C (en)
WO (1) WO1987002484A1 (en)

Families Citing this family (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4869635A (en) * 1988-03-31 1989-09-26 Caterpillar Industrial Inc. Apparatus for controllably positioning a lift mast assembly of a work vehicle
GB8822795D0 (en) * 1988-09-28 1988-11-02 Gen Electric Co Plc Automated vehicle control
FI85260C (en) * 1990-07-11 1992-03-25 Vesa Kaehoenen FOERFARANDE OCH SYSTEM FOER ATT INSTAELLA HYDRAULTRYCKET SOM VERKAR PAO ETT HYDRAULISKT GRIPORGAN.
GB2259823A (en) * 1991-09-17 1993-03-24 Radamec Epo Limited Navigation system
IT1275664B1 (en) * 1994-11-16 1997-10-17 Consorzio Telerobot SYSTEM FOR THE CONTROL AND AUTOMATIC DRIVING OF A FORK LIFTING GROUP
DE19541379C2 (en) * 1995-11-07 2001-01-18 Fraunhofer Ges Forschung Method for determining the position of a vehicle in a driving plane
DE19738163A1 (en) * 1997-09-01 1999-03-11 Siemens Ag Method for docking an autonomous mobile unit using a beacon
FR2799002B1 (en) * 1999-09-28 2006-07-21 Thomson Csf ACTIVE LASER IMAGING METHOD
JP2001206696A (en) * 2000-01-21 2001-07-31 Nippon Yusoki Co Ltd Forklift
US7473884B2 (en) 2005-04-21 2009-01-06 Avago Technologies Ecbu Ip (Singapore) Pte. Ltd. Orientation determination utilizing a cordless device
US7609249B2 (en) 2005-04-21 2009-10-27 Avago Technologies Ecbu Ip (Singapore) Pte. Ltd. Position determination utilizing a cordless device
FI117835B (en) 2005-06-22 2007-03-15 Sime Oy Positioning method
US7796119B2 (en) 2006-04-03 2010-09-14 Avago Technologies General Ip (Singapore) Pte. Ltd. Position determination with reference
BE1018160A3 (en) * 2008-05-26 2010-06-01 Egemin Nv Automatic guided vehicle, has load handling device connected to onboard computer, where vehicle is provided with detection unit, which comprises single sensor for determining relative position of load handling device
SE534240C2 (en) * 2008-12-05 2011-06-14 Datachassi Dc Ab Procedure and systems for providing docking assistance
US8995743B2 (en) 2011-10-19 2015-03-31 Crown Equipment Corporation Identifying and locating possible lines corresponding to pallet structure in an image
GB2513912B (en) * 2013-05-10 2018-01-24 Dyson Technology Ltd Apparatus for guiding an autonomous vehicle towards a docking station
US9378554B2 (en) 2014-10-09 2016-06-28 Caterpillar Inc. Real-time range map generation
US9449397B2 (en) 2014-10-15 2016-09-20 Caterpillar Inc. Real-time visual odometry system for determining motion of a machine with a range detection unit
US9990535B2 (en) 2016-04-27 2018-06-05 Crown Equipment Corporation Pallet detection using units of physical length
EP3269678B1 (en) 2016-07-14 2019-03-06 Toyota Material Handling Manufacturing Sweden AB Floor conveyor
EP3269679B1 (en) 2016-07-14 2019-09-11 Toyota Material Handling Manufacturing Sweden AB Floor conveyor
EP3269680B1 (en) 2016-07-14 2020-09-30 Toyota Material Handling Manufacturing Sweden AB Floor conveyor
US11459221B2 (en) * 2020-04-24 2022-10-04 Autoguide, LLC Robot for stacking elements

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
HU170235B (en) * 1974-05-29 1977-04-28
SE7804927L (en) * 1978-04-28 1979-10-29 Volvo Ab DEVICE FOR ORIENTATING, FOR EXAMPLE, A LIFTING RELATION IN RELATION TO A LOAD
US4331417A (en) * 1980-03-07 1982-05-25 Rapitsan Division, Lear Siegler, Inc. Vehicle alignment and method
FR2495797A1 (en) * 1980-12-09 1982-06-11 Onera (Off Nat Aerospatiale) Laser-optic automatic positioning for electric vehicle - uses fixed reference of three retro-reflectors and measures angle of reflected light to position vehicle
DE3305277A1 (en) * 1983-02-16 1984-08-16 Ing. Günter Knapp GmbH & Co. KG, Graz Method and device for the automatic storage and removal from storage of piece goods

Also Published As

Publication number Publication date
KR880700512A (en) 1988-03-15
EP0243493A1 (en) 1987-11-04
WO1987002484A1 (en) 1987-04-23

Similar Documents

Publication Publication Date Title
CA1277392C (en) Target and control system for positioning an automatically guided vehicle
US4678329A (en) Automatically guided vehicle control system
US4684247A (en) Target member for use in a positioning system
EP3465266B1 (en) Optical system for object detection and location
US11448735B2 (en) LiDAR with malfunction detection
US10408940B2 (en) Remote lidar with coherent fiber optic image bundle
US5745235A (en) Measuring system for testing the position of a vehicle and sensing device therefore
US4753569A (en) Robot calibration
US20050232466A1 (en) Method of recognizing and/or tracking objects
WO2002075350A1 (en) Method and device for determining an angular position of a reflector
US11803049B2 (en) Machine vision system and method with steerable mirror
JP2005504963A5 (en)
US4916302A (en) Apparatus for and method of measuring distances to objects present in a plurality of directions
US20220100979A1 (en) Machine vision system and method with on-axis aimer and distance measurement assembly
JP3146809B2 (en) Vehicle environment recognition device
US20230206478A1 (en) Imaging arrangement and corresponding methods and systems for depth map generation
KR20220076523A (en) Systems and Methods for Infrared Sensing
JPH09145320A (en) Three-dimensional input camera
KR20180081007A (en) Wide angle LiDAR and operation method thereof
KR100262182B1 (en) Vision sensor for amr and method for estimating of location using it
US11276192B2 (en) Interleaved image capture for dimensioning system
KR102506812B1 (en) Autonomous vehicle
GB2228157A (en) Optical sensor
US20230311770A1 (en) Vehicular camera focus test system using light collimator in controlled test chamber
US20220381886A1 (en) Detection system and method using a line detector

Legal Events

Date Code Title Description
MKLA Lapsed