GB2225852A - Object recognition system - Google Patents
Object recognition system Download PDFInfo
- Publication number
- GB2225852A GB2225852A GB8926000A GB8926000A GB2225852A GB 2225852 A GB2225852 A GB 2225852A GB 8926000 A GB8926000 A GB 8926000A GB 8926000 A GB8926000 A GB 8926000A GB 2225852 A GB2225852 A GB 2225852A
- Authority
- GB
- United Kingdom
- Prior art keywords
- lines
- image
- parallel
- vanishing
- parallel lines
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/60—Analysis of geometric attributes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Geometry (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
- Control Of Position, Course, Altitude, Or Attitude Of Moving Bodies (AREA)
Abstract
A system for the automatic recognition of parallel lines in a perspective image of a viewed object (8) comprises a recognition system (7) which detects edge lines of the image and recognises as parallel lines those lines which have a common vanishing direction (17). An algorithm generates a set of planes (14, 15, 16) for the image lines, together with a set of initial pairings between these lines. The quality of fit is decided by the deviation of the end points of the lines in a least squares fit to the common vanishing direction.
Description
Object Recognition
This invention relates to a system for detecting parallel lines in a perspective image of a viewed scene. It is intended for use primarily, but not exclusively, in a system for recognising an object, such as a pallet, which is to be picked up by an autonomous guided vehicle.
In a system as described in British Patent: 2,143,395, a vehicle can accurately determine its position and heading in a work space, such as the interior of a warehouse, by combining odometry with triangulated range measurements using reflecting bar-coded targets placed at known, fixed positions around the work space.
Since the vehicle can compute its position, it can move along software-computed paths between known end points, such as conveyors or machine tools. However, in such a system a vehicle cannot locate objects, such as pallets, which may be at random positions in an area, such as a loading bay.
Work has been done on systems which, by use of computer vision, attempt to recognise the shapes of objects.
In general, parallel lines in a viewed scene do not project parallel lines in an image; more often, they converge on a vanishing point. Determination of the vanishing point is the usual method used by object recognition systems for characterising the projected parable0 lines. However, this is a very unstable representation: the error-statistics are highly anisotropic, with a serious singularity when the real 3-dimensional parallel lines are nearly parallel to the image plane. In this case, the lines appear nearly parallel in the image and the vanishing point lies a very great (in the extreme, infinite) distance away from the centre of the image plane.
It is an object of the present invention to provide an improved system for the automatic recognition of parallel lines in a viewed scene.
According to the invention there is provided a system for the automatic recognition of parallel lines in a viewed scene, the system comprising means to produce a perspective image of the object; and means to detect lines in the perspective image and to recognise as parallel lines those lines which have a common vanishing direction.
An embodiment of the invention will now be described, by way of example, with reference to the accompanying drawings, in which
Figure 1 is a schematic block diagram of a vehicle navigation system incorporating the invention.
Figure 2 illustrates the concept of vanishing direction,
Figures 3, 4, 5 and 6 show, respectively, a pallet image, a corresponding edge map, a set of recognised parallel lines, and an orthogonal set of recognised parallel lines, and
Figures 7, 8, 9 and 10 show, respectively, an image of two pallets with boxes thereon, a corresponding edge map, a set of recognised vertical lines, and a set of recognised horizontal lines.
Referring to Figure 1 of the drawings, a vision system 1, comprising a video camera 2, is mounted on a vehicle 3, which is controlled for general movement around a warehouse or other site by signals from a base station 4, fed to the vehicle navigation system 5 over a radio link 6.
The vision system 1 is coupled to a data processing system 7 for processing data obtained from the vision system. The operation of the system will be described in relation to the viewing of a scene including one or more pallets 8, but it is applicable to the viewing of scenes including other objects having edges which comprise parallel lines.
In accordance with the present invention, the recognition of parallel lines is carried out on the basis of a concept of "vanishing direction", which alleviates the above-mentioned problems associated with vanishing point determination.
Referring to Figure 2, the vanishing direction might be defined simply as the direction V from the focal point 10 of the camera system 1 to the vanishing point 11 of a scene comprising, for example, two parallel, but apparently converging, lines 12 and 13.
However, this definition is not sufficient, as it degenerates just as the definition of the vanishing point does. Figure 2 serves to provide another, more useful definition. The lines 12 and 13 in the image, along with the focal point 10, define a set of planes 14, 15 and 16 in space. If these lines all go through the vanishing point 16, then the common intersection of the corresponding planes is a line 17 in space joining the focal point to the vanishing point.
This construction can be carried out even if the lines appear parallel in the image; the vanishing direction is then the direction of a line in space which is parallel to the image plane and to the observed lines in the image plane.
The vanishing direction has a further advantage. If we have an accurate knowledge of the position of the focal point 10 (from an appropriate calibration procedure) then the vanishing direction coincides with the direction of the parallel 3-dimensional lines the images of which are used in the construction. The reasoning is as follows: the real lines are parallel, and lie in the same plane as their projection in the image plane. If these lines never meet, then to lie in these planes they must be parallel to the intersection of the planes and therefore parallel to the vanishing direction.
A difficulty which arises is to find the 'best' vanishing direction, given a set of lines which are approximately converging.
In order to do this, a model of the way in which the positional errors in the image can occur is needed. This model will then allow quantification of the fit between the image lines and the estimated vanishing point. The model used is relatively simple: the centre-points of the lines are assumed to be correct, while the
endpoints are allowed to deviate. The measure of fit is chosen to be the square of the endpoint deviation (in order to facilitate a
least-squares fit). While this measure is fairly accurate for longer
line segments, it allows excessive rotation of the shorter ones, the
orientation data of which is usually more reliable than the measure would suggest.
The processing system 7 (Figure 1) operates to find the
vanishing direction in accordance with an algorithm as defined
below,and which includes the following steps:
1. the equation of the plane defined by the image line and
the focal point is parameterised in terms of the endpoint deviation
of the line.
2. given this equation, a measure of fit between a set of
image lines and an estimated vanishing direction is calculated.
3. an iterative least-squares algorithm is then developed to find the optimal vanishing direction.
The following conventions are observed, namely that the
focal centre is chosen to lie at the origin of the co-ordinate system
of the cameras, and the image plane is parallel to the x-y plane.
The vanishing direction is represented by a vector p; planes (which in this case all pass through the origin) are
represented by their normal vectors; the vectors and are
respectively the vectors from the focal centre to the midpoint of an
image line, the direction vector of the line, and the normal to that
line (chosen such that (t, , k) is right-handed).
The line given by#v,#t generates a plane with normal v At thus given an endpoint deviation#the plane normal becomes: v A(t + 27 n) where 1 is the length of the line. A constraint is #
that the vanishing direction p should lie in the corresponding plane:
This is required for each #, #, ni, tj.
From the above, the endpoint deviation required for a given plane to contain the vanishing direction can be determined:
Given this, we then need a method of choosing p such that the combined measure of fit is minimised: Fe fj2, is chosen such that the fit is a least-squares fit. The next section describes a method of iteratively determining the optimal direction p.
A least-squares fit for p is not trivial, essentially due to the non-linearity of the problem, and to the fact that the variation occurs only orthogonally to#. The present approach to the problem is as follows:
consider the function in cartesian co-ordinates of the plane orthogonal to the initial estimate of p; approximate this function by a second-order polynomial in two variables, the stationary point of which can then be found analytically.
Given:
we find:
All of the \tj values are independent of the magnitude of p so all of the Vryj values lie in the plane orthogonal to p. Hence, we can consider the two components of VF in this plane as two functions, the zero-crossing of which is to be found. Two unit vectors a and b in the plane orthogonal to p are chosen such that:
Then gradients of both components are taken:
Linearising the gradient, we know that x = p - s (s is the solution) must be orthogonal to F7(b. VF), as the component of VF in thel-direction is already zero, as we wish it to be. It is also required that 7. Y( . 7 F) = Iv Fl .The required direction is given by u = V( 7F)Ar; so the final result is:
From this, the new estimate p of the vanishing direction is#' =p - x.
The above algorithm applies to finding the vanishing direction based on a given set of lines. The lines in an image must then be grouped into these parallel sets. This is effected as follows. The line segments extracted from the image are sorted in descending order of length and the shortest are discarded on the basis of a fixed threshold. Then, lines which are within 100 of each other are paired up. These pairs are used as starting points for finding parallel sets.The following steps are then carried out:
1. an initial vanishing direction is determined from the pair;
2. the values of7/are calculated for every line, based on the estimated value of p; 3. lines for which 7 is below a given threshold are included in the set, while the others are excluded from the set;
4. if the set has not changed, then the cycle has been completed;
5. otherwise perform a single iteration of the least-squares algorithm described above, and return to step 2.
Figure 3 shows an image of a pallet, produced by the vision system 1, and Figure 4 shows edge lines 20 which are found by the processing system 7, by use of any known edge detection method (Canny, etc). From these detected edges, the system 7 operates in accordance with the invention to detect sets of parallel lines 21,22 in one horizontal direction (Figure 5) and in the orthogonal horizontal direction (Figure 6), using the vanishing direction concept.
Figure 7 shows an image of a complicated scene, in which confusion is caused by the presence of two pallets, each of which is partly obscured by boxes which are resting on the pallets. Figure 8 shows the corresponding highly-complicated edge map derived by the processing system, and Figures 9 and 10 show vertical and horizontal sets of parallel lines 23,24 recognised by the present system.
It is envisaged that two operations might then be performed on the parallel lines detected by the system in accordance with the invention. Firstly, the sets of parallel lines might be investigated to check whether there are cyclically spaced groups of lines amongst them, indicating, for example, the presence of slats of a pallet or ribs on a container. Secondly, the sets of parallel lines might be compared with stored models, to determine whether there is sufficient correspondence to be able to decide that the object being viewed is a pallet or is another particular object for which data are stored.
If the system were able to determine that the object being viewed is indeed the object which is to be picked up, the navigation system would then drive the vehicle to the correct pick up position in front of the object.
Claims (4)
1. A system for the automatic recognition of parallel lines in a viewed scene, the system comprising means to produce a perspective image of the object; and means to detect lines in the perspective image and to recognise as parallel lines those lines which have a common vanishing direction.
2. A system as claimed in any preceding claim, wherein the means to produce a perspective image comprises a video camera system.
3. A system for the automatic recognition of an object, substantially as hereinbefore described with reference to the accompanying drawings.
4. A guidance and control system for a guided vehicle, including a recognition system as claimed in any preceding claim.
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
GB888826880A GB8826880D0 (en) | 1988-11-17 | 1988-11-17 | Vehicle control |
Publications (2)
Publication Number | Publication Date |
---|---|
GB8926000D0 GB8926000D0 (en) | 1990-01-10 |
GB2225852A true GB2225852A (en) | 1990-06-13 |
Family
ID=10646996
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
GB888826880A Pending GB8826880D0 (en) | 1988-11-17 | 1988-11-17 | Vehicle control |
GB8926000A Withdrawn GB2225852A (en) | 1988-11-17 | 1989-11-17 | Object recognition system |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
GB888826880A Pending GB8826880D0 (en) | 1988-11-17 | 1988-11-17 | Vehicle control |
Country Status (2)
Country | Link |
---|---|
GB (2) | GB8826880D0 (en) |
WO (1) | WO1990005957A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2305050A (en) * | 1995-09-08 | 1997-03-26 | Orad Hi Tec Systems Ltd | Determining the position of a television camera for use in a virtual studio employing chroma keying |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5245422A (en) * | 1991-06-28 | 1993-09-14 | Zexel Corporation | System and method for automatically steering a vehicle within a lane in a road |
JP5979904B2 (en) | 2012-02-20 | 2016-08-31 | キヤノン株式会社 | Image processing apparatus, ophthalmic imaging system, and image processing method |
-
1988
- 1988-11-17 GB GB888826880A patent/GB8826880D0/en active Pending
-
1989
- 1989-11-17 WO PCT/GB1989/001376 patent/WO1990005957A1/en not_active Application Discontinuation
- 1989-11-17 GB GB8926000A patent/GB2225852A/en not_active Withdrawn
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB2305050A (en) * | 1995-09-08 | 1997-03-26 | Orad Hi Tec Systems Ltd | Determining the position of a television camera for use in a virtual studio employing chroma keying |
Also Published As
Publication number | Publication date |
---|---|
WO1990005957A1 (en) | 1990-05-31 |
GB8926000D0 (en) | 1990-01-10 |
GB8826880D0 (en) | 1988-12-21 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Nitzan | Three-dimensional vision structure for robot applications | |
EP2460629B1 (en) | Control method for localization and navigation of mobile robot and mobile robot using same | |
Ohya et al. | Vision-based navigation by a mobile robot with obstacle avoidance using single-camera vision and ultrasonic sensing | |
US6952488B2 (en) | System and method for object localization | |
EP0156342A2 (en) | Three-dimensional vision system | |
US20050008220A1 (en) | Method, apparatus, and program for processing stereo image | |
EP1477934A2 (en) | Image processing apparatus | |
JP7252581B2 (en) | Article detection device, article detection method, and industrial vehicle | |
KR900002509B1 (en) | Apparatus for recognizing three demensional object | |
CN114170521B (en) | Forklift pallet butt joint identification positioning method | |
Xie et al. | Active and intelligent sensing of road obstacles: Application to the European Eureka-PROMETHEUS project | |
GB2225852A (en) | Object recognition system | |
Billah et al. | Calibration of multi-lidar systems: Application to bucket wheel reclaimers | |
Kita et al. | Localization of pallets on shelves in a warehouse using a wide-angle camera | |
JPH03161223A (en) | Fitting of work | |
JPH07103715A (en) | Method and apparatus for recognizing three-dimensional position and attitude based on visual sense | |
Kita et al. | Detection and localization of pallets on shelves using a wide-angle camera | |
CN111028264B (en) | Rotation robust three-dimensional object detection optimization method and device | |
EP0307120A2 (en) | Object recognition | |
Pedersini et al. | Multi-camera parameter tracking | |
Kita et al. | Image Measurement Method for Automatic Insertion of Forks into Inclined Pallet | |
JPH11300683A (en) | Bag-like workpiece attitude detecting device | |
JPS6294230A (en) | Hole searching system | |
JPH05164569A (en) | Run road detector of traveling vehicle | |
JP4013732B2 (en) | Vehicle external recognition device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WAP | Application withdrawn, taken to be withdrawn or refused ** after publication under section 16(1) |