US5245675A - Method for the recognition of objects in images and application thereof to the tracking of objects in sequences of images - Google Patents
Method for the recognition of objects in images and application thereof to the tracking of objects in sequences of images Download PDFInfo
- Publication number
- US5245675A US5245675A US07/772,328 US77232891A US5245675A US 5245675 A US5245675 A US 5245675A US 77232891 A US77232891 A US 77232891A US 5245675 A US5245675 A US 5245675A
- Authority
- US
- United States
- Prior art keywords
- image
- zone
- zones
- signature
- differences
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/42—Global feature extraction by analysis of the whole pattern, e.g. using frequency domain transformations or autocorrelation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/24—Character recognition characterised by the processing or recognition method
- G06V30/248—Character recognition characterised by the processing or recognition method involving plural approaches, e.g. verification by template match; Resolving confusion among similar patterns, e.g. "O" versus "Q"
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/62—Extraction of image or video features relating to a temporal dimension, e.g. time-based feature extraction; Pattern tracking
Definitions
- the invention relates to the field of image processing and, more particularly, an object of the invention is a method for the recognition of shapes in images and its application to the tracking of objects in sequences of images.
- the recognition of objects in an image generally relies on a search, in the image, for the patterns that most resemble known models.
- the model of recognition of an object may take different forms but always characterizes a precise representation of the characteristics of the object to be recognized. It may be a matrix of dots, a collection of geometric elements, a set of contours or, again, an analysis in a particular field (Fourier transform, Hadamard transform wavelets etc.).
- An object of the invention is a method, and the corresponding device, for the recognition of shapes, wherein the reliability and robustness of the results are obtained without any excessive cost in terms of computation time.
- the method uses, for this recognition, a decomposition into singular values of the image (or S.V.D. for singular value decomposition).
- the singular value decomposition of an object represented by a matrix of dots characterizes this object by its "signature" on a projection space defined by two projectors.
- This signature constituted by the singular values of the object matrix, defines the recognition model.
- An object is then defined by three matrices referenced U, V and S.
- the matrix S is an ordered diagonal matrix. All the values of the diagonal correspond to the signature of the object.
- the matrices U and V are the matrices of the projection.
- An object is recognized if its signature, after projection, is close to the signature sought. For, there is an almost univocal relationship between the signature and the zone to which it corresponds.
- the recognition is done according to an iterative method on the eigen values.
- a first sorting operation is carried out on all the possible zones, in using only one first eigen value.
- a sorting will be done from among the zones selected at the previous step on a greater number of eigen values.
- the method therefore consists, initially, in carrying out a rough sorting operation on a large number of zones, and then in carrying out an increasingly fine sorting operation on an increasingly limited number of zones.
- the computations done in each step are re-used in the following step.
- a first object of the invention is a method for the recognition of objects in images comprising:
- said method uses a decomposition of the matrices of luminance values into singular values according to a matrix product of projection matrices and an ordered diagonal matrix, said ordered diagonal matrix constituting a characteristic signature of the zone considered, and wherein, for the recognition phase, differences between the signature of the model zone and the signatures of the zones of the image considered are computed and image zones are selected on the basis of the least differences according to an iterative method such that:
- the p 2 first coefficients of the signature matrices, p 2 being greater than 1, are considered for a reduced set of n 1 image zones selected in the previous step as image zones exhibiting the least differences;
- a number p i , greater than p i-1 , of matrix coefficients is considered for a reduced set of n 1 image zones selected in the step (i-1) because they exhibit the smallest differences, with i varying from 3 to k-1.
- a second object of the invention lies in the application of the method to the tracking of objects in sequences of images.
- FIGS. 1 and 2 respectively illustrate an image in which there is selected a zone comprising an object to be recognized and a second image in which this object should be recovered;
- FIG. 3 is an explanatory diagram
- FIG. 4 illustrates the recognition method according to the invention.
- Singular values decomposition is a promising method in the field of image processing for applications of shape compression, filtering and recognition.
- transformation by singular value decomposition consists in analyzing an image (or part of an image), in which the luminance values of the points form a matrix A with a size (m, n) that is rectangular or square, into a matrix product defined by:
- diagonal elements s 1 . . .s p are singular values of the matrix A and are positive real numbers (positive square roots of the eigen values) classified in descending order:
- the columns u i of [U] are the eigen vectors of [A] t ⁇ [A].
- [U] [u 1 ,u 2 . . . u p ]; u i is a column vector with a size m
- [V] [v 1 , v 2 . . . v n ]; v i is a column vector with a size p
- I p is the identity matrix with a dimension (p,p).
- the diagonal matrix [S] is such that:
- Singular value decomposition is computed by applying the Golub/Reinsch algorithm relating to the Golub/Kahan method which is a standard one in matrix processing operations.
- This method consists, firstly, in transforming the matrix [A] into a bidiagonal matrix by the HOUSEHOLDER method and then, on the basis of this bidiagonal shape, in computing the singular values.
- This method is efficient, fast and convergent. However, it is difficult to adapt it to multiprocessor computation.
- the vectors u 1 and v 1 are given respectively by the vectors u k+1 and v k+1 respectively coming from the last iteration.
- This type of algorithm may give rise to problems of convergence and stability. However, it is well suited to applications that us only the first singular values.
- the computation time needed to carry out this decomposition irrespectively of the method adopted, represents only a minute part of the computation time which will be used for the recognition of the pattern in an image. Furthermore, the learning process is not subjected to the time constraints dictated during the recognition stage.
- the energy of the image is concentrated on the first coefficients (s 1 , S 2 , . . . ).
- the last coefficients (. . . s p-1 , s p ) correspond to the details and to the noise contained in the image.
- This property is used for applications to the recognition of shapes in the image.
- the recognition is obtained by considering the set [S] to be the signature of the model to be recognized.
- the shape recognition is then obtained in two stages.
- the first stage consists in carrying out a learning operation on the model to be recognized, defined by a rectangular (or square) matrix of the luminance values that most closely surround the object to the characterized.
- the learning operation consists in computing the three matrices [U], [V], [S] that define this model completely and uniquely.
- the second stage consists in making a search, in an image to be processed, of that part of an image which corresponds best to the model.
- This search stage is conducted by seeking to recognize the object A in the image M by means of its signature [S].
- the signatures [S lc ] of each of the image zones M lc that form the image M are computed and then compared with the signature S of the object to be recognized.
- An image zone M lc as illustrated in FIG. 3 is a matrix A' with a size (m,n) equal to that of the object matrix A, extracted from the image at the position (l,c).
- the corresponding signature [S lc ] is computed in the same way as that of the matrix A.
- This difference e measures the similarity of the signatures of the model and of the image zone examined.
- the search for the pattern in the image could theoretically be done by considering all the zones with a size m x n of the image by shifts and by making a computation, for each of these zones, of a signature and the difference between this signature and the signature of the object.
- the position of the pattern is then the position of the image zone from which the smallest difference e has been computed.
- This search done on the entire image, in taking account all the eigen values, entails an operation which, given the computations to be performed, is hardly realistic and this approach cannot be used to resolve the problem which is one of optimizing the ratio between computation time and reliability and robustness of the method.
- a first possibility would be to limit the number of eigen values. This is a delicate operation, for the number of eigen values to be preserved depends on the characteristics of the objects to be sorted out: if two objects are close together, then a large number of eigen values needs to be considered in order to differentiate them whereas, if these objects are very different, a limited number of eigen values would be enough.
- This method consists in making a selection, at each step, of a set of candidate image zones, according to an increasing resolution and on a search field of decreasing size:
- the first step selects a number n 1 of zones according to the descending-order values of the difference e computed from a single eigen value, for all the zones of the image.
- the second step entails a sorting, from among these n 1 image zones, of a number n 2 of image zones according to the descending-order values of the differences computed, in considering p 2 eigen values.
- the difference is written as: ##EQU4##
- n i zones are chosen, again according to a sorting criterion based on the descending-order values of differences, from among the n i-1 candidate zones chosen at the previous step, i-1, the difference being computed according to P i eigen values.
- the last selection step enables the selection of the image zone leading to the minimum difference from all the eigen values of the matrices corresponding to the zones chosen in the previous step.
- the number of zones and of eigen values chosen for the computations in the different steps are therefore such that:
- a final step consists in effecting a comparison between the signature of the object zone and the signature of the image zone chosen in the last step of the iterative method, the chosen image zone leading to the recognition of the object only if the signatures are close enough, the corresponding difference being, for example, smaller than a predetermined fixed threshold.
- a number of iterations equal to 0 has been used (irrespectively of the size of the learning object).
- the number of eigen values used corresponds to the rank of the iteration (1st iteration, 1 eigen value; 2nd iteration, eigen values; . . . 5th iteration, 5 eigen values).
- the size of the object to be searched for is assumed to be greater than 10 ⁇ 10 pixels.
- the first iteration therefore consists in a rough and fast selection of image zones liable to meet the characteristics of the object zone used for the learning, and the successive iterations enable an increasingly finer selection of the possible image zones, until the final step when only one zone is selected.
- Improvements are useful, notably if the object to be recognized has undergone a uniform variation in luminance from the model to the processed image.
- the object to be sought in the image may indeed have undergone a variation in luminance due to a different illumination (visible imaging) or to a temperature modification (infrared imaging).
- An image zone A is considered to be formed by two components:
- a luminance variation component AL also represented by a matrix with a size m ⁇ n.
- II is a matrix with a dimension n x m constituted by ones and where the coefficient k represents the difference in mean luminance between the candidate image zone and the learning object zone.
- the coefficient k should be computed for each sub-imate at the first iteration and should be preserved for the following iterations:
- the method has been implemented in introducing an offset on the luminance values of 100 (which introduces a saturation of the order of 2% in the image).
- the sensitivity of the method has been verified by adding a Gaussian noise to the luminance values: the difference increases as a function of the increase in the noise value (four eigen values have been used).
- the influence of the rotation of the object in the image has also been analyzed.
- the computed difference changes as a function of the rotation of the sub-imate. An increase in the difference as a function of the rotation angle is observed, but the effect of the difference on the detection of an object remains small for rotation angles of 10° and 20° successively. The recognition is no longer accurate for an angel greater than or equal to 30°.
- the set of eigen values represents a good signature of the object and enables reliability and robustness in the recognition of this object.
- each step corresponds to a resolution.
- the initialization is done with a rough resolution (on the first eigen values).
- the last step enables the selection of the object in a fine resolution (all the eigen values are considered).
- the first step of the method consists in performing a first sorting of the image zones on the first eigen value
- a sorting is done on the image zones selected in the previous step, on a greater number of eigen values
- This method can be used, for example, in an application to the tracking and identification, in a fixed sequence of images, of the signature of the tracked object which has to be regularly refreshed in order to take account of the modifications of shapes due to the observation at different angles and maskings often seen in applications of this type.
- This method is naturally implemented by means of appropriately programmed computer processing and calculating means which, on the basis of the luminance values of the pixels of the images, enable the selection, in a reference image, of an image zone corresponding to the object to be detected, the processing of this object zone in order to determine therefrom the signature as defined here above, and then the processing of the images in which the object is searched for, according to the above-described iterative method.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The disclosed method comprises a learning phase, during which the object zone to be recognized, selected in a learning image, is modelized, and a recognition phase during which the modelized zone is compared with image zones of a same size taken taken from the images to be processed. To characterize the image zones that are models of objects to be recognized or image zones to be processed, the method uses a decomposition of the matrices of the luminance values into singular values according to a matrix product of projection matrices and an ordered diagonal matrix, this ordered diagonal matrix constituting a characteristic signature of the zone considered; for the recognition phase, differences between the signature of the model zone and the signatures of the zones of the image considered are computed and image zones are selected on the basis of the least differences according to an iterative method which consists in conducting a rough sorting operation on a large number of zones in an initial stage, and then an increasingly finer sorting operation on an increasingly limited number of zones. The computations done in each step are re-used in the next step.
Description
1. Field of the Invention
The invention relates to the field of image processing and, more particularly, an object of the invention is a method for the recognition of shapes in images and its application to the tracking of objects in sequences of images.
2. Description of the Prior Art
Presently known methods used to recognize objects in images use binary correlation or multilevel correlation in the spatial domain or in a transformed domain, or methods of comparison using shape descriptors.
The recognition of objects in an image generally relies on a search, in the image, for the patterns that most resemble known models.
The model of recognition of an object may take different forms but always characterizes a precise representation of the characteristics of the object to be recognized. It may be a matrix of dots, a collection of geometric elements, a set of contours or, again, an analysis in a particular field (Fourier transform, Hadamard transform wavelets etc.).
These methods are mostly very costly in terms of computation time in cases where they are required to be very reliable and robust.
An object of the invention is a method, and the corresponding device, for the recognition of shapes, wherein the reliability and robustness of the results are obtained without any excessive cost in terms of computation time. To this end, the method uses, for this recognition, a decomposition into singular values of the image (or S.V.D. for singular value decomposition).
The singular value decomposition of an object represented by a matrix of dots characterizes this object by its "signature" on a projection space defined by two projectors. This signature, constituted by the singular values of the object matrix, defines the recognition model. An object is then defined by three matrices referenced U, V and S. The matrix S is an ordered diagonal matrix. All the values of the diagonal correspond to the signature of the object. The matrices U and V are the matrices of the projection.
An object is recognized if its signature, after projection, is close to the signature sought. For, there is an almost univocal relationship between the signature and the zone to which it corresponds.
According to the invention, the recognition is done according to an iterative method on the eigen values.
In the first step of the method, a first sorting operation is carried out on all the possible zones, in using only one first eigen value.
At the ith step, a sorting will be done from among the zones selected at the previous step on a greater number of eigen values.
The method therefore consists, initially, in carrying out a rough sorting operation on a large number of zones, and then in carrying out an increasingly fine sorting operation on an increasingly limited number of zones. The computations done in each step are re-used in the following step.
A first object of the invention is a method for the recognition of objects in images comprising:
a learning phase during which the object zone to be recognized, selected in a learning image, is modelized, and
a recognition phase during which the modelized zone is compared with image zones of a same size, taken from the images to be processed,
wherein, for the characterization of the image zones that are models of objects to be recognized or of the image zones to be processed, said method uses a decomposition of the matrices of luminance values into singular values according to a matrix product of projection matrices and an ordered diagonal matrix, said ordered diagonal matrix constituting a characteristic signature of the zone considered, and wherein, for the recognition phase, differences between the signature of the model zone and the signatures of the zones of the image considered are computed and image zones are selected on the basis of the least differences according to an iterative method such that:
in a first step, only the first coefficient of the ordered signature matrices is considered for the computation of the differences, but this is done for all the image zones possible;
then, in a second step, the p2 first coefficients of the signature matrices, p2 being greater than 1, are considered for a reduced set of n1 image zones selected in the previous step as image zones exhibiting the least differences;
in an ith step, a number pi, greater than pi-1, of matrix coefficients is considered for a reduced set of n1 image zones selected in the step (i-1) because they exhibit the smallest differences, with i varying from 3 to k-1.
in a kth and last step, all the coefficients of the signature matrices of the image zones selected in the second last step are considered to select the image zone having the smallest difference in signature with respect to the model zone.
A second object of the invention lies in the application of the method to the tracking of objects in sequences of images.
The invention will be understood more clearly and other features will appear from the following description, made with reference to the appended figures.
FIGS. 1 and 2 respectively illustrate an image in which there is selected a zone comprising an object to be recognized and a second image in which this object should be recovered;
FIG. 3 is an explanatory diagram;
FIG. 4 illustrates the recognition method according to the invention.
Singular values decomposition is a promising method in the field of image processing for applications of shape compression, filtering and recognition.
As indicated briefly here above, transformation by singular value decomposition consists in analyzing an image (or part of an image), in which the luminance values of the points form a matrix A with a size (m, n) that is rectangular or square, into a matrix product defined by:
[A]=[U] [S] [V].sup.t, where [V].sup.t is the transposed matrix of [V] and where the matrix S is an ordered diagonal matrix such that:
[S]=diag(s.sub.1,s.sub.2,s.sub.3,s.sub.4, . . . s.sub.p)
with p=min(m,n).
These diagonal elements s1 . . .sp are singular values of the matrix A and are positive real numbers (positive square roots of the eigen values) classified in descending order:
s1 >s2 >s2 >. . . >sp.
The matrices [U] and [V], with respective sizes of (m,p) and (p,n) group together the left-hand singular values (U) and right-hand singular values (V) of A in their columns. The columns ui of [U] are the eigen vectors of [A]t ·[A].
[U]=[u1,u2 . . . up ]; ui is a column vector with a size m
[V]=[v1, v2 . . . vn ]; vi is a column vector with a size p
These matrices have the following particular features:
[U].sup.t [U]=[V]·[V].sup.t =I.sub.p
where Ip is the identity matrix with a dimension (p,p).
The diagonal matrix [S] is such that:
[S]=diag (s.sub.1, s.sub.2, 2.sub.3, . . . s.sub.r, . . . s.sub.p)
with
s.sub.1 >s.sub.2 . . . s.sub.r >0; s.sub.r+1 = . . . =s.sub.p =o
where r is the rank of the matrix A. The eigen vectors and the singular values are defined by the following relationships:
A·v.sub.i =s.sub.i ·u.sub.i
A.sup.t u.sub.i =s.sub.i ·v.sub.i
The transformation by singular value decomposition can therefore be represented in the following form: ##EQU1##
Several methods have been developed to determine the matrices [U], [V], [S].
Singular value decomposition is computed by applying the Golub/Reinsch algorithm relating to the Golub/Kahan method which is a standard one in matrix processing operations.
This method consists, firstly, in transforming the matrix [A] into a bidiagonal matrix by the HOUSEHOLDER method and then, on the basis of this bidiagonal shape, in computing the singular values. This method is efficient, fast and convergent. However, it is difficult to adapt it to multiprocessor computation.
Another type of method, called the SVD power algorithm, is being currently developed. This method uses an iterative method:
su=[A].sub.t V
and
sv=[A].sup.t u
if k is the preceding iteration rank:
u.sup.(k+1) =[A]v(k)/||[A]v.sup.(k) ||
v.sup.(k+1) =[A]U(k+1)+1)/||[A].sup.t u.sup.(k+1) ||
The iterative method is continued until the instant when:
||u.sup.(k+1) -u.sup.(k) ||
is smaller than a fixed value. The first singular value is then equal to:
s.sub.1 =||[A].sup.t u.sup.(k+1) ||,
The vectors u1 and v1 are given respectively by the vectors uk+1 and vk+1 respectively coming from the last iteration.
To obtain the following pair of vectors (u2, v2), the method is reiterated in considering no longer the matrix A bu the matrix A' defined by:
[A]'=[A]-s.sub.1 u.sub.1 v1.sup.t
This type of algorithm may give rise to problems of convergence and stability. However, it is well suited to applications that us only the first singular values.
Finally, another category of algorithm is based on the Hestenes/Nash method known as "one side transformation" which enables searching for the singular values of a matrix through the application of sequences of plane rotations to the columns or lines. This method enables only a complete decomposition of the image and is especially suited to small-sized matrices.
The computation time needed to carry out this decomposition, irrespectively of the method adopted, represents only a minute part of the computation time which will be used for the recognition of the pattern in an image. Furthermore, the learning process is not subjected to the time constraints dictated during the recognition stage.
The choice of the decomposition method is therefore not of critical importance, and either of the above described methods may be used, since the optimization is indispensable above all at the level of the recognition method proper.
The algorithm used to compute the different matrices depends on the type of computer (parallel, series etc.) that will be used to compute the singular value decomposition, and is not described in detail in the present description since it is within the scope of those skilled in the art.
In using a decomposition such as this, the energy of the image is concentrated on the first coefficients (s1, S2, . . . ). The last coefficients (. . . sp-1, sp) correspond to the details and to the noise contained in the image.
This property is used for applications to the recognition of shapes in the image. The recognition is obtained by considering the set [S] to be the signature of the model to be recognized. The shape recognition is then obtained in two stages.
The first stage consists in carrying out a learning operation on the model to be recognized, defined by a rectangular (or square) matrix of the luminance values that most closely surround the object to the characterized. The learning operation consists in computing the three matrices [U], [V], [S] that define this model completely and uniquely.
The second stage consists in making a search, in an image to be processed, of that part of an image which corresponds best to the model.
This search stage is conducted by seeking to recognize the object A in the image M by means of its signature [S]. To this end, the signatures [Slc ] of each of the image zones Mlc that form the image M are computed and then compared with the signature S of the object to be recognized.
An image zone Mlc as illustrated in FIG. 3 is a matrix A' with a size (m,n) equal to that of the object matrix A, extracted from the image at the position (l,c). The corresponding signature [Slc ] is computed in the same way as that of the matrix A.
The comparison between the two signatures, [S] and Slc ], which should enable an appreciation of the resemblance, is done by computing the difference or distance between the two matrices [S] and [Slc ]. This difference is equal to: ##EQU2##
with
[S.sub.lc ]=diag(s'.sub.2, . . . s'.sub.2. . . s'.sub.p)
This difference e measures the similarity of the signatures of the model and of the image zone examined.
The search for the pattern in the image could theoretically be done by considering all the zones with a size m x n of the image by shifts and by making a computation, for each of these zones, of a signature and the difference between this signature and the signature of the object. The position of the pattern is then the position of the image zone from which the smallest difference e has been computed.
This search, done on the entire image, in taking account all the eigen values, entails an operation which, given the computations to be performed, is hardly realistic and this approach cannot be used to resolve the problem which is one of optimizing the ratio between computation time and reliability and robustness of the method. There is therefore provision, according to the invention, to limit the number of computations. A first possibility would be to limit the number of eigen values. This is a delicate operation, for the number of eigen values to be preserved depends on the characteristics of the objects to be sorted out: if two objects are close together, then a large number of eigen values needs to be considered in order to differentiate them whereas, if these objects are very different, a limited number of eigen values would be enough.
To limit the number of computations, a so-called "top down" iterative method has been used. This method consists in making a selection, at each step, of a set of candidate image zones, according to an increasing resolution and on a search field of decreasing size:
the first step selects a number n1 of zones according to the descending-order values of the difference e computed from a single eigen value, for all the zones of the image.
The difference for each zone is computed as follows:
e=(s.sub.1 -s'.sub.1).sup.2
where: ##EQU3##
The n1 positions of the zones leading to the smallest differences are kept and will be used in the next step.
the second step entails a sorting, from among these n1 image zones, of a number n2 of image zones according to the descending-order values of the differences computed, in considering p2 eigen values. The difference is written as: ##EQU4##
At the ith step, ni zones are chosen, again according to a sorting criterion based on the descending-order values of differences, from among the ni-1 candidate zones chosen at the previous step, i-1, the difference being computed according to Pi eigen values.
The last selection step enables the selection of the image zone leading to the minimum difference from all the eigen values of the matrices corresponding to the zones chosen in the previous step. The number of zones and of eigen values chosen for the computations in the different steps are therefore such that:
n.sub.1 >n.sub.2 >n.sub.3 >. . . >1 (number of image zones)
1<p.sub.2 <p.sub.3 <. . . <p.sub.t (number of eigen values)
where pt represents the total number of eigen values.
A final step consists in effecting a comparison between the signature of the object zone and the signature of the image zone chosen in the last step of the iterative method, the chosen image zone leading to the recognition of the object only if the signatures are close enough, the corresponding difference being, for example, smaller than a predetermined fixed threshold.
In one embodiment, a number of iterations equal to 0 has been used (irrespectively of the size of the learning object).
Two constraints have been introduced:
1. For the first five iterations, the number of eigen values used corresponds to the rank of the iteration (1st iteration, 1 eigen value; 2nd iteration, eigen values; . . . 5th iteration, 5 eigen values). The size of the object to be searched for is assumed to be greater than 10×10 pixels.
2. In the last step, all the eigen values are considered.
The first iteration therefore consists in a rough and fast selection of image zones liable to meet the characteristics of the object zone used for the learning, and the successive iterations enable an increasingly finer selection of the possible image zones, until the final step when only one zone is selected.
For each iteration with a rank greater than 5, the number of eigen values taken into account has been computed according to the formula:
p.sub.i Ent [a*(exp(k*i)-1(]
where Ent signifies an integer part of the expression within square brackets,
and
a=25/(p.sub.t -100)
and
k=1/5 * log(p.sub.+ /5-1)
where
pt corresponds to the maximum number of eigen values;
with the above constraints
P5 =5 and P10 =pt.
Improvements are useful, notably if the object to be recognized has undergone a uniform variation in luminance from the model to the processed image. The object to be sought in the image may indeed have undergone a variation in luminance due to a different illumination (visible imaging) or to a temperature modification (infrared imaging).
In this case, the method as defined here above does not enable the recover of the object in the image. It is then necessary to introduce the following improvement:
An image zone A is considered to be formed by two components:
an information component Al represented by a matrix with a size m×n;
a luminance variation component AL also represented by a matrix with a size m×n. These two matrices are such that:
A=A1+AL
and
Al=k×II
where II is a matrix with a dimension n x m constituted by ones and where the coefficient k represents the difference in mean luminance between the candidate image zone and the learning object zone.
The sequence of the stages of the recognition method is identical to that described here above, except for the following modification:
The contribution of the luminance variation is removed from the matrix S' to give the reduced matrix:
SR'=S'-k·Ut * II * V
where the coefficients of the diagonal matrix SR' are such that:
sr'.sub.i =s'.sub.i -k corr (i,j)
The coefficient k should be computed for each sub-imate at the first iteration and should be preserved for the following iterations:
The value corr (i,j) is such that: ##EQU5## with k=1, m and r=1, n.
The elements of this matrix should be completed for each new iteration.
The difference e for each sub-image is then written as: ##EQU6## with i varying between 1 and pi.
For a uniform luminance variation and a small number of cases of saturation (this excludes any case an excessive saturation and non-linear responses from the sensors), the method has been implemented in introducing an offset on the luminance values of 100 (which introduces a saturation of the order of 2% in the image).
The sensitivity of the method has been verified by adding a Gaussian noise to the luminance values: the difference increases as a function of the increase in the noise value (four eigen values have been used).
the influence of the rotation of the object in the image has also been analyzed. The computed difference changes as a function of the rotation of the sub-imate. An increase in the difference as a function of the rotation angle is observed, but the effect of the difference on the detection of an object remains small for rotation angles of 10° and 20° successively. The recognition is no longer accurate for an angel greater than or equal to 30°.
The set of eigen values represents a good signature of the object and enables reliability and robustness in the recognition of this object.
For the implementation of the method, an algorithm optimizing the computation time has therefore been defined. Each step corresponds to a resolution. The initialization is done with a rough resolution (on the first eigen values). The last step enables the selection of the object in a fine resolution (all the eigen values are considered). To this end, in short:
the first step of the method consists in performing a first sorting of the image zones on the first eigen value;
at the ith step, a sorting is done on the image zones selected in the previous step, on a greater number of eigen values;
at the last step, only one image zone is chosen, in considering all the eigen values for the image zones of the previous step.
This method can be used, for example, in an application to the tracking and identification, in a fixed sequence of images, of the signature of the tracked object which has to be regularly refreshed in order to take account of the modifications of shapes due to the observation at different angles and maskings often seen in applications of this type.
This method is naturally implemented by means of appropriately programmed computer processing and calculating means which, on the basis of the luminance values of the pixels of the images, enable the selection, in a reference image, of an image zone corresponding to the object to be detected, the processing of this object zone in order to determine therefrom the signature as defined here above, and then the processing of the images in which the object is searched for, according to the above-described iterative method.
Claims (4)
1. In a pattern matching image processing method wherein a matching zone in an image to be read is matched with a template image, is searched and determined, comprising the steps of:
selecting, in a preliminary learning phase, an object zone to be recognized in a learning image;
modelizing, in a recognition phase, said modelized zone in a new image to be processed by considering several image zones of an identical size as said modelized zone, taken from said new image, by decomposing matrices of luminance values of said modelized zone and of each of said image zones into singular values according to a matrices product of a projection matrix and an order diagonal matrix which comprises a characteristic signature of each zone under consideration, and by computing differences between the signature of said model zone and the signature of each image zone according to an iterative method comprising the following steps:
in a first step, computing said differences by using only the first coefficient of the order diagonal matrices, and selecting a set of ni image zones exhibiting the least differences;
in a second step, computing said differences by using a set of p2 coefficients of the order diagonal matrices relative to said modelized zone and to said set of n1 image zones, for p2 greater than 1, and selecting a set of n2 image zones exhibiting the least differences;
in an ith step, computing said differences by using a set of pi coefficients of the order diagonal matrices relative to said modelized zone and to a set of ni-1 image zones selected in the step (i-1), with pi greater than pi+1, and selecting a set of ni image zones exhibiting the smallest differences, with i varying from 3 to K-1
in a Kth step, computing said differences by using all the coefficients of the order diagonal matrices, and selecting a last image zone exhibiting the smallest difference;
a final step of comparison between the signatures according to which the object of the model zone is recognized in the sub-image selected at the end of the recognition phase only if its signature is close to the signature of the model zone, the difference between their respective signatures being below a predetermined threshold.
2. A method according to claim 1, wherein (s1, . . . s'p) are set as coefficients of the ordered signature matrix of an image zone under consideration wherein said difference is computed at the first step and is defined as (s1 -s'1) and at a step j subsequent to said ith step, the difference computed is set equal to ##EQU7## i=1 to pj where j is the rank of the step of the iterative method and pj is the number of eigen values used at said step.
3. A method according to claim 1 or 2, wherein the iterative method is such that, for the first steps, the number of coefficients of the signature matrices used for the computation of differences is equal to the iteration rank and wherein, for the following steps, the number of coefficients if determined so as to increase gradually up to the final value pt, where pt is the maximum number of coefficients of the signature matrices.
4. A method according to claim 3 wherein, when the object to be processed undergoes luminance variations, the signature matrix of the sub-image analyzed is processed beforehand for the extraction therefrom of the variation in luminance, so that the signatures of the model zone of the object and of the analyzed sub-image are comparable.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
FR9012445 | 1990-10-09 | ||
FR9012445A FR2667711B1 (en) | 1990-10-09 | 1990-10-09 | METHOD FOR RECOGNIZING OBJECTS IN IMAGES AND ITS APPLICATION TO TRACKING OBJECTS IN IMAGE SEQUENCES. |
Publications (1)
Publication Number | Publication Date |
---|---|
US5245675A true US5245675A (en) | 1993-09-14 |
Family
ID=9401064
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US07/772,328 Expired - Fee Related US5245675A (en) | 1990-10-09 | 1991-10-07 | Method for the recognition of objects in images and application thereof to the tracking of objects in sequences of images |
Country Status (5)
Country | Link |
---|---|
US (1) | US5245675A (en) |
EP (1) | EP0480806A1 (en) |
JP (1) | JPH04299469A (en) |
KR (1) | KR920008624A (en) |
FR (1) | FR2667711B1 (en) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5625715A (en) * | 1990-09-07 | 1997-04-29 | U.S. Philips Corporation | Method and apparatus for encoding pictures including a moving object |
US5710833A (en) * | 1995-04-20 | 1998-01-20 | Massachusetts Institute Of Technology | Detection, recognition and coding of complex objects using probabilistic eigenspace analysis |
WO1998043105A2 (en) * | 1997-03-21 | 1998-10-01 | Autodesk, Inc. | Object tracking system using statistical modeling and geometric relationship |
US6360013B1 (en) * | 1999-03-16 | 2002-03-19 | Academia Sinica | Fast method and system for template matching acquiring global optima |
US6385610B1 (en) * | 1998-10-16 | 2002-05-07 | Computer Associates Think, Inc. | Method and apparatus for identifying and visualizing differences in object model diagrams |
US6401007B1 (en) * | 1999-04-16 | 2002-06-04 | Kalish Canada Inc. | Gage element for performing automatic setting of a machine for manipulating articles, such as a palletization machine |
US6636619B1 (en) * | 1999-07-07 | 2003-10-21 | Zhongfei Zhang | Computer based method and apparatus for object recognition |
US20040022413A1 (en) * | 2002-08-01 | 2004-02-05 | Albus John E. | Correlation tracker breaklock detection |
US20040101200A1 (en) * | 2002-11-26 | 2004-05-27 | Larry Peele | Method and apparatus for image processing to detect changes in a scene |
US20050160401A1 (en) * | 1999-10-16 | 2005-07-21 | Computer Associates Think, Inc. | System and method for adding user-defined objects to a modeling tool |
US20060225030A1 (en) * | 1999-10-16 | 2006-10-05 | Deffler Tad A | Method and System for Generating Dynamic Comparison Models |
US20060285723A1 (en) * | 2005-06-16 | 2006-12-21 | Vassilios Morellas | Object tracking system |
DE102015007434A1 (en) | 2015-06-15 | 2016-12-15 | Mediabridge Technology GmbH | information device |
CN110533043A (en) * | 2018-05-23 | 2019-12-03 | 华南理工大学 | The accelerated method of the Recognition with Recurrent Neural Network for handwritten Kanji recognition based on SVD |
CN112912922A (en) * | 2018-10-29 | 2021-06-04 | 京瓷株式会社 | Image processing device, camera, moving object, and image processing method |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0792370B2 (en) * | 1993-06-02 | 1995-10-09 | 株式会社エイ・ティ・アール人間情報通信研究所 | Object and light source information extraction method and apparatus using the same |
CN117857040B (en) * | 2024-03-05 | 2024-06-04 | 建投物联股份有限公司 | Internet of things equipment security auxiliary method, device, equipment and storage medium |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5101441A (en) * | 1988-09-21 | 1992-03-31 | Yamaha Hatsudoki K.K. | Method of and device for processing image |
-
1990
- 1990-10-09 FR FR9012445A patent/FR2667711B1/en not_active Expired - Lifetime
-
1991
- 1991-10-04 EP EP91402654A patent/EP0480806A1/en not_active Withdrawn
- 1991-10-07 US US07/772,328 patent/US5245675A/en not_active Expired - Fee Related
- 1991-10-07 KR KR1019910017506A patent/KR920008624A/en active IP Right Grant
- 1991-10-08 JP JP3287335A patent/JPH04299469A/en not_active Withdrawn
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5101441A (en) * | 1988-09-21 | 1992-03-31 | Yamaha Hatsudoki K.K. | Method of and device for processing image |
Non-Patent Citations (4)
Title |
---|
IEEE Pacific Rim Conference on Communications, Computers and Signal Processing, Jun. 4 5, 1987, Victoria, B. C., Canada; IEEE New York, USA, pp. 77 80; Zhi Qiang Liu, et al.; Image Detection by Singular Value Decomposition of Image Matrices . * |
IEEE Pacific Rim Conference on Communications, Computers and Signal Processing, Jun. 4-5, 1987, Victoria, B. C., Canada; IEEE New York, USA, pp. 77-80; Zhi-Qiang Liu, et al.; "Image Detection by Singular Value Decomposition of Image Matrices". |
IEEE Transactions on Pattern Analysis and Machine Intelligence; vol. PAMI7, No. 3, May, 1985; New York, USA; pp. 338 344; Ardeshir Goshtasby; Template Matching in Rotating Images . * |
IEEE Transactions on Pattern Analysis and Machine Intelligence; vol. PAMI7, No. 3, May, 1985; New York, USA; pp. 338-344; Ardeshir Goshtasby; "Template Matching in Rotating Images". |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5625715A (en) * | 1990-09-07 | 1997-04-29 | U.S. Philips Corporation | Method and apparatus for encoding pictures including a moving object |
US5710833A (en) * | 1995-04-20 | 1998-01-20 | Massachusetts Institute Of Technology | Detection, recognition and coding of complex objects using probabilistic eigenspace analysis |
WO1998043105A2 (en) * | 1997-03-21 | 1998-10-01 | Autodesk, Inc. | Object tracking system using statistical modeling and geometric relationship |
US5845009A (en) * | 1997-03-21 | 1998-12-01 | Autodesk, Inc. | Object tracking system using statistical modeling and geometric relationship |
WO1998043105A3 (en) * | 1997-03-21 | 1998-12-30 | Autodesk Inc | Object tracking system using statistical modeling and geometric relationship |
US7711685B1 (en) | 1998-10-16 | 2010-05-04 | Computer Associates Think, Inc. | Method and system for an extensible macro language |
US6385610B1 (en) * | 1998-10-16 | 2002-05-07 | Computer Associates Think, Inc. | Method and apparatus for identifying and visualizing differences in object model diagrams |
US6360013B1 (en) * | 1999-03-16 | 2002-03-19 | Academia Sinica | Fast method and system for template matching acquiring global optima |
US6401007B1 (en) * | 1999-04-16 | 2002-06-04 | Kalish Canada Inc. | Gage element for performing automatic setting of a machine for manipulating articles, such as a palletization machine |
US6636619B1 (en) * | 1999-07-07 | 2003-10-21 | Zhongfei Zhang | Computer based method and apparatus for object recognition |
US7734457B2 (en) | 1999-10-16 | 2010-06-08 | Computer Associates Think, Inc. | Method and system for generating dynamic comparison models |
US20050160401A1 (en) * | 1999-10-16 | 2005-07-21 | Computer Associates Think, Inc. | System and method for adding user-defined objects to a modeling tool |
US20060225030A1 (en) * | 1999-10-16 | 2006-10-05 | Deffler Tad A | Method and System for Generating Dynamic Comparison Models |
US7046823B2 (en) * | 2002-08-01 | 2006-05-16 | Raytheon Company | Correlation tracker breaklock detection |
US20040022413A1 (en) * | 2002-08-01 | 2004-02-05 | Albus John E. | Correlation tracker breaklock detection |
US7149361B2 (en) | 2002-11-26 | 2006-12-12 | Lockheed Martin Corporation | Method and apparatus for image processing to detect changes in a scene |
US20040101200A1 (en) * | 2002-11-26 | 2004-05-27 | Larry Peele | Method and apparatus for image processing to detect changes in a scene |
US7720257B2 (en) * | 2005-06-16 | 2010-05-18 | Honeywell International Inc. | Object tracking system |
US20060285723A1 (en) * | 2005-06-16 | 2006-12-21 | Vassilios Morellas | Object tracking system |
DE102015007434A1 (en) | 2015-06-15 | 2016-12-15 | Mediabridge Technology GmbH | information device |
CN110533043A (en) * | 2018-05-23 | 2019-12-03 | 华南理工大学 | The accelerated method of the Recognition with Recurrent Neural Network for handwritten Kanji recognition based on SVD |
CN110533043B (en) * | 2018-05-23 | 2022-04-08 | 华南理工大学 | SVD-based acceleration method of recurrent neural network for handwritten Chinese character recognition |
CN112912922A (en) * | 2018-10-29 | 2021-06-04 | 京瓷株式会社 | Image processing device, camera, moving object, and image processing method |
US11675875B2 (en) * | 2018-10-29 | 2023-06-13 | Kyocera Corporation | Image processing apparatus, camera, mobile body, and image processing method |
Also Published As
Publication number | Publication date |
---|---|
FR2667711A1 (en) | 1992-04-10 |
JPH04299469A (en) | 1992-10-22 |
FR2667711B1 (en) | 1992-11-27 |
KR920008624A (en) | 1992-05-28 |
EP0480806A1 (en) | 1992-04-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5245675A (en) | Method for the recognition of objects in images and application thereof to the tracking of objects in sequences of images | |
US7450765B2 (en) | Increasing accuracy of discrete curve transform estimates for curve matching in higher dimensions | |
US7136505B2 (en) | Generating a curve matching mapping operator by analyzing objects of interest and background information | |
Wu et al. | Gabor wavelet representation for 3-D object recognition | |
US7139432B2 (en) | Image pattern matching utilizing discrete curve matching with a mapping operator | |
US20090074299A1 (en) | Increasing accuracy of discrete curve transform estimates for curve matching in four or more dimensions | |
Van de Weijer et al. | Curvature estimation in oriented patterns using curvilinear models applied to gradient vector fields | |
CN112001257A (en) | SAR image target recognition method and device based on sparse representation and cascade dictionary | |
Ravichandran et al. | Advanced in-plane rotation-invariant correlation filters | |
Elghoul et al. | Fast global SA (2, R) shape registration based on invertible invariant descriptor | |
US7171048B2 (en) | Pattern matching system utilizing discrete curve matching with a mapping operator | |
US7133538B2 (en) | Pattern matching utilizing discrete curve matching with multiple mapping operators | |
US7120301B2 (en) | Efficient re-sampling of discrete curves | |
Phillips | Matching pursuit filter design | |
Koutaki et al. | Fast and high accuracy pattern matching using multi-stage refining eigen template | |
Wang et al. | See SIFT in a rain: divide-and-conquer SIFT key point recovery from a single rainy image | |
Bourbakis et al. | Handwriting recognition using a reduced character method and neural nets | |
Tsatsanis et al. | Object detection and classification using matched filtering and higher-order statistics | |
Casasent et al. | Analysis of kernel distortion-invariant filters | |
Wahba et al. | Primitive concept identification in a given set of wafer maps | |
Ghods et al. | Accurate detection of human face position in the environmental images using Gabor wavelet transformations | |
Ozturk et al. | System of associative relationships (SOAR) | |
Wang et al. | Feature Transformation for Few-Shot Learning | |
Rahim et al. | Iris Recognition using Histogram Analysis via LPQ and RI-LPQ Method | |
Awwal et al. | Complex associative memory neural network model for invariant pattern recognition |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: THOMSON-CSF, FRANCE Free format text: ASSIGNMENT OF ASSIGNORS INTEREST.;ASSIGNORS:FERRE, ALAIN;VILLALON, YANNICK;REEL/FRAME:006408/0774 Effective date: 19911112 |
|
REMI | Maintenance fee reminder mailed | ||
LAPS | Lapse for failure to pay maintenance fees | ||
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 19970917 |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |