WO2008139399A2 - Method of determining motion-related features and method of performing motion classification - Google Patents

Method of determining motion-related features and method of performing motion classification Download PDF

Info

Publication number
WO2008139399A2
WO2008139399A2 PCT/IB2008/051843 IB2008051843W WO2008139399A2 WO 2008139399 A2 WO2008139399 A2 WO 2008139399A2 IB 2008051843 W IB2008051843 W IB 2008051843W WO 2008139399 A2 WO2008139399 A2 WO 2008139399A2
Authority
WO
WIPO (PCT)
Prior art keywords
motion
images
image
features
sequence
Prior art date
Application number
PCT/IB2008/051843
Other languages
French (fr)
Other versions
WO2008139399A3 (en
Inventor
Olivier Pietquin
Vasanth Philomin
Original Assignee
Philips Intellectual Property & Standards Gmbh
Koninklijke Philips Electronics N. V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Philips Intellectual Property & Standards Gmbh, Koninklijke Philips Electronics N. V. filed Critical Philips Intellectual Property & Standards Gmbh
Publication of WO2008139399A2 publication Critical patent/WO2008139399A2/en
Publication of WO2008139399A3 publication Critical patent/WO2008139399A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/20Movements or behaviour, e.g. gesture recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30196Human being; Person

Definitions

  • the invention relates to a method of determining motion-related features pertaining to the motion of an object, and to a method of performing motion classification for the motion of an object. Furthermore, the invention relates to a system for determining motion-related features pertaining to an object, and to a system for performing motion classification.
  • gestures One of the most natural communication means between humans is that of making movements with the hands or head, also called gestures, as a supplement to spoken communication, or as a means of communication in itself.
  • a simple augmentation to speech can comprise, for example, moving one's hand and arm to point in a particular direction when giving another person directions.
  • An example of communication solely by motion is sign language, based entirely on gestures. Since gesturing is second nature to humans, it follows that gesturing would be a natural and easy means of communication between humans and computers, for example in a home dialog system, in a dialog system in an office environment, for automatic sign language interpretation, etc. Owing to developments in technology in recent years, it is conceivable that the use of various dialog systems will become more and more widespread in the future, and that classification of a user's movements, i.e. gesture recognition, will play an increasingly important role.
  • gesture recognition systems have been proposed by state of the art methods. Generally, these proposed systems rely on image analysis to locate a user in each of a sequence of images and to analyse successive postures to determine the gesture he is making. Some systems require that the user wear a coloured suit or position sensors, a requirement which would be most undesirable for the average user of a home dialog system, for example. Other systems such as US 6,256,033 Bl propose extracting from an image a number of coordinates corresponding to certain points on the body of a user, and comparing these to coordinates corresponding to known gestures. An obvious disadvantage of such an approach is the very high level of computational effort required to analyse the images, and that the performance of the camera used to obtain the images must be sufficient to ensure sharp images.
  • WO 03/009218 proposes the use of two cameras to obtain a stereo image of a user making a gesture, and proceeds by removing the background in the images and tracking the upper body of the user using statistical framework for upper body segmentation of those parts of the images corresponding to the body of the user.
  • the computational level required for gesture recognition using this proposed system is very high, so that gesture recognition will be slow.
  • a further disadvantage of such systems is the amount of hardware involved, namely a high- resolution stereo camera or multiple cameras.
  • a disadvantage common to all of the conventional gesture recognition systems is that a user must perform a gesture slowly enough to ensure that a certain level of image quality is maintained, i.e.
  • the present invention provides a method of determining motion-related features pertaining to the motion of an object, which method comprises obtaining a sequence of images of the object and processing at least parts of the images to extract a number of first-order features from the images.
  • sequence of images can include all the images captured - for example by a camera - over a certain period of time, or only a selection of these images.
  • An image of an object for which motion-related features are to be determined can be processed in its entirety, or it may be that only a certain region of the image is processed, for instance, when only a part of the image is relevant.
  • the first- order features that are extracted from the image or image part can be blur-related features, gradient values, or any other real value that can be extracted from the image, such as, for example, a value describing the colour of a pixel.
  • the term "first-order" feature means a feature that can essentially be extracted from the image information in a first stage of computation.
  • pixel colour values can be directly extracted from the image information, and gradient values can be obtained by directly processing the image data.
  • a number of statistical values pertaining to the first-order features is computed in a second stage of computation, and these statistical values are then combined to give one or more histograms. Based on these histograms, motion-related features are determined for the object.
  • These second-order motion- related features can be any information given by or extracted from the histograms, identifying the areas in the image that are of relevance with regard to the motion of the object, for example, blurred areas in an image caused by a moving object.
  • the histograms themselves can be regarded as kind of a second-order motion-related feature.
  • object as used here is to be interpreted to mean a user, the body of a user, or part of the user's body, but could also be any moving object that is capable of moving in a pre-defined manner, such as a pendulum, a robot or other type of machine, etc.
  • object and user will be used interchangeably in the following.
  • the present invention also provides a method of performing motion classification for the motion of an object captured in a sequence of images, which method comprises determining motion-related features for the object captured in the images in the manner described above, and using the motion-related features to classify the motion of the object.
  • An obvious advantage of the method according to the invention is that motion or gesture classification can be achieved without resorting to the complex methods of image analysis required by the state of the art gesture recognition systems, and the algorithms used in the method according to the invention to extract the first- order features from an image are easily available and are well-known to a person skilled in the art, such as established algorithms used in image compression. Since statistical information pertaining to the first-order features is simply tallied or counted in histograms, which can then easily and very quickly be analysed to obtain the motion- related features, the proposed method offers clear advantages over image analysis methods used in the state of the art.
  • the method according to the invention does not require identification or localisation of the object, e.g. a person, in the image, and no modelling of body parts is required.
  • Another advantage is that a simple, low-resolution camera such as a webcam is quite sufficient for generating the images, allowing the system according to the invention to be realised cheaply and effectively. Furthermore, since the actual motion of an object does not need to be tracked directly, a low frame rate for the camera is sufficient, so that a cheap camera is quite sufficient.
  • a corresponding system for determining motion-related features pertaining to an object comprises an image source for providing a sequence of images of the object, a processing unit for processing at least parts of the images to extract a number of first- order features from the image, a computation unit for computing a number of statistical values pertaining to the first-order features, a combining unit for combining the statistical values to give a number of histograms and for determining the motion-related features for the object, based on these histograms.
  • a corresponding system for performing motion classification comprises such a system for determining motion-related features pertaining to the object, and also a classification unit for classifying the motion of the object using the motion- related features.
  • a digital image generated by an image source such as a camera contains various types of characterizing information, such as brightness, colour or grey-scale values, contour or edge sharpness, etc., which, as mentioned above, can be used as first- order features for the method according to the invention.
  • characterizing information such as brightness, colour or grey-scale values, contour or edge sharpness, etc.
  • colour or grey-scale values might be of interest, for example to track the colour of the object as it appears to move across the image.
  • information describing the level of blur in an image or a part of an image can be particularly informative.
  • the first-order features extracted from the image are blur- related features.
  • Information given by contours or edges in an image is particularly useful for image processing techniques such as object or pattern recognition. Some edges in the image might be sharp (in focus) or blurred (out of focus). For a camera with sufficient depth of field, a blurred part in an image region will indicate that an object in that part of the image was moving while the image was being generated.
  • extraction of blur-related features from an image is preferably carried out by performing a wavelet transform on image data, for a number of scales, to determine a series of wavelet coefficients for each point in the image.
  • the wavelet coefficients because their values depend on the discrepancies in value of neighbouring pixels, and the level of discrepancy depends in turn on the level of blur in the image, are blur-related features which ultimately provide motion information about the image.
  • Such a method, using blur-related features as first-order features is particularly advantageous since the very fact that a moving hand, head etc. will introduce blur into an image is used to positive effect. Blur is no longer an undesirable side-effect, but provides valuable information upon which the motion classification, e.g. gesture recognition, is based.
  • the type of wavelet transform to be used might depend to some extent on the size of the image.
  • the image data of a larger image with sufficient resolution could be used as input to a dyadic wavelet transform (DWT), whereas a continuous wavelet transform (CWT) could be used for images of small size, having only a relatively low resolution, such as those generated by a typical webcam.
  • DWT dyadic wavelet transform
  • CWT continuous wavelet transform
  • the wavelet coefficients thus obtained might then be processed to obtain blur-related features for the image.
  • Each pixel or point of the image corresponds to a wavelet coefficient in each scale of the transform, and each scale has the same number of points as the image. Therefore, after carrying out a 10-scale transform, there will be ten coefficients for each point in the image.
  • the modulus of the coefficient associated with the pixel will be a maximum in each scale.
  • the evolution of the ten coefficients across scales associated with the edge pixel provides information about the derivability of the image at that point.
  • the image gradient can be computed for that pixel.
  • the direction and the intensity, or magnitude, of the image gradient provide information about the direction of blur and the velocity of an object if the considered pixel belongs to a blurred edge of the moving object. Since the gradient direction is orthogonal to the edge or contour, it provides information about the shape of the object.
  • the log of slope of the coefficients' evolution across scales provides an estimate of the so-called Lipschitz coefficient, or exponent, for that pixel.
  • the Lipschitz exponent is a measure of the derivability of the image at the considered pixel and therefore a measure of the degree of continuity. If this Lipschitz exponent is positive, the pixel is located on a smooth or blurred edge. Equally, if the Lipschitz exponent is zero or negative, the pixel is located on a sharp edge. Performing the wavelet transform over a number of scales effectively provides a third dimension in addition to the two dimensions of the image, and it is in this third dimension that the evolution of the coefficients is observed.
  • a wavelet transform is generally performed for a series of pixels, such as the pixels from a certain row or certain column in a set of image data. Therefore, Lipschitz exponents and the image gradient intensity and direction computed from the resulting wavelet coefficients provide information about the degree of blur for neighbouring pixels in a row or column and the direction of the blur, if applicable. For example, a set of mostly negative Lipschitz exponents for a row of pixels would indicate that this row of pixels comprises well-defined edges. On the other hand, mostly positive- valued Lipschitz exponents would indicate that there are smooth transitions between the pixels in this row, implying a lack of sharp edges and therefore possibly that this row belongs to a blurred region in the image.
  • the wavelet transform might be performed across successive rows of pixels and successive columns of pixels, so that the wavelet transform is performed numerous times for each image or part of the image.
  • First-order features such as the Lipschitz exponents and wavelet coefficient gradients described above are computed for each wavelet transform operation, i.e. for each processed row and column.
  • An effective way of collating the information provided by the first-order features might be to round up the values for the first-order features to certain discrete values, and to simply count the number of occurrences of each value. Therefore, according to the invention, statistical values, such as the number of occurrences of a particular coefficient, pertaining to a certain kind of first-order feature extracted from at least part of an image are combined in a histogram for this first-order feature.
  • a counter may be assigned to each discrete coefficient value, and, whenever this value occurs, the counter is incremented by one. The values thus accumulated are tallied or collected in the histogram, which can be visualised as a simple bar chart.
  • a number of images might be captured, for example by a camera, following the movement of the user from start to finish.
  • Such a motion or movement is generally referred to as a gesture, and the various stages or positions in the movement are generally referred to as postures, without limiting the scope of the invention in any way.
  • An overall gesture is then given by a sequence of postures, and each posture in turn is captured in a sequence of images.
  • the images will show the user and a blurred region to the left of the image, which blurred region effectively "moves" across the images over time.
  • first-order features can be extracted for all of the images in such a volume, and combined to give motion-related information about the entire sequence.
  • first-order features are extracted for the virtual volume, for example by performing the wavelet transform for each of the images in the sequence, and combining statistical values for these features in one or more volume histograms.
  • a sequence of specific areas is identified in a sequence of images, and first-order features are extracted for the sub-volume given by the sequence of specific areas over time. Identification of the specific area of interest in an image is easily achieved by means of the first-order features for that image, for example by analysing the number of occurrences of a first-order feature for various regions of one of a sequence of images. For example, an entire image can be partitioned or sub-divided into a number of segments or tiles, preferably in such a way that the segment or tiles overlap.
  • a sub-image histogram for a first-order feature can be compiled for each segment or tile.
  • a sub-image histogram of mostly zero values or positive values for Lipschitz coefficients indicates that this region of the image contains blur. Therefore, all regions of the image containing blur can be quickly identified.
  • a selection can be formed comprising these regions which can be visualised by an imaginary rectangle drawn around a blurred moving arm in an image, and defined, for example, by the image coordinates for this rectangle. These image coordinates can then be used to locate the same specific area in all the images of a sequence.
  • each tile or segment can be analysed by pixel colour value, for instance for a moving object with a colour different from the background colour in the images. Only tiles containing pixels of a certain colour might be regarded as being of interest, so that the remaining tiles might be disregarded.
  • the step of locating the specific area of interested can be carried out at regular intervals, so that the specific area tracks the moving object in the images. In this way, even motion or gestures involving movement of the user over the entire image can be analysed. Any motion or gesture can be broken down into a series of essentially distinct consecutive stages, which, when taken together, give the overall motion or gesture.
  • the term "position" refers to any stage in the motion of an object or user that is captured in a sequence of images. Histograms of statistical values for one or more first-order features can be compiled for each specific area of a sequence of images. When considered together, therefore, the information in these histograms is characteristic of that sequence of specific areas. And, since a sequence of images can be used to determine a position being held or a posture being made, the information of the histograms of a sequence of specific areas is therefore characteristic of the position being held or the posture being made.
  • the statistical values pertaining to a type of first-order feature extracted from the sequence of specific areas in a sequence of images are combined in a volume histogram for that feature.
  • a volume histogram describing a position of an object, a posture or part of a gesture, will be used in the motion classification process described in more detail below. While the regions of the image defined by the specific area contain useful information about the type of motion or gesture being made, the remaining regions of the image - the "complementary area" - can also be of use.
  • the complementary area essentially comprises the pixels not included in the specific area.
  • a specific area in an image when a specific area in an image is identified, its complementary area is also identified, and, for a sequence of complementary areas, statistical values pertaining to a certain kind of first-order feature extracted from the sequence of complementary areas are combined in a complementary volume histogram for this first-order feature.
  • the specific area and/or the complementary area of an image are preferably identified by means of the first-order features determined for that image as explained above.
  • an appropriate boosting algorithm can be used to analyse the image data in order to identify the area or areas of interest.
  • Various boosting algorithms are available and will be known to a person skilled in the art.
  • the motion-related features obtained from a sequence of images can be used to classify the motion made by an object and captured in the sequence of images.
  • a motion-related feature might be a "distance" between two histograms. Such a distance might be computed between histograms of a specific area and the corresponding complementary area in an image, or it might be calculated between histograms of successive images in the image sequence.
  • a distance between histograms can be computed using a standard statistical distribution technique or by comparing the main properties of the histograms, such as the main gradient direction or main Lipschitz coefficient in the case of blur-related first-order features, or any other simple property of the histogram.
  • volume histograms sub-volume histograms and complementary volume histograms (simply referred to as "volume histograms" collectively in the following), generated as described above using the first-order features of the image, can ultimately describe or characterise the positions of an object in motion. Therefore, in a preferred method according to the invention, such volume histograms are analysed to obtain position-characteristic features which can be used to classify the motion.
  • position-characteristic features such as a ratio of positive histogram values to negative histogram values in a volume histogram for a sequence of images can give an indication of the position and direction of motion of the object.
  • Another type of position-characteristic information might be a distance between two volume histograms for different image sequences.
  • volume histogram One way of using a volume histogram might be to compare it, or a derivative of the volume histogram, to previously generated data for various different motion sequences or for a variety of gestures. For example, a volume histogram obtained for a posture tracked by a sequence of images can be compared to a number of
  • prototype postures for a collection of gesture models, such as a collection of state- transition models for a number of gestures.
  • one or more candidate gestures can be determined which comprise the posture corresponding to the volume histogram.
  • the number of candidate gestures can be narrowed down until the gesture most likely corresponding to the sequence of postures has been identified.
  • the position- characteristic features for a motion are used to identify a state in a generative model which is based on position or posture sub-units for a motion or gesture, e.g. a state-transition model for a motion.
  • a state corresponds to a particular position in that motion, for example, to a certain posture of a user or to a certain stage in the motion of an object.
  • a preferred state-transition model well known to those skilled in the art is the Hidden Markov Model (HMM), and is used to determine the probability that, in a gesture or movement, a certain position is followed by another position.
  • HMM Hidden Markov Model
  • a distance between a sub-volume histogram and its complementary volume histogram could be calculated.
  • any appropriate method of statistical distribution comparison for obtaining a single real number from a series of volume histograms can be applied.
  • These single values can be used as input to a suitable algorithm, for example a boosting algorithm, in order to build a discriminative criterion.
  • a distance between a volume histogram, generated as described above using a sequence of specific areas in a sequence of images, and a complementary volume histogram, generated as described above using the corresponding sequence of complementary areas in a sequence of images is calculated to give a position-characteristic feature, which feature can then be used in the motion classification procedure.
  • position-characteristic features are derived from the histograms (second-order motion-related features), they can also be regarded as a kind of third- order motion-related feature. Using these simple third-order features, a motion model corresponding to the positions in a motion can be identified, provided the prototype positions are characterised by this type of feature.
  • the steps of the methods for obtaining motion-related features and for performing motion classification as described above can be realised in the form of software modules to be run on a processor of a corresponding system for obtaining motion-related features or a system for performing motion classification, respectively.
  • Some of the functions for example certain image processing steps such as wavelet transforms in the case of extracting blur-related features, might be realised in the form of hardware, for example, as a dedicated application specific integrated circuit (ASIC) or field-programmable gate array (FPGA).
  • ASIC application specific integrated circuit
  • FPGA field-programmable gate array
  • Fig. 1 is a schematic representation of a camera capturing an image of a user making gesture
  • Fig. 2a is a schematic representation of a sequence of images of a person with associated motion information and blur-related features
  • Fig. 2b is a schematic representation of an image sub-divided into sub-images according to the invention.
  • Fig. 2c is a schematic representation of a specific area in an image according to the invention.
  • Fig. 3 a is a schematic representation of a virtual volume given by a sequence of specific areas from a sequence of images, and its associated volume histograms;
  • Fig. 3b is a representation of frontal view of a virtual volume, showing superimposed specific areas of a sequence of images;
  • Fig. 4 shows a state diagram of a state-transition model for a gesture, and a number of volume histograms;
  • Fig. 5 is a block diagram of a system for gesture recognition according to the invention.
  • Fig. 6 shows a flow chart for the steps of the method of performing gesture recognition according to the invention.
  • Fig. 1 shows a user 1 making a gesture in front of a camera 2, such as a webcam 2.
  • a camera 2 such as a webcam 2.
  • the user 1 is moving his arm in the direction of motion M, as a gesture or as part of a gesture.
  • the webcam generates an image f ⁇ of the user 1, and this is forwarded to a suitable processing unit, not shown in the diagram, such as a personal computer.
  • a suitable processing unit not shown in the diagram, such as a personal computer.
  • Any suitable type of camera can be used, but even a low-cost webcam 2 with a typical resolution of 320 x 480 pixels is sufficient for the purpose.
  • the user 1 can make gestures associated with commands commonly occurring in a dialog between the user 1 and a dialog system, e.g.
  • the webcam 2 can generate images of the user 1 at more or less regular intervals, giving a sequence of images fi, f 2 , ..., f n as shown in Fig.2a. From this sequence of images fi, f 2 , ..., f n , the first-order features can first be extracted.
  • the first- order features in the following examples used to illustrate the proposed method are blur- related features, chosen because of the advantages described in detail above. Without limiting the invention in any way, it is assumed that a method of gesture classification is being described.
  • FIG. 2a various snapshots of the user 1 have been captured in the sequence of images fi, f 2 , ..., f n . These snapshots of the user, taken together, combine to form a certain position or posture in an overall gesture. Edge- detection is then performed on an image fi, by carrying out a wavelet transform on the rows and columns of pixels in the image fi. Depending on the size of the image fi and the resolution of the camera 2, it might suffice to use every nth pixel, or it may be necessary to use each pixel. A set of wavelet coefficients is obtained in this way for each pixel, containing information about the discrepancies between neighbouring pixels, i.e. motion information given by the level of blur.
  • discrete blur-related features are derived, in this case the Lipschitz exponents and the image gradient.
  • the occurrences of each value of first-order feature are counted, for example the number of times a Lipschitz coefficient of 1.5 or an image gradient value of 0.5 ⁇ occurs, and collected in a histogram H LC , H G for that first-order feature.
  • the resulting histogram H LC , H G therefore contains blur-related information about the overall image fi, for example the proportion of the image fi that is blurred, or the level of sharpness of the edges in the image fi.
  • This process can be repeated for each image of the remaining images f 2 , ..., f n in the sequence.
  • performing edge analysis over the entire image is wasteful of resources, and can easily be circumvented.
  • a type of "sliding window" could be used to virtually travel over the image data, giving smaller sub- windows for which the wavelet transform is carried out.
  • the dimensions of the sliding window could also vary over time.
  • Fig. 2b shows another technique of reducing computational effort, easier to present visually.
  • an image f ⁇ is virtually divided into smaller sections, tiles, or sub-images.
  • Edge detection can be performed for each of the rows and columns in each of the sub-images of the overall image f ⁇ .
  • One such sub-image 20 is shown, for which first-order feature histograms sH L c, SHQ have been derived. Since the process is carried out for each sub-image in the image f ⁇ , such histograms are derived for each of the sub-images. These can then be analysed to decide which of the sub-images are actually of interest. Since the interesting elements of a user making a gesture are his moving limbs such as arm, hand, etc., and since moving elements will be accompanied by a certain level of blur, the "interesting" parts of the image f ⁇ can easily be located by examining the first-order feature histograms to determine which of them is characteristic of a blurred sub-image. As mentioned above, these interesting parts of the image can be located easily by using the type of "sliding window" on the image data, effectively minimising the computational effort.
  • Fig. 2c two such sub-images 21, 22 have been located in the image ft.
  • the image f ⁇ could be sub-divided into many more sub-images than are shown here, and the sub-images could also overlap.
  • only two sub-images 21, 22 of interest are shown.
  • These sub-images 21, 22 combine to form a specific area Ai, indicated in the diagram by a thick rectangle.
  • a user positioned in front of a camera or webcam and making gestures will usually stand more or less in the same place, and only move one or both of his hands and arms. Therefore, the specific area identified in one image or frame can simply be propagated from one image in the sequence to the next in order to define a specific area in each of the following images or frames in a sequence. This saves on computational power and resources, since only the image data in the specific areas actually changes significantly over time.
  • the rest of the image - the "complementary area" - remains to all intents and purposes the same, and can be regarded as stationary. This is illustrated in Fig. 3a, which shows a sequence of images fi, f 2 , ..., f n stacked vertically one behind the other.
  • a virtual volume V results.
  • the specific area Ai shown in the first frame fi is propagated through to all the following frames f 2 , ..., f n , so that these frames f 2 , ... , f n have their specific areas A 2 , ... , A n in the same relative positions.
  • a virtual sub-volume V s is created, excluding the complementary areas Ai', A 2 ', ..., A n ' of the images fi, f 2 , ..., f n .
  • first-order feature histograms are computed for each of the sub-images in the specific areas Ai, A 2 , ..., A n of the frames fi, f 2 , ..., f n .
  • a state-transition model G for a motion or gesture is shown, in this case a discrete Hidden Markov Model (HMM), in which a finite number of states Si, S 2 , S3, S4 correspond to a number of prototype positions of a motion, or postures of a gesture.
  • HMM Hidden Markov Model
  • This type of model is used to classify a motion, e.g. a gesture made by the user.
  • Transitions are weighted by the probabilities of stepping from one state to another, as shown, for example, by the probabilities P(S 3 ->S 4 ), P(Si->S 3 ), P(S 4 ->S 4 ), which give a measure of the probability of stepping or making the transition from state 3 to state 4, from state 1 to state 3, and from state 4 to state 4, respectively.
  • a volume histogram of a particular posture made by the user and captured in a sequence of images can be compared to a collection of prototype postures associated with various gestures.
  • the diagram shows four prototype posture histograms PHi, PH 2 , PH 3 , PH 4 for the states Si, S 2 , S 3 , S 4 of the state transition model G, but there may be any number of such state transition models available, each of which can be associated with various prototype postures.
  • These prototype posture histograms PHi, PH 2 , PH 3 , PH 4 can have been generated previously using a suitable learning algorithm.
  • volume histogram of a posture When comparing a volume histogram of a posture with the prototype postures, it may be found to be most similar to, for example, the prototype posture PHi associated with state Si of the state transition model.
  • the comparison simply involves computing a distance, as already described above, between the volume histogram and the prototype histogram and converting this distance into a probability that is a number between 0 and 1.
  • Subsequent volume histograms collected over time may be found to correspond to the consecutive states S 2 , S 3 , S 4 so that the gesture being made by the user would in all likelihood correspond to the gesture modelled by this state-transition model G.
  • the classification procedure can conclude that this state transition model G does not correspond to a candidate gesture.
  • Fig. 5 shows a block diagram of a system 4 for performing gesture classification, without limiting the scope of the invention in any way, comprising a camera 2 for obtaining a sequence of images f ⁇ , f 2 , ..., f n of a user 1 (in the diagram, only a single image f ⁇ is indicated).
  • Each image is first processed in a processing unit 5 to obtain a number of blur-related features 12.
  • the processing involves carrying out a wavelet transform on the image data to obtain Lipschitz exponents and wavelet coefficient gradients.
  • These blur-related first-order features 12 are then forwarded to a computation unit 6, in which statistical values 14 are computed for the first-order features 12.
  • the values of the first-order features 12 are first rounded up or down , as appropriate, to the closest of a set of pre-defined discrete values, such as -0.5, 0.0, 0.5, 1.0, etc., before the occurrences of each discrete value are counted, or tallied, to give a number of counts 14.
  • the counts 14 are combined to give a number of histograms H LC , H G , one for each type of blur-related feature, in this case two (Lipschitz exponents and wavelet coefficient gradients), which can be regarded as one kind of motion-related features pertaining to the image.
  • the histograms H LC , H G can be used to derive further motion- related features F m .
  • the ratio of the total number of positive values to the total number of zero or negative values in a histogram H LC , H G can be supplied as a motion-related feature F m for that histogram H LC , H G .
  • a histogram H LC , H G itself can be used as a motion-related feature F m .
  • the combining unit 15 might comprise a separate analysis unit for performing any required histogram analysis, but is shown as a single entity in the illustration.
  • the units and blocks 2, 5, 6, 15 described up to this point comprise a system 3 for determining motion-related features, and which comprises the front end of the system 4 for performing gesture classification.
  • the histograms H LC , H G are forwarded to a second combining unit 8, where they are collected and combined with previous histograms derived from image data of previous images in the sequence of images, or virtual image 'volume', generated by the camera 2 of the user 1 performing a gesture.
  • the output of the second combining unit 8 is then a set of volume histograms VH LC , VH G , VH LC ', VH G ', where the volume histograms VH LC , VH G correspond to specific areas of interest identified in the sequence of images generated by the camera 2, and the complementary volume histograms VH LC ', VHQ' correspond to complementary areas in the sequence of images.
  • the volume histograms VH LC , VH G , VH LC ', VHQ', corresponding to certain postures made by the user 1 while performing the gesture are processed to obtain position-characteristic features F p , such as the distances between the volume histograms VH LC , VH G , or the distances between a volume histogram VH LC , VH G and its complementary volume histogram VH LC ', VH G '.
  • the volume histograms VH LC , VH G , VH LC ', VH G ' and the position- characteristic features F p are forwarded to a classification unit 7, where they analysed to classify the gesture made by the user.
  • prototype position or posture information 13 including prototype posture histograms PHi, PH 2 , PH 3 , PH 4 , is retrieved from a database 9 of candidate gestures.
  • a volume histogram VH LC , VH G , VH LC ', VH G ' can be compared in some suitable manner to a corresponding prototype histogram PHi, PH 2 , PH 3 , PH 4 to determine a number of candidate gestures.
  • the second combining unit 8 By comparing successive volume histograms, obtained over time as the user moves through the postures of the gesture he is making, the number of candidate gestures from the database can be narrowed down, until the most likely candidate gesture is identified.
  • the result of the classification can be forwarded as a suitable signal 10 to a further processing block 11, for example a gesture interpretation module.
  • a further processing block 11 for example a gesture interpretation module.
  • the second combining unit 8 is shown to be part of the system 4 for performing motion classification, this second combining unit 8 could also conceivably be integrated in a system 3 for determining motion-related features.
  • Fig. 6 shows a block diagram of the main steps in the method of motion classification according to the invention, for an embodiment based on the analysis of blur-related first-order features.
  • images 600 are obtained from an image source such as a camera.
  • Edge-detection is performed on the images 600 in an edge-detection block 501, for example by performing wavelet transform on the image data, to give blur-related information 601, such as a set of wavelet coefficients 601.
  • blur-related information 601 such as a set of wavelet coefficients 601.
  • a feature extraction block 502 to give a set of first-order features 602 for each set of wavelet coefficients 601.
  • a histogram compilation block 503 statistical values for the first-order features 602 are computed, and a number of histograms or volume histograms 603 for the first-order features for an image, a specific area of an image, a sequence of images, or a sequence of specific areas of an image is compiled.
  • the histograms 603 are input to a histogram analysis block 504, in which motion-related features or position-characteristic features 604 are derived from the histograms 603.
  • These features 604 in turn are input to a motion classification block 505, where they are used to determine the motion to which the corresponding position belongs.
  • an output signal 605 can indicate the identified motion, e.g. a gesture made by a user, or whether the motion classification has failed.
  • the present invention has been disclosed in the form of preferred embodiments and variations thereon, it will be understood that numerous additional modifications and variations could be made thereto without departing from the scope of the invention.
  • the invention can be used in any system in which a simple camera and a computer with sufficient computational resources are available.
  • the method and system for gesture recognition according to the invention could be used to recognise sign-language gestures and to convert these to text.
  • Another implementation might be a surveillance system to classify and recognise different types of moving objects such as people, cars, etc.
  • Another possible implementation might be in automated production processes, where the system and method according to the invention could be used to track items being moved and to control different stages of the production process.
  • a “unit” or “module” can comprises a number of units or modules, unless otherwise stated.

Abstract

The invention relates to a method of determining motion-related features (Fm) pertaining to the motion of an object (1), which method comprises obtaining a sequence of images (f1, f2,..., fn) of the object (1), processing at least parts of the images (f1, f2,..., fn) to extract a number of first-order features from the images (f1, f2,..., fn), computing a number of statistical values pertaining to the first-order features, combining the statistical values into a number of histograms (HLC, HG), and determining the motion- related features (Fm) for the object (1) based on the histograms (HLC, HG). Furthermore, the invention relates to a method for performing motion classification for the motion of an object (1) captured in a sequence of images (f1, f2,..., fn). Such a motion classification method comprises determining motion-related features (Fm) for the images (f1, f2,..., fn) using the described method of determining motion-related features (Fm), and using the motion-related features (Fm) to classify the motion of the object (1). The invention also relates to a system (3) for determining motion-related features (Fm) pertaining to an object (1) in an image (f1), and to a system (4) for performing motion classification for the motion of an object ( 1 ).

Description

METHOD OF DETERMINING MOTION-RELATED FEATURES AND METHOD OF PERFORMING MOTION CLASSIFICATION
The invention relates to a method of determining motion-related features pertaining to the motion of an object, and to a method of performing motion classification for the motion of an object. Furthermore, the invention relates to a system for determining motion-related features pertaining to an object, and to a system for performing motion classification.
One of the most natural communication means between humans is that of making movements with the hands or head, also called gestures, as a supplement to spoken communication, or as a means of communication in itself. A simple augmentation to speech can comprise, for example, moving one's hand and arm to point in a particular direction when giving another person directions. An example of communication solely by motion is sign language, based entirely on gestures. Since gesturing is second nature to humans, it follows that gesturing would be a natural and easy means of communication between humans and computers, for example in a home dialog system, in a dialog system in an office environment, for automatic sign language interpretation, etc. Owing to developments in technology in recent years, it is conceivable that the use of various dialog systems will become more and more widespread in the future, and that classification of a user's movements, i.e. gesture recognition, will play an increasingly important role.
Several gesture recognition systems have been proposed by state of the art methods. Generally, these proposed systems rely on image analysis to locate a user in each of a sequence of images and to analyse successive postures to determine the gesture he is making. Some systems require that the user wear a coloured suit or position sensors, a requirement which would be most undesirable for the average user of a home dialog system, for example. Other systems such as US 6,256,033 Bl propose extracting from an image a number of coordinates corresponding to certain points on the body of a user, and comparing these to coordinates corresponding to known gestures. An obvious disadvantage of such an approach is the very high level of computational effort required to analyse the images, and that the performance of the camera used to obtain the images must be sufficient to ensure sharp images. Other conventional systems attempt to solve the problem of gesture recognition by working with stereo image data obtained by using two or more cameras. For example, WO 03/009218 proposes the use of two cameras to obtain a stereo image of a user making a gesture, and proceeds by removing the background in the images and tracking the upper body of the user using statistical framework for upper body segmentation of those parts of the images corresponding to the body of the user. Again, the computational level required for gesture recognition using this proposed system is very high, so that gesture recognition will be slow. A further disadvantage of such systems is the amount of hardware involved, namely a high- resolution stereo camera or multiple cameras. A disadvantage common to all of the conventional gesture recognition systems is that a user must perform a gesture slowly enough to ensure that a certain level of image quality is maintained, i.e. that the images are not blurred. The solutions suggested by the state of the art are therefore too expensive, too slow, or too unwieldy for gesture recognition, particularly for systems that need to be non-intrusive, user-friendly, and inexpensive. Often, the effort required by these systems simply to determine information about movement of a user, e.g. to determine motion-related features suitable for use in a motion classification process, is just too costly.
It is therefore an object of the invention to provide easy and economical methods of determining motion-related features pertaining to an object in an image and for classifying the motion of an object. Moreover, it is an object of the invention to provide appropriate systems for these tasks. To this end, the present invention provides a method of determining motion-related features pertaining to the motion of an object, which method comprises obtaining a sequence of images of the object and processing at least parts of the images to extract a number of first-order features from the images. Thereby, the term "sequence of images" can include all the images captured - for example by a camera - over a certain period of time, or only a selection of these images. For example, for a camera with a fast shutter speed, it might suffice to take every tenth image, whereas, for a low-cost camera with slow shutter speed, it may be preferable to use all the images. An image of an object for which motion-related features are to be determined can be processed in its entirety, or it may be that only a certain region of the image is processed, for instance, when only a part of the image is relevant. The first- order features that are extracted from the image or image part can be blur-related features, gradient values, or any other real value that can be extracted from the image, such as, for example, a value describing the colour of a pixel. The term "first-order" feature means a feature that can essentially be extracted from the image information in a first stage of computation. For example, pixel colour values can be directly extracted from the image information, and gradient values can be obtained by directly processing the image data. According to the invention, a number of statistical values pertaining to the first-order features is computed in a second stage of computation, and these statistical values are then combined to give one or more histograms. Based on these histograms, motion-related features are determined for the object. These second-order motion- related features can be any information given by or extracted from the histograms, identifying the areas in the image that are of relevance with regard to the motion of the object, for example, blurred areas in an image caused by a moving object. In particular, the histograms themselves can be regarded as kind of a second-order motion-related feature._The term "object" as used here is to be interpreted to mean a user, the body of a user, or part of the user's body, but could also be any moving object that is capable of moving in a pre-defined manner, such as a pendulum, a robot or other type of machine, etc. The terms "object" and "user" will be used interchangeably in the following. The present invention also provides a method of performing motion classification for the motion of an object captured in a sequence of images, which method comprises determining motion-related features for the object captured in the images in the manner described above, and using the motion-related features to classify the motion of the object.
An obvious advantage of the method according to the invention is that motion or gesture classification can be achieved without resorting to the complex methods of image analysis required by the state of the art gesture recognition systems, and the algorithms used in the method according to the invention to extract the first- order features from an image are easily available and are well-known to a person skilled in the art, such as established algorithms used in image compression. Since statistical information pertaining to the first-order features is simply tallied or counted in histograms, which can then easily and very quickly be analysed to obtain the motion- related features, the proposed method offers clear advantages over image analysis methods used in the state of the art. The method according to the invention does not require identification or localisation of the object, e.g. a person, in the image, and no modelling of body parts is required. Another advantage is that a simple, low-resolution camera such as a webcam is quite sufficient for generating the images, allowing the system according to the invention to be realised cheaply and effectively. Furthermore, since the actual motion of an object does not need to be tracked directly, a low frame rate for the camera is sufficient, so that a cheap camera is quite sufficient.
A corresponding system for determining motion-related features pertaining to an object comprises an image source for providing a sequence of images of the object, a processing unit for processing at least parts of the images to extract a number of first- order features from the image, a computation unit for computing a number of statistical values pertaining to the first-order features, a combining unit for combining the statistical values to give a number of histograms and for determining the motion-related features for the object, based on these histograms.
In addition, a corresponding system for performing motion classification comprises such a system for determining motion-related features pertaining to the object, and also a classification unit for classifying the motion of the object using the motion- related features.
The dependent claims and the subsequent description disclose particularly advantageous embodiments and features of the invention. A digital image generated by an image source such as a camera contains various types of characterizing information, such as brightness, colour or grey-scale values, contour or edge sharpness, etc., which, as mentioned above, can be used as first- order features for the method according to the invention. For the purposes of classifying motion of an object captured in an image, colour or grey-scale values might be of interest, for example to track the colour of the object as it appears to move across the image. However, information describing the level of blur in an image or a part of an image can be particularly informative. If a region of an image is blurred, for example, it might be that this region corresponds to, say, the hand or arm of a user making a gesture or other movement in front of the camera. Therefore, in a particularly preferred embodiment of the invention, the first-order features extracted from the image are blur- related features. Information given by contours or edges in an image is particularly useful for image processing techniques such as object or pattern recognition. Some edges in the image might be sharp (in focus) or blurred (out of focus). For a camera with sufficient depth of field, a blurred part in an image region will indicate that an object in that part of the image was moving while the image was being generated. Several techniques are available for identifying the edges in an image, and the technique of wavelet transformation has been developed in recent years for such applications as efficient image compression. The technique will be known to a person skilled in the art, and need not be explained in detail here. It suffices to say that the technique of wavelet transformation can be used to quickly and reliably identify the edges in an image, and to obtain information as to the level of sharpness of these edges.
Therefore, according to a preferred embodiment of the invention, extraction of blur-related features from an image is preferably carried out by performing a wavelet transform on image data, for a number of scales, to determine a series of wavelet coefficients for each point in the image. The wavelet coefficients, because their values depend on the discrepancies in value of neighbouring pixels, and the level of discrepancy depends in turn on the level of blur in the image, are blur-related features which ultimately provide motion information about the image. Such a method, using blur-related features as first-order features, is particularly advantageous since the very fact that a moving hand, head etc. will introduce blur into an image is used to positive effect. Blur is no longer an undesirable side-effect, but provides valuable information upon which the motion classification, e.g. gesture recognition, is based.
The type of wavelet transform to be used might depend to some extent on the size of the image. For example, the image data of a larger image with sufficient resolution could be used as input to a dyadic wavelet transform (DWT), whereas a continuous wavelet transform (CWT) could be used for images of small size, having only a relatively low resolution, such as those generated by a typical webcam. The wavelet coefficients thus obtained might then be processed to obtain blur-related features for the image. Each pixel or point of the image corresponds to a wavelet coefficient in each scale of the transform, and each scale has the same number of points as the image. Therefore, after carrying out a 10-scale transform, there will be ten coefficients for each point in the image. For a point or pixel on an edge in the image, the modulus of the coefficient associated with the pixel will be a maximum in each scale. The evolution of the ten coefficients across scales associated with the edge pixel provides information about the derivability of the image at that point. Using the wavelet coefficients for a pixel in an image, the image gradient can be computed for that pixel. The direction and the intensity, or magnitude, of the image gradient provide information about the direction of blur and the velocity of an object if the considered pixel belongs to a blurred edge of the moving object. Since the gradient direction is orthogonal to the edge or contour, it provides information about the shape of the object. Similarly, low-order gradient values derived from the pixels in a row or column imply diffuse or smooth edges, and high- order gradient values imply sharp or distinct edges. Therefore, histograms of image gradient intensity and direction provide information about shapes and motion field in the considered areas of the images.
The log of slope of the coefficients' evolution across scales provides an estimate of the so-called Lipschitz coefficient, or exponent, for that pixel. The Lipschitz exponent is a measure of the derivability of the image at the considered pixel and therefore a measure of the degree of continuity. If this Lipschitz exponent is positive, the pixel is located on a smooth or blurred edge. Equally, if the Lipschitz exponent is zero or negative, the pixel is located on a sharp edge. Performing the wavelet transform over a number of scales effectively provides a third dimension in addition to the two dimensions of the image, and it is in this third dimension that the evolution of the coefficients is observed. A wavelet transform is generally performed for a series of pixels, such as the pixels from a certain row or certain column in a set of image data. Therefore, Lipschitz exponents and the image gradient intensity and direction computed from the resulting wavelet coefficients provide information about the degree of blur for neighbouring pixels in a row or column and the direction of the blur, if applicable. For example, a set of mostly negative Lipschitz exponents for a row of pixels would indicate that this row of pixels comprises well-defined edges. On the other hand, mostly positive- valued Lipschitz exponents would indicate that there are smooth transitions between the pixels in this row, implying a lack of sharp edges and therefore possibly that this row belongs to a blurred region in the image.
The wavelet transform might be performed across successive rows of pixels and successive columns of pixels, so that the wavelet transform is performed numerous times for each image or part of the image. First-order features such as the Lipschitz exponents and wavelet coefficient gradients described above are computed for each wavelet transform operation, i.e. for each processed row and column.
An effective way of collating the information provided by the first-order features might be to round up the values for the first-order features to certain discrete values, and to simply count the number of occurrences of each value. Therefore, according to the invention, statistical values, such as the number of occurrences of a particular coefficient, pertaining to a certain kind of first-order feature extracted from at least part of an image are combined in a histogram for this first-order feature. To this end, in a system for obtaining motion-related features pertaining to the motion of an object in an image, a counter may be assigned to each discrete coefficient value, and, whenever this value occurs, the counter is incremented by one. The values thus accumulated are tallied or collected in the histogram, which can be visualised as a simple bar chart. To determine the movement or gesture being made by a user, a number of images (also called frames) might be captured, for example by a camera, following the movement of the user from start to finish. Such a motion or movement is generally referred to as a gesture, and the various stages or positions in the movement are generally referred to as postures, without limiting the scope of the invention in any way. An overall gesture is then given by a sequence of postures, and each posture in turn is captured in a sequence of images. For a gesture or movement in which the user moves his right hand and arm, the images will show the user and a blurred region to the left of the image, which blurred region effectively "moves" across the images over time. If a sequence of images or frames are imagined to be upright and stacked one in front of the other, the resulting three-dimensional block is a virtual volume given by the sequence of images over time. First-order features can be extracted for all of the images in such a volume, and combined to give motion-related information about the entire sequence. In a preferred embodiment of the invention, therefore, first-order features are extracted for the virtual volume, for example by performing the wavelet transform for each of the images in the sequence, and combining statistical values for these features in one or more volume histograms.
However, as mentioned above, only a part of the image may actually be of interest. Therefore, to reduce unnecessary computation, preferably only those areas in the images which are actually of interest might be processed. For example, if the user is making a gesture with only one hand and arm, as described above, and the region in the image comprising this hand and arm is only a part of the overall image, this region or specific area can be identified in one image, and only the corresponding specific areas of successive images in the sequence of images need be processed. This takes advantage of the fact that a user will generally stay more or less in the same place while making a gesture in front of a camera. If the images or frames are imagined to be upright, stacked one behind the other, and the specific areas of each of the successive frames are imagined to be "connected", the result is a virtual sub-volume. In a particularly preferred embodiment of the invention, therefore, a sequence of specific areas is identified in a sequence of images, and first-order features are extracted for the sub-volume given by the sequence of specific areas over time. Identification of the specific area of interest in an image is easily achieved by means of the first-order features for that image, for example by analysing the number of occurrences of a first-order feature for various regions of one of a sequence of images. For example, an entire image can be partitioned or sub-divided into a number of segments or tiles, preferably in such a way that the segment or tiles overlap. A sub-image histogram for a first-order feature can be compiled for each segment or tile. For example, a sub-image histogram of mostly zero values or positive values for Lipschitz coefficients indicates that this region of the image contains blur. Therefore, all regions of the image containing blur can be quickly identified. A selection can be formed comprising these regions which can be visualised by an imaginary rectangle drawn around a blurred moving arm in an image, and defined, for example, by the image coordinates for this rectangle. These image coordinates can then be used to locate the same specific area in all the images of a sequence. Equally, each tile or segment can be analysed by pixel colour value, for instance for a moving object with a colour different from the background colour in the images. Only tiles containing pixels of a certain colour might be regarded as being of interest, so that the remaining tiles might be disregarded.
The step of locating the specific area of interested can be carried out at regular intervals, so that the specific area tracks the moving object in the images. In this way, even motion or gestures involving movement of the user over the entire image can be analysed. Any motion or gesture can be broken down into a series of essentially distinct consecutive stages, which, when taken together, give the overall motion or gesture. In the following, therefore, the term "position" refers to any stage in the motion of an object or user that is captured in a sequence of images. Histograms of statistical values for one or more first-order features can be compiled for each specific area of a sequence of images. When considered together, therefore, the information in these histograms is characteristic of that sequence of specific areas. And, since a sequence of images can be used to determine a position being held or a posture being made, the information of the histograms of a sequence of specific areas is therefore characteristic of the position being held or the posture being made.
Therefore, in a preferred embodiment of the invention, the statistical values pertaining to a type of first-order feature extracted from the sequence of specific areas in a sequence of images are combined in a volume histogram for that feature. Such a volume histogram, describing a position of an object, a posture or part of a gesture, will be used in the motion classification process described in more detail below. While the regions of the image defined by the specific area contain useful information about the type of motion or gesture being made, the remaining regions of the image - the "complementary area" - can also be of use. The complementary area essentially comprises the pixels not included in the specific area. Therefore, in a preferred embodiment of the invention, when a specific area in an image is identified, its complementary area is also identified, and, for a sequence of complementary areas, statistical values pertaining to a certain kind of first-order feature extracted from the sequence of complementary areas are combined in a complementary volume histogram for this first-order feature.
The specific area and/or the complementary area of an image are preferably identified by means of the first-order features determined for that image as explained above. For example, an appropriate boosting algorithm can be used to analyse the image data in order to identify the area or areas of interest. Various boosting algorithms are available and will be known to a person skilled in the art.
The motion-related features obtained from a sequence of images can be used to classify the motion made by an object and captured in the sequence of images. For example, a motion-related feature might be a "distance" between two histograms. Such a distance might be computed between histograms of a specific area and the corresponding complementary area in an image, or it might be calculated between histograms of successive images in the image sequence. A distance between histograms can be computed using a standard statistical distribution technique or by comparing the main properties of the histograms, such as the main gradient direction or main Lipschitz coefficient in the case of blur-related first-order features, or any other simple property of the histogram. These distance are ultimately an indication of the degree of change , with regard to the motion-related features, of the image regions. As described above, a sequence of images can be seen to correspond to a particular position or stationary part of the trajectory in the motion of an object, or to a particular posture of a gesture being made by a user, so that, according to the invention, sub-volume histograms and complementary volume histograms (simply referred to as "volume histograms" collectively in the following), generated as described above using the first-order features of the image, can ultimately describe or characterise the positions of an object in motion. Therefore, in a preferred method according to the invention, such volume histograms are analysed to obtain position-characteristic features which can be used to classify the motion. For example, position-characteristic features such as a ratio of positive histogram values to negative histogram values in a volume histogram for a sequence of images can give an indication of the position and direction of motion of the object. Another type of position-characteristic information might be a distance between two volume histograms for different image sequences.
One way of using a volume histogram might be to compare it, or a derivative of the volume histogram, to previously generated data for various different motion sequences or for a variety of gestures. For example, a volume histogram obtained for a posture tracked by a sequence of images can be compared to a number of
"prototype postures" for a collection of gesture models, such as a collection of state- transition models for a number of gestures. In this way, one or more candidate gestures can be determined which comprise the posture corresponding to the volume histogram. By identifying successive prototypes for the volume histograms of a sequence of postures, the number of candidate gestures can be narrowed down until the gesture most likely corresponding to the sequence of postures has been identified.
Therefore, in a preferred embodiment of the invention, the position- characteristic features for a motion, obtained from the analysis of a sequence of images, are used to identify a state in a generative model which is based on position or posture sub-units for a motion or gesture, e.g. a state-transition model for a motion. Such a state corresponds to a particular position in that motion, for example, to a certain posture of a user or to a certain stage in the motion of an object. A preferred state-transition model well known to those skilled in the art is the Hidden Markov Model (HMM), and is used to determine the probability that, in a gesture or movement, a certain position is followed by another position. However, simpler features can be derived from the histograms. For instance, as described above, a distance between a sub-volume histogram and its complementary volume histogram could be calculated. Essentially, any appropriate method of statistical distribution comparison for obtaining a single real number from a series of volume histograms can be applied. These single values can be used as input to a suitable algorithm, for example a boosting algorithm, in order to build a discriminative criterion.
Therefore, in another particularly preferred embodiment of the invention, a distance between a volume histogram, generated as described above using a sequence of specific areas in a sequence of images, and a complementary volume histogram, generated as described above using the corresponding sequence of complementary areas in a sequence of images, is calculated to give a position-characteristic feature, which feature can then be used in the motion classification procedure.
Since the position-characteristic features are derived from the histograms (second-order motion-related features), they can also be regarded as a kind of third- order motion-related feature. Using these simple third-order features, a motion model corresponding to the positions in a motion can be identified, provided the prototype positions are characterised by this type of feature.
The steps of the methods for obtaining motion-related features and for performing motion classification as described above can be realised in the form of software modules to be run on a processor of a corresponding system for obtaining motion-related features or a system for performing motion classification, respectively. Some of the functions, for example certain image processing steps such as wavelet transforms in the case of extracting blur-related features, might be realised in the form of hardware, for example, as a dedicated application specific integrated circuit (ASIC) or field-programmable gate array (FPGA).
Other objects and features of the present invention will become apparent from the following detailed descriptions considered in conjunction with the accompanying drawing. It is to be understood, however, that the drawings are designed solely for the purposes of illustration and not as a definition of the limits of the invention. Fig. 1 is a schematic representation of a camera capturing an image of a user making gesture; Fig. 2a is a schematic representation of a sequence of images of a person with associated motion information and blur-related features;
Fig. 2b is a schematic representation of an image sub-divided into sub-images according to the invention;
Fig. 2c is a schematic representation of a specific area in an image according to the invention;
Fig. 3 a is a schematic representation of a virtual volume given by a sequence of specific areas from a sequence of images, and its associated volume histograms;
Fig. 3b is a representation of frontal view of a virtual volume, showing superimposed specific areas of a sequence of images; Fig. 4 shows a state diagram of a state-transition model for a gesture, and a number of volume histograms;
Fig. 5 is a block diagram of a system for gesture recognition according to the invention;
Fig. 6 shows a flow chart for the steps of the method of performing gesture recognition according to the invention.
In the diagrams, like numbers refer to like objects throughout. Fig. 1 shows a user 1 making a gesture in front of a camera 2, such as a webcam 2. Here, the user 1 is moving his arm in the direction of motion M, as a gesture or as part of a gesture. The webcam generates an image fϊ of the user 1, and this is forwarded to a suitable processing unit, not shown in the diagram, such as a personal computer. Any suitable type of camera can be used, but even a low-cost webcam 2 with a typical resolution of 320 x 480 pixels is sufficient for the purpose. The user 1 can make gestures associated with commands commonly occurring in a dialog between the user 1 and a dialog system, e.g. "stop interaction", "wait", "go back", "continue", "help", etc.. This obviates the need for speech interaction, for example in noisy environments, or when the user 1 is communicating by means of sign language. The user can also provide additional gesture information to supplement a spoken command, e.g. by pointing in a particular direction and verbally instructing the camera to look in that direction, e.g. by saying "watch this way".
The webcam 2 can generate images of the user 1 at more or less regular intervals, giving a sequence of images fi, f2, ..., fn as shown in Fig.2a. From this sequence of images fi, f2, ..., fn, the first-order features can first be extracted. The first- order features in the following examples used to illustrate the proposed method are blur- related features, chosen because of the advantages described in detail above. Without limiting the invention in any way, it is assumed that a method of gesture classification is being described.
As illustrated in Fig. 2a, various snapshots of the user 1 have been captured in the sequence of images fi, f2, ..., fn. These snapshots of the user, taken together, combine to form a certain position or posture in an overall gesture. Edge- detection is then performed on an image fi, by carrying out a wavelet transform on the rows and columns of pixels in the image fi. Depending on the size of the image fi and the resolution of the camera 2, it might suffice to use every nth pixel, or it may be necessary to use each pixel. A set of wavelet coefficients is obtained in this way for each pixel, containing information about the discrepancies between neighbouring pixels, i.e. motion information given by the level of blur.
For each set of wavelet coefficients, discrete blur-related features are derived, in this case the Lipschitz exponents and the image gradient. The occurrences of each value of first-order feature are counted, for example the number of times a Lipschitz coefficient of 1.5 or an image gradient value of 0.5π occurs, and collected in a histogram HLC, HG for that first-order feature. The resulting histogram HLC, HG therefore contains blur-related information about the overall image fi, for example the proportion of the image fi that is blurred, or the level of sharpness of the edges in the image fi.
This process can be repeated for each image of the remaining images f2, ..., fn in the sequence. However, performing edge analysis over the entire image is wasteful of resources, and can easily be circumvented. For instance, a type of "sliding window" could be used to virtually travel over the image data, giving smaller sub- windows for which the wavelet transform is carried out. The dimensions of the sliding window could also vary over time. For ease of clarification, Fig. 2b shows another technique of reducing computational effort, easier to present visually. Here, an image fϊ is virtually divided into smaller sections, tiles, or sub-images. Edge detection can be performed for each of the rows and columns in each of the sub-images of the overall image fϊ. One such sub-image 20 is shown, for which first-order feature histograms sHLc, SHQ have been derived. Since the process is carried out for each sub-image in the image fϊ, such histograms are derived for each of the sub-images. These can then be analysed to decide which of the sub-images are actually of interest. Since the interesting elements of a user making a gesture are his moving limbs such as arm, hand, etc., and since moving elements will be accompanied by a certain level of blur, the "interesting" parts of the image fϊ can easily be located by examining the first-order feature histograms to determine which of them is characteristic of a blurred sub-image. As mentioned above, these interesting parts of the image can be located easily by using the type of "sliding window" on the image data, effectively minimising the computational effort.
In Fig. 2c, two such sub-images 21, 22 have been located in the image ft. Naturally, the image fϊ could be sub-divided into many more sub-images than are shown here, and the sub-images could also overlap. For the sake of simplicity, however, only two sub-images 21, 22 of interest are shown. These sub-images 21, 22 combine to form a specific area Ai, indicated in the diagram by a thick rectangle.
A user positioned in front of a camera or webcam and making gestures will usually stand more or less in the same place, and only move one or both of his hands and arms. Therefore, the specific area identified in one image or frame can simply be propagated from one image in the sequence to the next in order to define a specific area in each of the following images or frames in a sequence. This saves on computational power and resources, since only the image data in the specific areas actually changes significantly over time. The rest of the image - the "complementary area" - remains to all intents and purposes the same, and can be regarded as stationary. This is illustrated in Fig. 3a, which shows a sequence of images fi, f2, ..., fn stacked vertically one behind the other. If imaginary lines are drawn to connect the corners of these frames fϊ, f2, ..., fn, a virtual volume V results. The specific area Ai shown in the first frame fi is propagated through to all the following frames f2, ..., fn, so that these frames f2, ... , fn have their specific areas A2, ... , An in the same relative positions. Again, if imaginary lines are drawn to connect the corners of these specific areas Ai, A2, ..., An, a virtual sub-volume Vs is created, excluding the complementary areas Ai', A2', ..., An' of the images fi, f2, ..., fn. As already described under Fig. 2a and 2b, first-order feature histograms are computed for each of the sub-images in the specific areas Ai, A2, ..., An of the frames fi, f2, ..., fn. By combining these histograms into corresponding volume histograms VHLC, VHG, the information contained in these histograms is extended to include time as well. In other words, the changes in the level of blur in the specific areas of the images are tracked over time as well. This is illustrated graphically in Fig. 3b, in which the image data of the sequence of frames fi, f2, ..., fn have been superimposed in the image 30, and all the "interesting" information made by the user while making a gesture is found in the region 31 , corresponding to the specific areas Ai , A2, ... , An in the sequence of images fi , f2, ... , fn.
In Fig. 4, a state-transition model G for a motion or gesture is shown, in this case a discrete Hidden Markov Model (HMM), in which a finite number of states Si, S2, S3, S4 correspond to a number of prototype positions of a motion, or postures of a gesture. This type of model is used to classify a motion, e.g. a gesture made by the user. Transitions are weighted by the probabilities of stepping from one state to another, as shown, for example, by the probabilities P(S3->S4), P(Si->S3), P(S4->S4), which give a measure of the probability of stepping or making the transition from state 3 to state 4, from state 1 to state 3, and from state 4 to state 4, respectively.
For example, a volume histogram of a particular posture made by the user and captured in a sequence of images can be compared to a collection of prototype postures associated with various gestures. The diagram shows four prototype posture histograms PHi, PH2, PH3, PH4 for the states Si, S2, S3, S4 of the state transition model G, but there may be any number of such state transition models available, each of which can be associated with various prototype postures. These prototype posture histograms PHi, PH2, PH3, PH4 can have been generated previously using a suitable learning algorithm. When comparing a volume histogram of a posture with the prototype postures, it may be found to be most similar to, for example, the prototype posture PHi associated with state Si of the state transition model. The comparison simply involves computing a distance, as already described above, between the volume histogram and the prototype histogram and converting this distance into a probability that is a number between 0 and 1. Subsequent volume histograms collected over time may be found to correspond to the consecutive states S2, S3, S4 so that the gesture being made by the user would in all likelihood correspond to the gesture modelled by this state-transition model G. In the event that a further volume histogram does not correspond to the next state S2, the classification procedure can conclude that this state transition model G does not correspond to a candidate gesture.
Fig. 5 shows a block diagram of a system 4 for performing gesture classification, without limiting the scope of the invention in any way, comprising a camera 2 for obtaining a sequence of images fϊ, f2, ..., fn of a user 1 (in the diagram, only a single image fϊ is indicated). Each image is first processed in a processing unit 5 to obtain a number of blur-related features 12. In this embodiment, the processing involves carrying out a wavelet transform on the image data to obtain Lipschitz exponents and wavelet coefficient gradients. These blur-related first-order features 12 are then forwarded to a computation unit 6, in which statistical values 14 are computed for the first-order features 12. In this computation unit 6, the values of the first-order features 12 are first rounded up or down , as appropriate, to the closest of a set of pre-defined discrete values, such as -0.5, 0.0, 0.5, 1.0, etc., before the occurrences of each discrete value are counted, or tallied, to give a number of counts 14. In a combining unit 15, the counts 14 are combined to give a number of histograms HLC, HG, one for each type of blur-related feature, in this case two (Lipschitz exponents and wavelet coefficient gradients), which can be regarded as one kind of motion-related features pertaining to the image. In this unit 15, the histograms HLC, HG can be used to derive further motion- related features Fm. For instance, the ratio of the total number of positive values to the total number of zero or negative values in a histogram HLC, HG can be supplied as a motion-related feature Fm for that histogram HLC, HG. Equally, a histogram HLC, HG itself can be used as a motion-related feature Fm. The combining unit 15 might comprise a separate analysis unit for performing any required histogram analysis, but is shown as a single entity in the illustration.
The units and blocks 2, 5, 6, 15 described up to this point comprise a system 3 for determining motion-related features, and which comprises the front end of the system 4 for performing gesture classification.
The histograms HLC, HG are forwarded to a second combining unit 8, where they are collected and combined with previous histograms derived from image data of previous images in the sequence of images, or virtual image 'volume', generated by the camera 2 of the user 1 performing a gesture. The output of the second combining unit 8 is then a set of volume histograms VHLC, VHG, VHLC', VHG', where the volume histograms VHLC, VHG correspond to specific areas of interest identified in the sequence of images generated by the camera 2, and the complementary volume histograms VHLC', VHQ' correspond to complementary areas in the sequence of images.
In the second combining unit 8, the volume histograms VHLC, VHG, VHLC', VHQ', corresponding to certain postures made by the user 1 while performing the gesture are processed to obtain position-characteristic features Fp, such as the distances between the volume histograms VHLC, VHG, or the distances between a volume histogram VHLC, VHG and its complementary volume histogram VHLC', VHG'.
The volume histograms VHLC, VHG, VHLC', VHG' and the position- characteristic features Fp are forwarded to a classification unit 7, where they analysed to classify the gesture made by the user. To this end, prototype position or posture information 13, including prototype posture histograms PHi, PH2, PH3, PH4, is retrieved from a database 9 of candidate gestures. A volume histogram VHLC, VHG, VHLC', VHG' can be compared in some suitable manner to a corresponding prototype histogram PHi, PH2, PH3, PH4 to determine a number of candidate gestures. By comparing successive volume histograms, obtained over time as the user moves through the postures of the gesture he is making, the number of candidate gestures from the database can be narrowed down, until the most likely candidate gesture is identified. The result of the classification can be forwarded as a suitable signal 10 to a further processing block 11, for example a gesture interpretation module. Although the second combining unit 8 is shown to be part of the system 4 for performing motion classification, this second combining unit 8 could also conceivably be integrated in a system 3 for determining motion-related features.
Fig. 6 shows a block diagram of the main steps in the method of motion classification according to the invention, for an embodiment based on the analysis of blur-related first-order features. In a first step 500, images 600 are obtained from an image source such as a camera. Edge-detection is performed on the images 600 in an edge-detection block 501, for example by performing wavelet transform on the image data, to give blur-related information 601, such as a set of wavelet coefficients 601. These are processed in a feature extraction block 502 to give a set of first-order features 602 for each set of wavelet coefficients 601. In a histogram compilation block 503, statistical values for the first-order features 602 are computed, and a number of histograms or volume histograms 603 for the first-order features for an image, a specific area of an image, a sequence of images, or a sequence of specific areas of an image is compiled. The histograms 603 are input to a histogram analysis block 504, in which motion-related features or position-characteristic features 604 are derived from the histograms 603. These features 604 in turn are input to a motion classification block 505, where they are used to determine the motion to which the corresponding position belongs. Once the classification procedure is completed, an output signal 605 can indicate the identified motion, e.g. a gesture made by a user, or whether the motion classification has failed.
Although the present invention has been disclosed in the form of preferred embodiments and variations thereon, it will be understood that numerous additional modifications and variations could be made thereto without departing from the scope of the invention. The invention can be used in any system in which a simple camera and a computer with sufficient computational resources are available. For example, the method and system for gesture recognition according to the invention could be used to recognise sign-language gestures and to convert these to text. Another implementation might be a surveillance system to classify and recognise different types of moving objects such as people, cars, etc. Another possible implementation might be in automated production processes, where the system and method according to the invention could be used to track items being moved and to control different stages of the production process.
For the sake of clarity, it is to be understood that the use of "a" or "an" throughout this application does not exclude a plurality, and "comprising" does not exclude other steps or elements. A "unit" or "module" can comprises a number of units or modules, unless otherwise stated.

Claims

CLAIMS:
1. A method of determining motion-related features (Fm) pertaining to the motion of an object (1), which method comprises obtaining a sequence of images (fi, f2, ..., fn) of the object (1); processing at least parts of the images (fi, f2, ..., fn) to extract a number of first-order features from the images (fi, f2, ..., fn); computing a number of statistical values pertaining to the first-order features; combining the statistical values to give a number of histograms (HLC, HG); and determining the motion-related features (Fm) for the object (1) based on the histograms (HLC, HG).
2. A method according to claim 1, wherein the first-order features comprise blur-related features.
3. A method according to claim 2, wherein extraction of blur-related features from an image (fi, f2, ..., fn) comprises performing a wavelet transform on image data to determine a number of wavelet coefficients, and processing the wavelet coefficients to give the blur-related features.
4. A method according to any of the preceding claims, wherein first-order features are extracted for a volume (V) given by the sequence of images (fi, f2, ..., fn) over time.
5. A method according to any of the preceding claims, wherein a sequence of specific areas (Ai, A2, ..., An) is identified in the sequence of images (fi, f2, ..., fn), and first-order features are extracted for a sub-volume (V8) given by the sequence of specific areas (Ai, A2, ..., An) over time.
6. A method according to claim 4 or claim 5, wherein statistical values pertaining to a certain kind of first-order feature, extracted from a sequence of images (fi, f2, ..., fn) or from a sequence of specific areas (Ai, A2, ..., An) in a sequence of images (fi, f2, ..., fn), are combined in a volume histogram (VHLC, VHG) for this first- order feature.
7. A method according to claim 6, wherein a sequence of complementary areas (Ai', A2', ..., An') is identified in a sequence of images (fi, f2, ..., fn), where a complementary area (Ai', A2', ..., An') in an image (fi, f2, ..., fn) is complementary to the specific area (Ai, A2, ..., An) in that image (fi, f2, ..., fn), and statistical values pertaining to a certain kind of first-order feature extracted from the sequence of complementary areas (Ai', A2', ..., An') are combined in a complementary volume histogram (VHLC', VHQ') for this first-order feature.
8. A method according to any of claims 5 to 7, wherein the specific area (Ai, A2, ... , An) and/or the complementary area (Ai', A2', ... , An') of an image (fi, f2, ... , fn) are identified by means of the first-order features determined for that image (fi, f2, ...,
Q.
9. A method of performing motion classification for the motion of an object (1) captured in a sequence of images (fi, f2, ..., fn), which method comprises determining motion-related features (Fm) for the object (1) in the images (fi, f2, ..., fn) using the method of any of claims 1 to 8, and using the motion-related features (Fm) to classify the motion of the object (1).
10. A method according to claim 9, wherein volume histograms (VHLC, VHG, VHLC', VHQ') are generated using the method according to claims 6 or 7 to obtain position-characteristic features (Fp), and these position-characteristic features (Fp) are analysed to classify the motion of the object (1).
11. A method according to claim 10, wherein a distance between a volume histogram (VHLC, VHG) and a complementary volume histogram (VHLC', VHG') is calculated to obtain position-characteristic features (Fp), and these position-characteristic features (Fp) are analysed to classify the motion of the object (1).
12. A method according to claims 10 or 11, wherein the position- characteristic features are used to identify a state (Si, S2, S3, S4) in a state-transition model for a motion, which state (Si, S2, S3, S4) corresponds to a particular position in that motion.
13. A system (3) for determining motion-related features (Fm) pertaining to an obj ect ( 1 ) , which system comprises an image source (2) for providing a sequence of images (fi, f2, ..., fn) of the object (1); a processing unit (5) for processing at least parts of the images (fi, f2, ... , fn) to extract a number of first-order features from the images (fi, f2, ..., fn); - a computation unit (6) for computing a number of statistical values pertaining to the first-order features; a combining unit (15) for combining the statistical values into a number of histograms (HLC, HG) and for determining the motion-related features (Fm) based on the histograms (HLC, HG).
14. A system (4) for performing motion classification for the motion of an object (1), comprising a system (3) according to claim 13 for determining motion-related features (Fm) pertaining to the object, and a classification unit (7) for classifying the motion of the object (1) using the motion-related features (Fm).
15. A computer program product, directly loadable into the memory of a programmable device for use in a system for determining motion-related features pertaining to an object and/or a system for performing motion classification for the motion of an object, comprising software code portions for performing the steps of a > method according to claims 1 to 12 when said computer program product is run on the device.
PCT/IB2008/051843 2007-05-15 2008-05-09 Method of determining motion-related features and method of performing motion classification WO2008139399A2 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP07108267 2007-05-15
EP07108267.1 2007-05-15

Publications (2)

Publication Number Publication Date
WO2008139399A2 true WO2008139399A2 (en) 2008-11-20
WO2008139399A3 WO2008139399A3 (en) 2009-04-30

Family

ID=40002714

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2008/051843 WO2008139399A2 (en) 2007-05-15 2008-05-09 Method of determining motion-related features and method of performing motion classification

Country Status (2)

Country Link
TW (1) TW200910221A (en)
WO (1) WO2008139399A2 (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2010086866A1 (en) * 2009-02-02 2010-08-05 Eyesight Mobile Technologies Ltd. System and method for object recognition and tracking in a video stream
EP2428870A1 (en) * 2010-09-13 2012-03-14 Samsung Electronics Co., Ltd. Device and method for controlling gesture for mobile device
JP2014174776A (en) * 2013-03-11 2014-09-22 Lenovo Singapore Pte Ltd Recognition method of action of moving object and portable type computer
US8890803B2 (en) 2010-09-13 2014-11-18 Samsung Electronics Co., Ltd. Gesture control system
JP2015231517A (en) * 2014-05-13 2015-12-24 オムロン株式会社 Posture estimation device, posture estimation system, posture estimation method, posture estimation program, and computer-readable recording medium recording posture estimation program
CN107735813A (en) * 2015-06-10 2018-02-23 柯尼卡美能达株式会社 Image processing system, image processing apparatus, image processing method and image processing program
EP3309747A4 (en) * 2015-06-11 2018-05-30 Konica Minolta, Inc. Motion detection system, motion detection device, motion detection method, and motion detection program
US10096118B2 (en) 2009-01-13 2018-10-09 Futurewei Technologies, Inc. Method and system for image processing to classify an object in an image

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
TWI415032B (en) * 2009-10-30 2013-11-11 Univ Nat Chiao Tung Object tracking method
US9165187B2 (en) * 2012-01-12 2015-10-20 Kofax, Inc. Systems and methods for mobile image capture and processing
TWI779454B (en) * 2021-01-08 2022-10-01 財團法人資訊工業策進會 Motion recognition apparatus and method thereof

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2007004100A1 (en) * 2005-06-30 2007-01-11 Philips Intellectual Property & Standards Gmbh A method of recognizing a motion pattern of an obejct

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2007004100A1 (en) * 2005-06-30 2007-01-11 Philips Intellectual Property & Standards Gmbh A method of recognizing a motion pattern of an obejct

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ROOMS F ET AL: "Estimating image blur in the wavelet domain" INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNALPROCESSING, XX, XX, 8 July 2002 (2002-07-08), pages 1-6, XP002384823 *

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10096118B2 (en) 2009-01-13 2018-10-09 Futurewei Technologies, Inc. Method and system for image processing to classify an object in an image
US20110291925A1 (en) * 2009-02-02 2011-12-01 Eyesight Mobile Technologies Ltd. System and method for object recognition and tracking in a video stream
KR20110138212A (en) * 2009-02-02 2011-12-26 아이사이트 모빌 테크놀로지 엘티디 System and method for object recognition and tracking in a video stream
CN102356398A (en) * 2009-02-02 2012-02-15 视力移动技术有限公司 System and method for object recognition and tracking in a video stream
JP2012517044A (en) * 2009-02-02 2012-07-26 アイサイト モバイル テクノロジーズ リミテッド System and method for object recognition and tracking in video streams
WO2010086866A1 (en) * 2009-02-02 2010-08-05 Eyesight Mobile Technologies Ltd. System and method for object recognition and tracking in a video stream
KR101632963B1 (en) 2009-02-02 2016-06-23 아이사이트 모빌 테크놀로지 엘티디 System and method for object recognition and tracking in a video stream
US9405970B2 (en) * 2009-02-02 2016-08-02 Eyesight Mobile Technologies Ltd. System and method for object recognition and tracking in a video stream
US20160343145A1 (en) * 2009-02-02 2016-11-24 Eyesight Mobile Technologies Ltd. System and method for object recognition and tracking in a video stream
EP2428870A1 (en) * 2010-09-13 2012-03-14 Samsung Electronics Co., Ltd. Device and method for controlling gesture for mobile device
US8890803B2 (en) 2010-09-13 2014-11-18 Samsung Electronics Co., Ltd. Gesture control system
JP2014174776A (en) * 2013-03-11 2014-09-22 Lenovo Singapore Pte Ltd Recognition method of action of moving object and portable type computer
JP2015231517A (en) * 2014-05-13 2015-12-24 オムロン株式会社 Posture estimation device, posture estimation system, posture estimation method, posture estimation program, and computer-readable recording medium recording posture estimation program
EP3309748A4 (en) * 2015-06-10 2018-06-06 Konica Minolta, Inc. Image processing system, image processing device, image processing method, and image processing program
CN107735813A (en) * 2015-06-10 2018-02-23 柯尼卡美能达株式会社 Image processing system, image processing apparatus, image processing method and image processing program
EP3309747A4 (en) * 2015-06-11 2018-05-30 Konica Minolta, Inc. Motion detection system, motion detection device, motion detection method, and motion detection program

Also Published As

Publication number Publication date
WO2008139399A3 (en) 2009-04-30
TW200910221A (en) 2009-03-01

Similar Documents

Publication Publication Date Title
WO2008139399A2 (en) Method of determining motion-related features and method of performing motion classification
Chen et al. Repetitive assembly action recognition based on object detection and pose estimation
Wu et al. Recent advances in video-based human action recognition using deep learning: A review
US11237637B2 (en) Gesture recognition systems
Hasan et al. RETRACTED ARTICLE: Static hand gesture recognition using neural networks
US6072494A (en) Method and apparatus for real-time gesture recognition
EP2344980B1 (en) Device, method and computer program for detecting a gesture in an image, and said device, method and computer program for controlling a device
US7308112B2 (en) Sign based human-machine interaction
CN108171133B (en) Dynamic gesture recognition method based on characteristic covariance matrix
KR100421740B1 (en) Object activity modeling method
CN110826389B (en) Gait recognition method based on attention 3D frequency convolution neural network
CN108416780B (en) Object detection and matching method based on twin-region-of-interest pooling model
CN107480585B (en) Target detection method based on DPM algorithm
CN110929593A (en) Real-time significance pedestrian detection method based on detail distinguishing and distinguishing
Moore A real-world system for human motion detection and tracking
CN110717154A (en) Method and device for processing characteristics of motion trail and computer storage medium
CN112668492A (en) Behavior identification method for self-supervised learning and skeletal information
KR20120089948A (en) Real-time gesture recognition using mhi shape information
EP2790130A1 (en) Method for object recognition
Yu et al. Human motion recognition based on neural network
Świtoński et al. Human identification based on the reduced kinematic data of the gait
KR100679645B1 (en) Gesture recognition method using a virtual trellis plane
CN112926522A (en) Behavior identification method based on skeleton attitude and space-time diagram convolutional network
Vo et al. Dynamic gesture classification for Vietnamese sign language recognition
Hoque et al. Computer vision based gesture recognition for desktop object manipulation

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 08751191

Country of ref document: EP

Kind code of ref document: A2

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 08751191

Country of ref document: EP

Kind code of ref document: A2