GB2572497A - Video sequence processing - Google Patents

Video sequence processing Download PDF

Info

Publication number
GB2572497A
GB2572497A GB1905665.4A GB201905665A GB2572497A GB 2572497 A GB2572497 A GB 2572497A GB 201905665 A GB201905665 A GB 201905665A GB 2572497 A GB2572497 A GB 2572497A
Authority
GB
United Kingdom
Prior art keywords
pixel
filter
regions
partial
partial filters
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
GB1905665.4A
Other versions
GB2572497B (en
GB201905665D0 (en
Inventor
James Knee Michael
Weston Martin
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Snell Advanced Media Ltd
Original Assignee
Snell Advanced Media Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Snell Advanced Media Ltd filed Critical Snell Advanced Media Ltd
Priority to GB1905665.4A priority Critical patent/GB2572497B/en
Priority claimed from GB1206065.3A external-priority patent/GB2502047B/en
Publication of GB201905665D0 publication Critical patent/GB201905665D0/en
Publication of GB2572497A publication Critical patent/GB2572497A/en
Application granted granted Critical
Publication of GB2572497B publication Critical patent/GB2572497B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/503Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
    • H04N19/51Motion estimation or motion compensation
    • H04N19/513Processing of motion vectors
    • H04N19/521Processing of motion vectors for estimating the reliability of the determined motion vectors or motion vector field, e.g. for smoothing the motion vector field or for correcting motion vectors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/223Analysis of motion using block-matching
    • G06T7/238Analysis of motion using block-matching using non-full search, e.g. three-step search
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • G06T7/248Analysis of motion using feature-based methods, e.g. the tracking of corners or segments involving reference images or patches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/254Analysis of motion involving subtraction of images
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/80Details of filtering operations specially adapted for video compression, e.g. for pixel interpolation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10016Video; Image sequence

Abstract

In an apparatus for motion vector assignment partial filters are applied to two or more regions 402-409 of pixel-to-pixel dissimilarity values between a plurality of video frames in a sequence. Outputs of the partial filters are combined to estimate motion between the plurality of video frames. The partial filters may take a maximum or a mean of the pixel difference values in the regions. The regions may define line segments or octants arranged radially around a pixel 401 for which the motion is to be determined. The outputs may be combined by a non-linear operation, for example taking the minimum from diametrically opposed regions. The apparatus may be used during motion searching and the filtering may assess how well-matched pixels in regions near a pixel are to regions near a point indicated by a candidate motion vector.

Description

This invention relates to video sequence processing particularly in connection with motion estimation of video signals.
BACKGROUND OF THE INVENTION
In the estimation of motion vectors between video frames, motion vectors are assigned to pixels, or blocks of pixels, in each frame and describe the estimated displacement of each pixel or block in a next frame or a previous frame in the sequence of frames. In the following description, the motion estimation is considered to be dense meaning that a motion vector is calculated for every pixel. The definition of dense may be widened to cover the calculation of a motion vector for each small block in the picture, for each pixel in a subsampled version of the picture, or for each small region of arbitrary shape within which the motion is expected to be uniform. The invention can be applied with trivial modification to these wider cases.
Motion estimation has application in many image and video processing tasks, including video compression, motion-compensated temporal interpolation for standards conversion or slow-motion synthesis, motion-compensated noise reduction, object tracking, image segmentation, and, in the form of displacement estimation, stereoscopic 3D analysis and view synthesis from multiple cameras.
Most applications of motion estimation involve the “projection” (also described as “shifting”) of picture information forward or backward in time according to the motion vector that has been estimated. This is known as motion-compensated projection. The projection may be to the time instant of an existing frame or field, for example in compression, where a motion-compensated projection of a past or future frame to the current frame instant serves as a prediction of the current frame. Alternatively, the projection may be to a time instant not in the input sequence, for example in motion-compensated standards conversion, where information from a current frame is projected to an output time instant, where it will be used to build a motion-compensated interpolated output frame.
Some of the terminology used in describing motion estimation systems will now be described. Figure 1 shows one-dimensional sections through two successive
-2frames in a sequence of video frames. The horizontal axis of Figure 1 represents time, and the vertical axis represents position. Of course, the skilled person will recognise that Figure 1 is a simplification and that motion vectors used in image processing are generally two dimensional. The illustrated frames are: a previous or reference frame (101); and, the current frame (102). A motion vector (104) is shown assigned to a pixel (103) in the current frame. The motion vector indicates a point (105) in the reference frame which is the estimated source, in the reference frame, of the current frame pixel (103). This example shows a backward vector. Forward vectors may also be measured, in which case the reference frame is the next frame in the sequence rather than the previous frame.
The following descriptions assume that these frames are consecutive in the sequence, but the described processes are equally applicable in cases where there are intervening frames, for example in some compression algorithms. Temporal samples of an image will henceforth be referred to as fields, as would be the case when processing interlaced images. However, as the skilled person will appreciate, in non-interlaced image formats a temporal sample is represented by a frame; and, fields may be ‘de-interlaced’ to form frames within an image process. The spatial sampling of the image is not relevant to the discussion which follows.
An example of an algorithm that calculates motion vectors is disclosed in GB2188510. This algorithm is summarised in Figure 2 and assigns a single vector to every pixel of a current field in a sequence of fields. The process of Figure 2 is assumed to operate sequentially on the pixels of the current field; the pixel whose vector assignment is currently being determined will be referred to as the current pixel. The current field (202) and the previous field (201) are applied to a phase correlation unit (203) which calculates a menu (204) for every pixel of the current field consisting of a number (three in this example) of candidate motion vectors. Each candidate vector controls a respective member of a set of shift units (205) which, for every pixel in the current field, displaces the previous field (201) by the respective candidate vector to produce a shifted pixel corresponding to the current pixel of the current field in the respective member of the set of displaced fields (206).
-3A set of error calculation units (207) produces a set of error values (208), one error value for every menu vector for every pixel of the current field. Each of the error calculation units (207) subtracts the respective one of the displaced fields (206) from the current field (202) and rectifies the result to produce a field of difference magnitudes, which are known as displaced field differences or “DFDs”. Each of the error calculation units (207) spatially filters its respective field of DFDs in a filter centred on the current pixel to give an error value for that pixel and menu vector. This spatially filtered DFD is the error value for the respective current pixel and vector. The set three error values (208) for the current pixel are compared in a comparison unit (209), which finds the minimum error value. The comparison unit (209) outputs a candidate index (210), which identifies the vector that gave rise to the minimum error value. The candidate index (210) is then applied to a vector selection unit (211) to select the identified candidate from the menu of vectors (204) as the respective output assigned vector (212) for the current pixel.
An important property of DFDs will now be described. If a candidate motion vector for a pixel describes the true motion of that pixel, then we would expect the DFD to be small, and only non-zero because of noise in the video sequence. If the candidate motion vector is incorrect, then the DFD may well be large, but it might be coincidentally small. For example, a rising waveform in one field may match a falling waveform in the displaced field at the point where they cross. Alternatively, a pixel may be in a plain area or in a one-dimensional edge, in which case several motion vectors would give rise to a small or even a zero DFD value. This inconvenient property of DFDs is sometimes referred to as the aperture problem and leads to the necessity of spatially filtering the DFDs in order to take information from nearby pixels into account in determining the error value for a pixel.
In the example of Figure 2, each error calculation block (207) filters the DFDs with a two-dimensional filter, a typical example of which is a 5 x 5 runningaverage filter. It is this rectified and filtered error that is used for comparison of candidate motion vectors. Figure 3 illustrates the positions of the 25 samples involved in the running-average filter. The 5x5 arrangement of 25 samples comprises the samples within the rectangular filter window (302) and is centred on the current pixel position (301).
-4Choosing the size of the two-dimensional DFD filter involves a trade-off between reliability and spatial accuracy of the resulting assigned motion vector field. If, on the one hand, the filter is large, then the effect of noise on the filtered error value is reduced and the filter is more likely to take into account nearby detail in the picture which might help to distinguish reliably between candidate motion vectors. However, a large filter is also more likely to take in pixel data from one or more objects whose motion is properly described by different motion vectors, in which case it will fail to give a low error value for any candidate motion vector, even for one that is correct for the pixel in question.
If, on the other hand, the filter is small, it is more likely to involve pixels from only one object and so is more likely to return a low error value for the correct motion vector. However, it will be less likely to reject wrong motion vectors and will be more susceptible to noise.
The inventors have observed that, for critical picture material, there is no choice of filter size which yields satisfactory performance in all aspects of reliability, noise immunity, spatial accuracy and sensitivity. However, the inventors have recognized that it is possible to design an improved displaced field difference filter which combines the reliability and noise immunity of a large conventional filter with the sensitivity and spatial accuracy of a small filter, while avoiding the disadvantages of each.
SUMMARY OF THE INVENTION
The invention consists of a method and apparatus for filtering displaced field differences arising from candidate motion vectors, characterised in that the filter window is decomposed into regions that are filtered separately and whose outputs are combined by a non-linear operation.
BRIEF DESCRIPTION OF THE DRAWINGS
An example of the invention will now be described with reference to the drawings in which:
Figure 1 is a diagram showing current and previous frames in an image sequence and a backward motion vector extending from a pixel in the current frame;
-5Figure 2 is a block diagram of apparatus for assigning backward motion vectors to pixels according to the prior art;
Figure 3 is a diagram of a filter window according to the prior art;
Figure 4 is a diagram of a set of filter windows according to a first embodiment of the invention;
Figure 5 is a block diagram of an improved filter according to a first embodiment of the invention.
Figure 6 is a diagram of a set of filter windows according to a second embodiment of the invention;
Figure 7 is a block diagram of an improved filter according to a second embodiment of the invention.
DETAILED DESCRIPTION OF THE INVENTION
As explained in the introduction, a displaced field difference filter operates on a set of DFDs representing difference values, between current field pixels and respective displaced field pixels for a particular motion vector. Typically the difference values are rectified prior to filtering so that the magnitudes of the errors are represented by the DFDs. The filter takes contributions from the DFDs for a number of pixels within a filter window surrounding a current pixel; the DFD for the current pixel may also be used. Contributions from these DFDs are used to form an error value for the current pixel.
The input DFD values being filtered arise from a candidate motion vector, or from a smoothly varying motion vector field, calculated by known methods. In the description that follows, the term motion vector refers either to a constant vector over a region or to a smoothly varying vector field.
Displaced field difference filters according to examples of the invention will now be described. In each case the filter output is an error value for a particular motion vector at a particular pixel position within a current field, this pixel position will be referred to as the current pixel. The filter input DFD values will be referred to as samples, and the DFD corresponding to the current pixel will be described
-6as the current sample. The positions of samples correspond with the positions of the respective current field pixels used to calculate the respective DFDs.
The filter window of a first exemplary embodiment of the invention is illustrated in Figure 4, to which reference is now directed. The filter is given access to a number of contributing samples surrounding the current sample (401). Only samples that are used by the filter are shown in Figure 4; other samples in the vicinity of the current sample are not shown, typically there will be intermediate, unused samples forming part of an orthogonal spatial sampling structure for the current field. The contributing samples are grouped into eight line segments (402 to 409) in a star pattern centred on the current sample (401). The choice of this pattern is a compromise between economy and ease of access to samples in a hardware implementation, and the need to cover a reasonably wide area surrounding the current sample. In this particular example, each line segment contains seven samples, though other sizes are possible without departing from the scope of the invention.
The object of the filter is to give a high output if the motion vector that gave rise to the contributing samples is the wrong motion vector for the position of the current sample (401), and to give a low output if the motion vector is correct. If we begin with the assumption that the validity or invalidity of a motion vector extends across the area covered by the star pattern, then a high sample value somewhere in the pattern constitutes evidence that the motion vector is incorrect, and a suitable nonlinear filtering operation would be to take the maximum of the sample values across the pattern. However, it is quite possible that a boundary between two differently moving objects, for example the line shown (410) will cross the area. In this case, if the motion vector that gave rise to the sample is the one describing the motion of the right-hand object, we would expect the samples to the right of the line to have low values and those to the left to have at least some high values. We observe that, if the eight line segments in the star pattern are grouped into pairs of diametrically opposite segments (402 with 403;
404 with 405; 406 with 407; and, 408 with 409) then one segment of each pair will be expected to contain low sample values. The operation of the first inventive filter is therefore to take maximum values in each line segment, and then to take the minimum of the two maxima within each pair. This operation produces four values, all of which we expect to be low if the motion vector is correct. A further
-7operation of the filter is therefore to take the maximum of the four minima. Finally, it is important for spatial accuracy to take account of the current sample. This is done by combining its value with the output of the filter so far defined, for example by taking the mean square value.
An alternative description of the first exemplary inventive filter will now be given with reference to the block diagram in Figure 5. The filter receives an input stream of samples (500) corresponding to the DFDs for a current field and a particular motion vector. The samples are ordered according to a scanning raster so that when they are passed through a chain of delay elements (510) suitable choices for the delay values give access to the 57 (in this example) samples at the locations shown in the star pattern of Figure 4. The output of the delay chain (510) takes the form of eight sets (502 to 509) of seven samples each, where output (502) corresponds to line segment (402), output (503) to line segment (403), and so on, together with the central sample (501), corresponding to current sample (401)). The maximum value of each of the eight sets is found in respective maximum-value calculation units (512) to (519). The resulting maximum values (522) to (529) are applied in pairs to minimum-value calculation units (532), (534), (536) and (538) so as to find the respective minimum values from diametrically-opposite filter window segments. The resulting minimum values (542), (544), (546) and (548) are applied to a maximum-value calculation unit (550) whose output (551) is combined (553) with the current sample (501) by taking the root-mean-square value, which form the filtered DFD output (554).
Possible variations of this filter will now be described. In a first variation, the eight maximum-value calculation units (512) to (519) are replaced by eight averaging units. This variation can improve the noise immunity of the filter. In a second variation, the subsequent maximum-value unit (550) is likewise replaced by an averaging unit.
It will be apparent to the skilled person that other choices of processing elements may also be used. For example, units (512) to (519) may calculate: a mean square value; a combination of the mean and the maximum; or, other rank-order values such as the second or third highest value. Similarly, unit (550) may also take: a mean square value; a combination of the mean and the maximum; or, the second highest value. Such decisions are a trade-off between robustness to
-8noise and sensitivity to data, and between reliability and the capability of handling motion vector boundaries that are more complex in shape.
A displaced field difference filter according to a second exemplary embodiment of the invention will now be described. The second filter is more reliable than those previously described, at the cost of an increase in complexity. Figure 6 shows the samples involved in the second filter, based on an example window size of 15x15. In place of the eight 7-sample line segments shown in Figure 4, this filter has eight octants (602) to (609) each containing 28 samples. (In Figure 6 the sample positions in alternate octants are indicated by open circles so as to indicate more clearly the allocation of samples to octants.) The average value of the samples within each octant is taken, and subsequent processing may be the same as that of the first filter.
Preferably however, the final combining step, (553) of Figure 5, may be replaced by a linear combination of the output of the four-value mean (550 in Figure 5) with the output of a conventional 5x5 running average filter whose window (610) is also shown in Figure 6.
The architecture of the second filter may be based on Figure 5, with the output of delay chain (510) now consisting of eight sets of 28 samples. However, a more efficient implementation is as shown in Figure 7, where the chain of delay elements and the mean-value calculations at its output are replaced by octantshaped running-average filters which may be constructed, for example, as described in UK patent application 1113569.6, with additional simplifications that exploit the fact that the octants have shared boundaries.
Referring to Figure 7, the input stream of samples (700) is applied to eight octantshaped running-average filters (712) to (719) whose outputs (722) to (729) are applied in pairs to minimum-value calculation units (732), (734), (736) and (738) so as to find the respective minimum values from diametrically-opposite filter window segments. The resulting minimum values (742), (744), (746) and (748) are applied to an averaging unit (750) whose output (751) is linearly combined (753) with the output (752) of a 5 x 5 running-average filter (702) applied to a suitably delayed version (701) of the input (700), to produce a final filtered DFD output (754). A typical linear combination in block (753) is to add 75% of the
-9output (751) of the averaging unit (750) to 25% of the output (752) of the 5 x 5 running-average filter (702).
The invention so far described involves filter windows of particular sizes and shapes. It will be apparent to the skilled person that other sizes and shapes may be chosen without departing from the scope of the invention. For example, the line segments of the star pattern in Figure 4 may contain fewer or more than the seven samples shown. The pattern may also have fewer or more than the eight line segments shown. Likewise, the square window shown in Figure 6 may be smaller or larger than the 15 x 15 window shown, and the eight octants may be replaced by suitable numbers of other shapes, for example four quadrants or sixteen sedecants. The window need not be square: for example, windows that are polygonal with other than four sides, or that are approximately circular, may also be used. It is also possible to combine error value samples from overlapping segments of the filter window without departing from the scope of the invention.
The above description is based on displaced field differences. Other measures of pixel-to-pixel dissimilarity may also be used, including but not limited to: nonlinear functions of displaced field difference, displaced field differences between noisereduced fields, Euclidean or other distances between multidimensional signals, for example RGB signals, and differences between feature point descriptors.
The implementations of the filters have been described in terms of serial processing of streams of values, typically ordered according to a scanning raster. Of course the skilled person will appreciate that many other implementations of the inventive filters are possible, including, for example, the use of randomaccess field or frame stores or programmable apparatus. And, as explained in the introduction, filtering according to the invention may be applied to measures of dissimilarity between subsamples or regions of an image.
Although motion-compensated processing of images is typically applied to a time sequence of images where the sequence of images is a time sequence, the same process may be used with spatial image sequences, where the sequence is a sequence of different views of a common scene, or a sequence of different views captured in a time sequence. The current invention is equally applicable to the processing of these other types of image sequence.

Claims (20)

1. In video sequence processing, a method of filtering motion-compensated pixel-to-pixel dissimilarity values in which the filter aperture is decomposed into two or more regions and the outputs of partial filters applied to each region are combined by a non-linear operation.
2. A method according to claim 1 in which the dissimilarity value is a rectified displaced field difference.
3. A method according to claim 1 in which the regions are non-overlapping.
4. A method according to claim 1 in which the non-linear combination process includes taking minimum values of partial-filter outputs from pairs of regions that are diametrically opposite each other in the filter aperture.
5. A method according to claim 4 in which the partial filters operate on radial line segments.
6. A method according to claim 5 in which the number of radial line segments is eight.
7. A method according to claim 4 in which the partial filters operate on sectors of the filter aperture.
8. A method according to claim 7 in which the sectors are octants and the number of sectors is eight.
9. A method according to claim 1 in which the partial filtering operation is a rank-order operation.
10. A method according to claim 1 in which the partial filtering operation is an averaging operation.
11. A method according to claim 4 in which the minimum values from pairs of regions are processed by a rank-order operation.
12. A method according to claim 4 in which the minimum values from pairs of regions are processed by an averaging operation.
13. In video sequence processing, apparatus for filtering motioncompensated pixel-to-pixel dissimilarity values in which the filter aperture is decomposed into two or more regions and the outputs of partial filters applied to each region are combined by a non-linear operation.
14. Apparatus according to claim 13 in which the dissimilarity value is a rectified displaced field difference.
15. Apparatus according to claim 13 in which the regions are nonoverlapping.
16. Apparatus according to claim 13 in which the non-linear combination process includes taking minimum values of partial-filter outputs from pairs of regions that are diametrically opposite each other in the filter aperture.
17. Apparatus according to claim 16 in which the partial filters operate on radial line segments, preferably in which the number of radial line segments is eight.
18. Apparatus according to claim 16 in which the partial filters operate on sectors of the filter aperture, preferably in which the sectors are octants and the number of sectors is eight.
19. The apparatus according to claim 11 or 12, wherein the partial filters are configured to filter motion-compensated pixel-to-pixel dissimilarity values.
5
20. The apparatus according to claim 11 or 12, wherein the partial filters are configured to filter displaced frame pixel-to-pixel dissimilarity values.
19. Apparatus according to claim 13 in which the partial filtering operation is a rank-order operation.
20. Apparatus according to claim 13 in which the partial filtering operation is an averaging operation.
21. Apparatus according to claim 16 in which the minimum values from pairs of regions are processed by a rank-order operation.
22. Apparatus according to claim 16 in which the minimum values from pairs of regions are processed by an averaging operation.
23. A method of video sequence processing, comprising the steps of deriving a candidate motion vector representing the displacement of an object between first and second images of the video sequence, each image
- 12 being formed of pixels; using the motion vector to project from the first image to the second image a current pixel and a plurality of pixels neighbouring the current pixel, the current pixel and the neighbouring pixels defining a filter aperture; providing for each pixel in the filter aperture a dissimilarity value indicative of the dissimilarity between the pixel in the second image and the pixel projected from the first image; and deriving from the plurality of dissimilarity values in the filter aperture an error value for the candidate motion vector; characterised in that the filter aperture is decomposed into two or more regions; the dissimilarity values in each region are spatially filtered and the outputs of the spatial filters applied to each region are combined by a non-linear operation to provide the error value for the candidate motion vector.
24. Programmable apparatus programmed to implement a method according to any one of Claims 1 to 12 or 23.
25. A computer program product adapted to cause programmable apparatus to implement a method according to any one of Claims 1 to 12 or 23.
Amendments to the claims are as follows:
1. An apparatus for motion estimation in image or video processing, the apparatus comprising:
a filter aperture decomposed into two or more regions forming partial filters configured to filter pixel-to-pixel dissimilarity values between a plurality of video frames in a video sequence; and a motion estimator configured to estimate motion between the plurality of video frames by combining respective outputs of the partial filters; wherein the motion estimator is configured to combine the respective outputs of the partial filters by a non-linear operation.
2. The apparatus according to claim 1, wherein each of the pixel-to-pixel dissimilarity values is a rectified displaced field difference.
3. The apparatus according to claim 1, wherein each of the two or more regions are nonoverlapping.
4. The apparatus according to claim 1, wherein the non-linear operation comprises taking minimum values of the respective outputs of the partial filters from pairs of regions that are diametrically opposite each other in the filter aperture.
5. The apparatus according to claim 4, wherein the partial filters operate on radial line segments.
6. The apparatus according to claim 5, wherein the radial line segments comprise eight radial line segments, with respective pairs of the radial line segments diametrically opposing each other.
7. The apparatus according to claim 4, wherein the partial filters are configured to operate on sectors of the filter aperture.
8. The apparatus according to claim 7, wherein the sectors are octants and the number of sectors is eight.
9. The apparatus according to claim 1, wherein the partial filters are configured to filter motion-compensated pixel-to-pixel dissimilarity values.
10. The apparatus according to claim 1, wherein the partial filters are configured to filter displaced frame pixel-to-pixel dissimilarity values.
11. An apparatus for motion estimation in image or video processing, the apparatus comprising:
a spatial filter configured to operate on pixel-to-pixel dissimilarity values, with the spatial filter having a filter aperture that is decomposed into two or more filter windows that each define respective subsets of the pixel-to-pixel dissimilarity values of a plurality of video frames, and a plurality of partial filters applied respectively to each filter window; and a combiner configured to combine the respective outputs of the partial filters by performing a non-linear operation on the respective outputs from respective pairs of filter windows that are diametrically opposite each other in the filter aperture.
12. The apparatus of claim 11, wherein the non-linear operation comprises taking minimum values of the respective outputs from respective pairs of filter windows that are diametrically opposite each other in the filter aperture.
13. The apparatus according to claim 11 or 12, wherein each of the pixel-to-pixel dissimilarity values is a rectified displaced field difference.
14. The apparatus according to claim 11 or 12, wherein two or more regions corresponding the two or more filter windows are non-overlapping regions in each video frame of the plurality of video frames.
15. The apparatus according to claim 11, wherein the partial filters operate on radial line segments.
16. The apparatus according to claim 15, wherein the radial line segments comprise eight radial line segments, with respective pairs of the radial line segments diametrically opposing each other.
17. The apparatus according to claim 11, wherein the partial filters are configured to operate on sectors of the filter aperture.
CO
18. The apparatus according to claim 17, wherein the sectors are octants and the number of sectors is eight.
GB1905665.4A 2012-04-04 2012-04-04 Video sequence processing Active GB2572497B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
GB1905665.4A GB2572497B (en) 2012-04-04 2012-04-04 Video sequence processing

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
GB1206065.3A GB2502047B (en) 2012-04-04 2012-04-04 Video sequence processing
GB1905665.4A GB2572497B (en) 2012-04-04 2012-04-04 Video sequence processing

Publications (3)

Publication Number Publication Date
GB201905665D0 GB201905665D0 (en) 2019-06-05
GB2572497A true GB2572497A (en) 2019-10-02
GB2572497B GB2572497B (en) 2019-12-18

Family

ID=66810124

Family Applications (1)

Application Number Title Priority Date Filing Date
GB1905665.4A Active GB2572497B (en) 2012-04-04 2012-04-04 Video sequence processing

Country Status (1)

Country Link
GB (1) GB2572497B (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0899687A1 (en) * 1997-08-26 1999-03-03 Matsushita Electric Industrial Co., Ltd. Method and apparatus for detecting motion vector and image coding apparatus

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP0899687A1 (en) * 1997-08-26 1999-03-03 Matsushita Electric Industrial Co., Ltd. Method and apparatus for detecting motion vector and image coding apparatus

Also Published As

Publication number Publication date
GB2572497B (en) 2019-12-18
GB201905665D0 (en) 2019-06-05

Similar Documents

Publication Publication Date Title
EP2711891B1 (en) Stereo image processing device and stereo image processing method
EP0540762B1 (en) Method for detecting moving vector and apparatus therefor, and system for processing image signal using the apparatus
US20170085912A1 (en) Video sequence processing
Bestagini et al. Detection of temporal interpolation in video sequences
EP2064671B1 (en) Method and apparatus for interpolating an image
WO2001074072A1 (en) Processing sequential video images to detect image motion among interlaced video fields or progressive video images
US11849137B2 (en) Setting selection values for motion estimation vectors based on remote motion vectors of interpolated frames
EP2107521A2 (en) Detecting a border region in an image
JP4213035B2 (en) Occlusion detector and method for detecting an occlusion region
EP1958451B1 (en) Motion vector field correction
US8306123B2 (en) Method and apparatus to improve the convergence speed of a recursive motion estimator
EP2178289B1 (en) Method and unit for motion detection based on a difference histogram
US8565309B2 (en) System and method for motion vector collection for motion compensated interpolation of digital video
US20110216831A1 (en) Apparatus and method for motion vector filtering based on local image segmentation and lattice maps
US7881500B2 (en) Motion estimation with video mode detection
US20080144716A1 (en) Method For Motion Vector Determination
WO2004082294A1 (en) Method for motion vector determination
Park et al. Covariance-based adaptive deinterlacing method using edge map
US8462170B2 (en) Picture attribute allocation
US9648339B2 (en) Image processing with segmentation using directionally-accumulated difference-image pixel values
GB2572497A (en) Video sequence processing
GB2513112A (en) Video sequence processing
KR100949137B1 (en) Apparatus for video interpolation, method thereof and computer recordable medium storing the method
KR20020044715A (en) Apparatus and method for compensating video motions
GB2500909A (en) Selecting motion vectors on the basis of acceleration