WO2003049036A2 - Discriminating between synthetic and natural image regions - Google Patents

Discriminating between synthetic and natural image regions Download PDF

Info

Publication number
WO2003049036A2
WO2003049036A2 PCT/IB2002/005066 IB0205066W WO03049036A2 WO 2003049036 A2 WO2003049036 A2 WO 2003049036A2 IB 0205066 W IB0205066 W IB 0205066W WO 03049036 A2 WO03049036 A2 WO 03049036A2
Authority
WO
WIPO (PCT)
Prior art keywords
probability
pixels
values
image
particular pixel
Prior art date
Application number
PCT/IB2002/005066
Other languages
French (fr)
Other versions
WO2003049036A3 (en
Inventor
Giulio Ferretti
Johannes W. G. C. Vos
Paola Carrai
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics N.V. filed Critical Koninklijke Philips Electronics N.V.
Priority to AU2002351070A priority Critical patent/AU2002351070A1/en
Publication of WO2003049036A2 publication Critical patent/WO2003049036A2/en
Publication of WO2003049036A3 publication Critical patent/WO2003049036A3/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/40Analysis of texture
    • G06T7/41Analysis of texture based on statistical description of texture
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20092Interactive image processing based on input by user
    • G06T2207/20104Interactive definition of region of interest [ROI]

Definitions

  • Classification unit for and method of discriminating between synthetic and natural image regions
  • the invention relates to a method of discriminating between synthetic image regions and natural image regions in an image.
  • the invention further relates to a classification unit for discriminating between synthetic image regions and natural image regions in an image.
  • the invention further relates to an image display apparatus comprising: means to receive an image; a classification unit for discriminating between synthetic image regions and natural image regions in the image; an image processing unit to process the image based on output of the classification unit; and a display device to display the processed image.
  • natural or synthetic content-dedicated algorithms By discriminating between the data representing regions of the image that are either classified as natural or synthetic, natural or synthetic content-dedicated algorithms can then be employed so as to provide for further, and particularly appropriate, and accurate, signal processing applications. Without such segmentation, application of an algorithm to the whole image occurs and disadvantages can arise. For example, the same image-enhancement algorithms applied to both natural and synthetic regions of an image may produce significant improvements in the perceived quality of the natural image regions but may lead disadvantageously to artifacts in the synthetic parts of the image.
  • US-A-6195459 discloses an algorithm arranged for discriminating between natural and synthetic regions of an image which provides for a block-analysis of the image with subsequent clustering of blocks found likely to fall either in the synthetic or natural category.
  • The, generally rectangular, area formed by such clustered blocks is then refined and either accepted as a synthetic or accepted as a natural region responsive to further analysis steps, or discarded.
  • such a known arrangement is disadvantageously limited in the range of graphics patterns that can be accurately identified and also with regard to its general accuracy and efficiency and its sensitivity to noise.
  • this known algorithm is arranged to operate in accordance with a method that is considered unnecessarily complex and which exhibits a relatively high computational load which can disadvantageously restrict the usability of the algorithm in some circumstances.
  • the method comprises: a number of probability estimation steps, each step estimating for a particular pixel of the image an elementary probability value representing a probability of the particular pixel of being located in one of the natural image regions based on values of pixels of a group of pixels in a neighborhood of the particular pixel; and a combination step of calculating for the particular pixel a final probability value representing the probability of the particular pixel of being located in one of the natural image regions by combining the respective elementary probability values estimated in the probability estimation steps.
  • the mix of estimation steps result in a reliable final probability value indicating the probability of being located in one of the natural image regions, even by using values of very few pixels.
  • values of pixels e.g. luminance or color levels of the pixels. Only a small portion of the image is used to calculate the final value of probability.
  • the group of pixels corresponds to a block of pixels.
  • the block of pixels comprises typically 2x2, 3x3, 3x5, 3x7 or 5x5 pixels.
  • the method according to the invention can be applied in low- cost small delay streaming data processing. E.g. in the case of a block of 3x3 pixels the method can be used to label pixels of an image to be displayed with a measure of
  • the output is not a binary classification, but comprises a range of values. This is especially useful if content dependent processing downstream is designed to pass softly from one type of processing to another as the content changes, hence reducing artifacts due to content misclassification.
  • Estimation steps are carried out on a group of pixels in the neighborhood of the particular pixel.
  • the particular pixel is included in the group of pixels and is the central pixel of the group of pixels.
  • the same group of pixels is used in the various probability estimation steps related to the particular pixel.
  • the group of pixels is extended with additional pixels.
  • An embodiment of the method according to the invention is characterized in that in a first one of the probability estimation steps, the associated elementary probability value is estimated by dividing a number of different pixel values that is present in the group of pixels by a value that is related to the number of pixels in the group of pixels. This estimation step is simple and fast.
  • This probability estimation step results in a probability value NOV (Number Of Values) which equals 1 if the number of distinct pixel values that is present in the group of pixels is equal to the number of pixels in the group of pixels, whereas the probability value NOV equals 0 in a flat portion of the image.
  • NOV Number Of Values
  • An embodiment of the method according to the invention is characterized in that in a second one of the probability estimation steps, the associated elementary probability value is estimated by means of weighted summation of differences between pixel values of pixels of the group and the differences corresponding to distances between non-zero bins in a histogram of pixel values of the group of pixels.
  • this probability estimation step a simplified histogram analysis is carried out. Since natural images often comprise colors or luminance values very close to each other, small differences are scored better than large differences.
  • the histogram or array of present pixel values is scanned and the differences between the present pixel values are calculated. Each difference is then weighted and added to compute the second probability value.
  • This second probability value is further called SOV (Separation of Values).
  • An embodiment of the method according to the invention is characterized in that in a third one of the probability estimation steps, the associated elementary probability value is estimated by calculating absolute values of two-directional gradients of pixel values of pixels of the group of pixels. Use is made of the knowledge that artificial graphic elements usually spread horizontally or vertically. That is why two-directional gradients are calculated to emphasis diagonal gradients.
  • An embodiment of the method according to the invention is characterized in that in the third one of the probability estimation steps, the associated elementary probability value is calculated by means of weighting a sum of the absolute values of the two-directional gradients.
  • the associated elementary probability value is calculated by means of weighting a sum of the absolute values of the two-directional gradients.
  • a scalar value which can later on be used for succeeding calculations for the particular pixel
  • multiple gradient values which have been calculated for the group of pixels have to be combined. To achieve this the non-zero values are averaged and weighted.
  • An embodiment of the method according to the invention is characterized in that in the third one of the probability estimation steps a LUT is applied for weighting the sum of the absolute values of the two-directional gradients.
  • a Look-Up Table (LUT) is very easy approach to implement functions.
  • Preferably also a weighting function for the calculation of the second probability value is implemented by means of a LUT.
  • An embodiment of the method according to the invention is characterized in that in the combination step the final probability value is calculated by summation of the elementary probability values estimated in the probability estimation steps, divided by the number of elementary probability values. This is a fast and easy approach.
  • An embodiment of the method according to the invention is characterized in that in the combination step the final probability value is calculated by means of a thresholded power sum of the elementary probability values estimated in the probability estimation steps. It is preferable that the final probability value is high if one of the probability values is relatively high regardless to the other probability values being calculated during the estimation steps.
  • Fig. 1 schematically shows an embodiment of the classification unit according to the invention
  • Fig. 2 shows an example of a weighting function for the gradient estimator
  • Fig. 3 schematically shows an embodiment of the image display apparatus.
  • Fig. 1 schematically shows an embodiment of the classification unit 100 for discriminating between synthetic image regions 116 and natural image regions 118 in an image 114.
  • the classification unit 100 comprises: three probability estimators 102-106, each estimators arranged to estimate for a particular pixel 120 of the image 114 respective probability values related to a probability of being located in one of the natural image regions 118 based on values of pixels of a block 122 of pixels in a neighborhood of the particular pixel 120; and a combination unit 108 designed to calculate for the particular pixel 120 one final probability value related to the probability of being located in one of the natural image regions 118 by combining the respective probability values estimated by the probability estimators 102-106.
  • an image 114 is provided on the input connector 110 of the classification unit 100 .
  • This image 114 has one large synthetic region 116 comprising text and other types of graphics, and has one natural image region 118 representing photographic data which has been captured by a camera and then digitized.
  • the classification unit 100 is arranged to classify the pixels of the image 114 by using a sliding window approach.
  • the output of the classification unit 100 is provided at the output connector 112.
  • This output is a two-dimensional matrix 124 in which each element corresponds to a respective pixel of the image 114.
  • the values of the elements represent the probability of being located in a natural image region 118.
  • a portion 126 of the elements of the two-dimensional matrix 124 is labeled as "natural".
  • This embodiment is arranged to process images with a maximum of 256 distinct luminance values.
  • the output is not a binary classification, but the range of values of the elements of the two-dimensional matrix 124 has also 256 different values. This is especially useful if content dependent processing downstream is designed to pass softly from one type of processing to another as the content changes, hence reducing artifacts due to content misclassification.
  • the scale ranges from 0 for synthetic to 255 for natural.
  • the first probability estimator 102 is designed to calculate the probability value NOV (Number Of Values). This corresponds with the calculation as specified by Equation 1:
  • the second probability estimator 104 is designed to- calculate the probability value SOV (Separation of Values). This probability value is calculated by means of weighted summation of differences between pixel values of pixels of the block 122, with the differences corresponding to distances between non-zero bins in a histogram of pixel values of the group of pixels.
  • the second estimator 104 is designed to perform a simplified histogram analysis. The histogram or array of present pixel values is scanned and the differences between the present values are calculated.
  • each difference is then weighted and added to compute the probability value SOV .
  • the distance or separation between each value of this list and the next value in the list is calculated. This is illustrated by means of an example.
  • the block 122 of pixels comprises 9 pixels with the following values: ⁇ 1,1,3,4,7,7,250,255,255 ⁇ .
  • the following values can be distinguished: ⁇ 1,3,4,7,250,255 ⁇ and the differences between the distinguishable values are ⁇ 2,1,3,243,5 ⁇ . See table 1.
  • Equation 3 An example of a weighting function is given in Equation 3:
  • npix is equal to the total number of pixels in the block of pixels
  • k a constant that is related to the maximum separation which is expected in a natural image region: typically 30-40. Since natural images often have values that do not differ much in small portions of the image, small differences are scored higher than large differences. Hence, SOV equals 1 in the case of a natural image region, whereas the SOV equals 0 in a synthetic image region.
  • the third probability estimator 106 is designed to calculate the probability value GRD (Gradient). This probability value is calculated by means of calculating absolute values of two-directional gradients of pixel values of pixels of the block. Using the knowledge that artificial graphic elements usually spread horizontally or vertically, the two-
  • GRD Since is a local property in a particular location of the picture, GRD is only dxdy calculated using very small blocks of pixels e.g. 2x2 or 3x3 around the particular pixel, thus reducing the computational effort.
  • the horizontal differences of the original n x m block of pixels are differentiated again vertically, obtaining a block of pixels with dimensions (n - 1) x (m - 1) .
  • n - 1) x (m - 1) .
  • m - 1) a block of pixels with dimensions (n - 1) x (m - 1) .
  • Lum p2 p5 p8 p2> p6 p9
  • the combination unit 108 is arranged to calculate for each pixel the final probability value FPV based on the probability values which have been determined for each pixel.
  • the final probability value FPV can be calculated by means " of a summation or by means of a thresholded power sum.
  • the summation is given by Equation 6 and the power sum by Equation 7.
  • Equation 7 The effect of applying Equation 7 is that the value of FPV is low in the case that the values of the probability values NO V , SOV and GRD are all relatively low, but the value of FPV is near to the maximum if one of the probability values is relatively high, regardless to the values of the other probability values.
  • Fig. 2 shows an example of a weighting function W (AGRD) for the gradient estimator 106.
  • the x-axis 204 corresponds the parameter AGRD
  • the y-axis corresponds with probability value GRD .
  • the weighting function 200 according to Equation 5 is depicted.
  • the weighting function gives maximum score to the smallest non-zero value and gives a lower score to higher values. Above a predetermined threshold 208 all scores are equal to zero.
  • the weighting function can be implemented by means of a LUT.
  • Fig. 3 schematically shows an embodiment of the image display apparatus 300 according to the invention.
  • the image display apparatus comprises: means to receive an image 302; a classification unit 100 for discriminating between synthetic image regions and natural image regions in the image. This classification unit 100 is as described in connection with Fig. 1; - an image processing unit 306 to process the image based on output of the classification unit 100; and a display device 308 to display the processed image.
  • image data will be provided to the display apparatus 300 via the input connector 310 as a video signal.
  • the image data might e.g. be rendered by a computer system and converted to a video signal by the video controller of the computer system. It can be either a analogues or a digital signal.
  • the image Before being displayed on the display device 308 the image is processed by the image processing unit 306. As a control signal the output of the classification unit 100 is provided to the image processing unit 306. Appropriate processing is performed depending on the type of data: natural image regions are processed differently form synthetic image regions.
  • the image classification and image processing coufd also be performed by the computer system before the image data is sent to a display device.

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Probability & Statistics with Applications (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

A classification unit (100) for discriminating between synthetic image regions (116) and natural image regions (118) in an image (114) comprises a number of probability estimators (102-106), each for estimating for a particular pixel (120) of the image respective probability values related to a probability of being located in one of the natural image regions (118) based on values of pixels of a group (122) of pixels in a neighborhood of the particular pixel (120). The classification unit (100) further comprises a combination unit (108) designed to calculate for the particular pixel (120) one final probability value related to the probability of being located in one of the natural image regions (118) by combining the probability values estimated by the respective probability estimators (102-106).

Description

Classification unit for and method of discriminating between synthetic and natural image regions
The invention relates to a method of discriminating between synthetic image regions and natural image regions in an image.
The invention further relates to a classification unit for discriminating between synthetic image regions and natural image regions in an image. The invention further relates to an image display apparatus comprising: means to receive an image; a classification unit for discriminating between synthetic image regions and natural image regions in the image; an image processing unit to process the image based on output of the classification unit; and a display device to display the processed image.
Many aspects of signal processing applications, such as feature extraction and content driven processing, compression and retrieval operations, are heavily dependent upon the ability to accurately segment the image into regions that are considered likely to represent a natural image, such as a photo or video image, and regions likely to represent so-called synthetic images such as computer generated text and/or graphics.
By discriminating between the data representing regions of the image that are either classified as natural or synthetic, natural or synthetic content-dedicated algorithms can then be employed so as to provide for further, and particularly appropriate, and accurate, signal processing applications. Without such segmentation, application of an algorithm to the whole image occurs and disadvantages can arise. For example, the same image-enhancement algorithms applied to both natural and synthetic regions of an image may produce significant improvements in the perceived quality of the natural image regions but may lead disadvantageously to artifacts in the synthetic parts of the image.
Thus, it can prove inappropriate to attempt to enhance the complete image without first seeking to discriminate, and separate, the natural regions of the image from synthetic regions of the image. Once such different regions have been identified, respective appropriate processing algorithms can then be applied.
Of course, further advantages can arise in handling the image data in this manner. For example, the automatic optimization of the bandwidth utilization in coding applications such as arranging a fax machine to adopt separate encoding schemes for video images and for pure text/graphics content can be achieved.
US-A-6195459 discloses an algorithm arranged for discriminating between natural and synthetic regions of an image which provides for a block-analysis of the image with subsequent clustering of blocks found likely to fall either in the synthetic or natural category. The, generally rectangular, area formed by such clustered blocks is then refined and either accepted as a synthetic or accepted as a natural region responsive to further analysis steps, or discarded. However, such a known arrangement is disadvantageously limited in the range of graphics patterns that can be accurately identified and also with regard to its general accuracy and efficiency and its sensitivity to noise.
Also, this known algorithm is arranged to operate in accordance with a method that is considered unnecessarily complex and which exhibits a relatively high computational load which can disadvantageously restrict the usability of the algorithm in some circumstances.
It is an object of the invention to provide a method of the kind described in the opening paragraph which can be applied in real-time image processing applications relatively easy.
This object of the invention is achieved in that the method comprises: a number of probability estimation steps, each step estimating for a particular pixel of the image an elementary probability value representing a probability of the particular pixel of being located in one of the natural image regions based on values of pixels of a group of pixels in a neighborhood of the particular pixel; and a combination step of calculating for the particular pixel a final probability value representing the probability of the particular pixel of being located in one of the natural image regions by combining the respective elementary probability values estimated in the probability estimation steps.
The mix of estimation steps result in a reliable final probability value indicating the probability of being located in one of the natural image regions, even by using values of very few pixels. With values of pixels is meant e.g. luminance or color levels of the pixels. Only a small portion of the image is used to calculate the final value of probability. Preferably the group of pixels corresponds to a block of pixels. The block of pixels comprises typically 2x2, 3x3, 3x5, 3x7 or 5x5 pixels. The method according to the invention can be applied in low- cost small delay streaming data processing. E.g. in the case of a block of 3x3 pixels the method can be used to label pixels of an image to be displayed with a measure of
"naturality", adding a negligible delay of one to two times the scan time of one line of the display device. Most classification methods work off-line and use all pixels of the entire image.
Another advantage is that the output is not a binary classification, but comprises a range of values. This is especially useful if content dependent processing downstream is designed to pass softly from one type of processing to another as the content changes, hence reducing artifacts due to content misclassification.
Estimation steps, requiring a relatively low computational effort, are carried out on a group of pixels in the neighborhood of the particular pixel. Preferably the particular pixel is included in the group of pixels and is the central pixel of the group of pixels.
Preferably the same group of pixels is used in the various probability estimation steps related to the particular pixel. However it is possible that for some probability estimation steps the group of pixels is extended with additional pixels.
Below three basically different estimations steps are disclosed. It should be noted that other probability estimations steps can be applied too and that the number of probability estimation steps can also differ from three. The probability estimation steps are called first, second and third respectively. However these names are used for identification purposes only. These names are not related to any order or combination of probability estimation steps. An embodiment of the method according to the invention is characterized in that in a first one of the probability estimation steps, the associated elementary probability value is estimated by dividing a number of different pixel values that is present in the group of pixels by a value that is related to the number of pixels in the group of pixels. This estimation step is simple and fast. This probability estimation step results in a probability value NOV (Number Of Values) which equals 1 if the number of distinct pixel values that is present in the group of pixels is equal to the number of pixels in the group of pixels, whereas the probability value NOV equals 0 in a flat portion of the image.
An embodiment of the method according to the invention is characterized in that in a second one of the probability estimation steps, the associated elementary probability value is estimated by means of weighted summation of differences between pixel values of pixels of the group and the differences corresponding to distances between non-zero bins in a histogram of pixel values of the group of pixels. In this probability estimation step a simplified histogram analysis is carried out. Since natural images often comprise colors or luminance values very close to each other, small differences are scored better than large differences. In this embodiment the histogram or array of present pixel values is scanned and the differences between the present pixel values are calculated. Each difference is then weighted and added to compute the second probability value. This second probability value is further called SOV (Separation of Values). An embodiment of the method according to the invention is characterized in that in a third one of the probability estimation steps, the associated elementary probability value is estimated by calculating absolute values of two-directional gradients of pixel values of pixels of the group of pixels. Use is made of the knowledge that artificial graphic elements usually spread horizontally or vertically. That is why two-directional gradients are calculated to emphasis diagonal gradients.
An embodiment of the method according to the invention is characterized in that in the third one of the probability estimation steps, the associated elementary probability value is calculated by means of weighting a sum of the absolute values of the two-directional gradients. In order to calculate a scalar value, which can later on be used for succeeding calculations for the particular pixel, multiple gradient values which have been calculated for the group of pixels have to be combined. To achieve this the non-zero values are averaged and weighted.
An embodiment of the method according to the invention is characterized in that in the third one of the probability estimation steps a LUT is applied for weighting the sum of the absolute values of the two-directional gradients. A Look-Up Table (LUT) is very easy approach to implement functions. Preferably also a weighting function for the calculation of the second probability value is implemented by means of a LUT.
An embodiment of the method according to the invention is characterized in that in the combination step the final probability value is calculated by summation of the elementary probability values estimated in the probability estimation steps, divided by the number of elementary probability values. This is a fast and easy approach.
An embodiment of the method according to the invention is characterized in that in the combination step the final probability value is calculated by means of a thresholded power sum of the elementary probability values estimated in the probability estimation steps. It is preferable that the final probability value is high if one of the probability values is relatively high regardless to the other probability values being calculated during the estimation steps.
Modifications of the method and variations thereof may correspond to modifications and variations thereof of the classification unit and of the image display apparatus described.
These and other aspects of the method, of the classification unit and of the image display apparatus according to the invention will become apparent from and will be elucidated with respect to the implementations and embodiments described hereinafter and with reference to the accompanying drawings, wherein:
Fig. 1 schematically shows an embodiment of the classification unit according to the invention; Fig. 2 shows an example of a weighting function for the gradient estimator; and
Fig. 3 schematically shows an embodiment of the image display apparatus.
Corresponding reference numerals have the same meaning in all of the Figs.
Fig. 1 schematically shows an embodiment of the classification unit 100 for discriminating between synthetic image regions 116 and natural image regions 118 in an image 114. The classification unit 100 comprises: three probability estimators 102-106, each estimators arranged to estimate for a particular pixel 120 of the image 114 respective probability values related to a probability of being located in one of the natural image regions 118 based on values of pixels of a block 122 of pixels in a neighborhood of the particular pixel 120; and a combination unit 108 designed to calculate for the particular pixel 120 one final probability value related to the probability of being located in one of the natural image regions 118 by combining the respective probability values estimated by the probability estimators 102-106.
On the input connector 110 of the classification unit 100 an image 114 is provided. This image 114 has one large synthetic region 116 comprising text and other types of graphics, and has one natural image region 118 representing photographic data which has been captured by a camera and then digitized. The classification unit 100 is arranged to classify the pixels of the image 114 by using a sliding window approach. The output of the classification unit 100 is provided at the output connector 112. This output is a two-dimensional matrix 124 in which each element corresponds to a respective pixel of the image 114. The values of the elements represent the probability of being located in a natural image region 118. In Fig. 1 it is depicted that a portion 126 of the elements of the two-dimensional matrix 124 is labeled as "natural". This embodiment is arranged to process images with a maximum of 256 distinct luminance values. The output is not a binary classification, but the range of values of the elements of the two-dimensional matrix 124 has also 256 different values. This is especially useful if content dependent processing downstream is designed to pass softly from one type of processing to another as the content changes, hence reducing artifacts due to content misclassification. The scale ranges from 0 for synthetic to 255 for natural.
The first probability estimator 102 is designed to calculate the probability value NOV (Number Of Values). This corresponds with the calculation as specified by Equation 1:
_ (number of different values in the block -1) (number of pixels in the block -I)
Hence, NOV equals 1 if the number of distinct pixel values that is present in the group of pixels is equal to the number of pixels in the group of pixels, whereas the NOV equals 0 in a flat portion of the image. The second probability estimator 104 is designed to- calculate the probability value SOV (Separation of Values). This probability value is calculated by means of weighted summation of differences between pixel values of pixels of the block 122, with the differences corresponding to distances between non-zero bins in a histogram of pixel values of the group of pixels. The second estimator 104 is designed to perform a simplified histogram analysis. The histogram or array of present pixel values is scanned and the differences between the present values are calculated. Each difference is then weighted and added to compute the probability value SOV . Once the list of values that is present in the block 122 is ordered, the distance or separation between each value of this list and the next value in the list is calculated. This is illustrated by means of an example. Suppose that the block 122 of pixels comprises 9 pixels with the following values: {1,1,3,4,7,7,250,255,255}. Then the following values can be distinguished: {1,3,4,7,250,255} and the differences between the distinguishable values are {2,1,3,243,5}. See table 1. These differences are called separations, St
Table 1
Figure imgf000008_0003
Each separation St is then weighted and added to compute the probability value SOV . This corresponds with the calculation as specified by Equation 2: nc-l
SOV = ∑Ws(Si)> (2)
!=1 where nc is the number of different values in the block and Ws ( ) a weighting function for variable n . An example of a weighting function is given in Equation 3:
Figure imgf000008_0001
where npix is equal to the total number of pixels in the block of pixels, and k a constant that is related to the maximum separation which is expected in a natural image region: typically 30-40. Since natural images often have values that do not differ much in small portions of the image, small differences are scored higher than large differences. Hence, SOV equals 1 in the case of a natural image region, whereas the SOV equals 0 in a synthetic image region. The third probability estimator 106 is designed to calculate the probability value GRD (Gradient). This probability value is calculated by means of calculating absolute values of two-directional gradients of pixel values of pixels of the block. Using the knowledge that artificial graphic elements usually spread horizontally or vertically, the two-
directional derivative order to emphasize diagonal gradients.
Figure imgf000008_0002
d Lum .
Since is a local property in a particular location of the picture, GRD is only dxdy calculated using very small blocks of pixels e.g. 2x2 or 3x3 around the particular pixel, thus reducing the computational effort.
In terms of pixel values, the horizontal differences of the original n x m block of pixels are differentiated again vertically, obtaining a block of pixels with dimensions (n - 1) x (m - 1) . For example, given a block of pixels with dimension of 3 x 3 pixels: pi pA p7
Lum = p2 p5 p8 p2> p6 p9
The horizontal gradient is computed giving the new matrix:
~(pA - p\) (p7 - pA) dlυm (P5 -p2)(p8 -p5) dx (p6 - p3)(P9 - p6) which is then differentiated again vertically. After that the absolute values are determined: f d ^2\vcm> "abs((/75 - P2) - (pA - pi)) abs((/;8 - p5) - (p7 - pA)) 2ndGrad = abs dxdy _abs(G?6 - p3) - (p5 - p2)) abs((/>9 - pβ) - (p8 - p5))
Then the non zero values are averaged:
^2ndGrady
AGRD =- nonzerovalues (4) number of non zero values Finally the gradient GRD is determined by applying an appropriate weighting function Wg .
(See Fig. 2)
GRD = Wg(AGRD) (5)
The combination unit 108 is arranged to calculate for each pixel the final probability value FPV based on the probability values which have been determined for each pixel. Hence, when the probability values NOV , SOV and GRD have been calculated for a pixel, the final probability value FPV can be calculated by means "of a summation or by means of a thresholded power sum. The summation is given by Equation 6 and the power sum by Equation 7.
(NOV + SOV + GRD)
FPV = 255. (6)
FPV = min(255,[(l6 - NOV)2 + (16 - SOV)2 + (16 GRD)2 J) (7)
The effect of applying Equation 7 is that the value of FPV is low in the case that the values of the probability values NO V , SOV and GRD are all relatively low, but the value of FPV is near to the maximum if one of the probability values is relatively high, regardless to the values of the other probability values.
Fig. 2 shows an example of a weighting function W (AGRD) for the gradient estimator 106. The x-axis 204 corresponds the parameter AGRD , and the y-axis corresponds with probability value GRD . Hence the weighting function 200 according to Equation 5 is depicted. The weighting function gives maximum score to the smallest non-zero value and gives a lower score to higher values. Above a predetermined threshold 208 all scores are equal to zero. The weighting function can be implemented by means of a LUT.
Fig. 3 schematically shows an embodiment of the image display apparatus 300 according to the invention. The image display apparatus comprises: means to receive an image 302; a classification unit 100 for discriminating between synthetic image regions and natural image regions in the image. This classification unit 100 is as described in connection with Fig. 1; - an image processing unit 306 to process the image based on output of the classification unit 100; and a display device 308 to display the processed image. Typically image data will be provided to the display apparatus 300 via the input connector 310 as a video signal. The image data might e.g. be rendered by a computer system and converted to a video signal by the video controller of the computer system. It can be either a analogues or a digital signal. Before being displayed on the display device 308 the image is processed by the image processing unit 306. As a control signal the output of the classification unit 100 is provided to the image processing unit 306. Appropriate processing is performed depending on the type of data: natural image regions are processed differently form synthetic image regions.
The image classification and image processing coufd also be performed by the computer system before the image data is sent to a display device.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention and that those skilled in the art will be able to design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be constructed as limiting the claim. The word 'comprising' does not exclude the presence of elements or steps not listed in a claim. The word "a" or "an" preceding an element does not exclude the presence of a plurality of such elements. The invention can be implemented by means of hardware comprising several distinct elements and by means of a suitable programmed computer. In the unit claims enumerating several means, several of these means can be embodied by one and the same item of hardware.

Claims

CLAIMS:
1. A method of discriminating between synthetic image regions (116) and natural image regions (118) in an image (114), characterized in that the method comprises: a number of probability estimation steps, each step estimating for a particular pixel (120) of the image an elementary probability value representing a probability of the particular pixel of being located in one of the natural image regions (118) based on values of pixels of a group (122) of pixels in a neighborhood of the particular pixel (120); and a combination step of calculating for the particular pixel (120) a final probability value representing the probability of the particular pixel of being located in one of the natural image regions (118) by combining the respective elementary probability values estimated in the probability estimation steps.
2. A method as claimed in Claim 1 , characterized in that in a first one of the probability estimation steps, the associated elementary probability value is estimated by dividing a number of different pixel values that is present in the group (122) of pixels by a value that is related to the number of pixels in the group (122) of pixels.
3. A method as claimed in Claim 1, characterized in that in a second one of the probability estimation steps, the associated elementary probability value is estimated by means of weighted summation of differences between pixel values of pixels of the group and the differences corresponding to distances between non-zero bins in a histogram of pixel values of the group (122) of pixels.
4. A method as claimed in Claim 1 , characterized in that in a third one of the probability estimation steps, the associated elementary probability value is estimated by calculating absolute values of two-directional gradients of pixel values of pixels of the group (122) of pixels.
5. A method as claimed in Claim 4, characterized in that in the third one of the probability estimation steps, the associated elementary probability value is calculated by means of weighting a sum of the absolute values of the two-directional gradients.
6. A method as claimed in Claim 5, characterized in that in the third one of the probability estimation steps a LUT is applied for weighting the sum of the absolute values of the two-directional gradients.
7. A method as claimed in Claim 1, characterized in that in the combination step the final probability value is calculated by summation of the elementary probability values estimated in the probability estimation steps, divided by the number of elementary probability values.
8. A method as claimed in Claim 1, characterized in that in the combination step the final probability value is calculated by means of a thresholded power sum of the elementary probability values estimated in the probability estimation steps.
9. A classification unit (100) for discriminating between synthetic image regions (116) and natural image regions (118) in an image (114), characterized in that the classification unit (100) comprises: a number of probability estimators (102-106), each estimator (102-106) arranged to estimate for a particular pixel (120) of the image an elementary probability value representing a probability of the particular pixel of being located in one of the natural image regions (118) based on values of pixels of a group (122) of pixels in a neighborhood of the particular pixel (120); and a combination unit (108) for calculating for the particular pixel (120) a final probability value representing the probability of the particular pixel of being located in one of the natural image regions (118) by combining the respective elementary probability values estimated by the probability estimators (102-106).
10. A classification unit (100) as claimed in Claim 9 and arranged with means for carrying out the steps according to the method of any of the claims 1-8.
11. An image display apparatus (300) comprising: means to receive an image (302); a classification unit (100) for discriminating between synthetic image regions (116) and natural image regions (118) in the image (114); an image processing unit (306) to process the image (114) based on output of the classification unit (100); and a display device (308) to display the processed image, characterized in that the classification unit (100) comprises:
- a number of probability estimators (102-106), each estimator (102-106) arranged to estimate for a particular pixel (120) of the image an elementary probability value representing a probability of the particular pixel of being located in one of the natural image regions (118) based on values of pixels of a group (122) of pixels in a neighborhood of the particular pixel (120); and
- a combination unit (108) for calculating for the particular pixel (120) a final probability value representing the probability of the particular pixel of being located in one of the natural image regions (118) by combining the respective elementary probability values estimated by the probability estimators (102-106).
PCT/IB2002/005066 2001-12-04 2002-11-29 Discriminating between synthetic and natural image regions WO2003049036A2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU2002351070A AU2002351070A1 (en) 2001-12-04 2002-11-29 Discriminating between synthetic and natural image regions

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP01204678 2001-12-04
EP01204678.5 2001-12-04

Publications (2)

Publication Number Publication Date
WO2003049036A2 true WO2003049036A2 (en) 2003-06-12
WO2003049036A3 WO2003049036A3 (en) 2004-06-03

Family

ID=8181349

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2002/005066 WO2003049036A2 (en) 2001-12-04 2002-11-29 Discriminating between synthetic and natural image regions

Country Status (2)

Country Link
AU (1) AU2002351070A1 (en)
WO (1) WO2003049036A2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006087666A1 (en) * 2005-02-16 2006-08-24 Koninklijke Philips Electronics N.V. Method for natural content detection and natural content detector
WO2013144942A1 (en) * 2012-03-28 2013-10-03 I.C.V.T. Ltd. Controlling a compression of an image according to a degree of photo-realism

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5546474A (en) * 1993-12-21 1996-08-13 Hewlett-Packard Company Detection of photo regions in digital images
US6195459B1 (en) * 1995-12-21 2001-02-27 Canon Kabushiki Kaisha Zone segmentation for image display

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5546474A (en) * 1993-12-21 1996-08-13 Hewlett-Packard Company Detection of photo regions in digital images
US6195459B1 (en) * 1995-12-21 2001-02-27 Canon Kabushiki Kaisha Zone segmentation for image display

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
J[HNE B: "Digital Image Processing - 3rd Edition" 1995 , SPRINGER-VERLAG , BERLIN HEIDELBERG XP002269187 Chapter 12 (p. 219-230): Classification *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2006087666A1 (en) * 2005-02-16 2006-08-24 Koninklijke Philips Electronics N.V. Method for natural content detection and natural content detector
WO2013144942A1 (en) * 2012-03-28 2013-10-03 I.C.V.T. Ltd. Controlling a compression of an image according to a degree of photo-realism
US9241165B2 (en) 2012-03-28 2016-01-19 Beamr Imaging Ltd Controlling a compression of an image according to a degree of photo-realism

Also Published As

Publication number Publication date
AU2002351070A1 (en) 2003-06-17
AU2002351070A8 (en) 2003-06-17
WO2003049036A3 (en) 2004-06-03

Similar Documents

Publication Publication Date Title
JP4746050B2 (en) Method and system for processing video data
JP4017489B2 (en) Segmentation method
JP3964327B2 (en) Method and apparatus for determining a region of interest in an image and image transmission method and apparatus
US6263113B1 (en) Method for detecting a face in a digital image
US8121403B2 (en) Methods and systems for glyph-pixel selection
JP4498422B2 (en) Pixel classification method and image processing apparatus
JP2005505870A (en) Method and apparatus for identifying different regions of an image
JP2008148298A (en) Method and apparatus for identifying regions of different content in image, and computer readable medium for embodying computer program for identifying regions of different content in image
EP1918872A2 (en) Image segmentation method and system
US7606414B2 (en) Fusion of color space data to extract dominant color
CN111028276A (en) Image alignment method and device, storage medium and electronic equipment
CN110399873A (en) ID Card Image acquisition methods, device, electronic equipment and storage medium
CN108205657A (en) Method, storage medium and the mobile terminal of video lens segmentation
KR20050085355A (en) Method and apparatus for removing false edges from a segmented image
CN106331746B (en) Method and apparatus for identifying watermark location in video file
US8311269B2 (en) Blocker image identification apparatus and method
EP3014528B1 (en) Determining barcode locations in documents
CN109543487B (en) Automatic induction triggering method and system based on bar code edge segmentation
CN112949423B (en) Object recognition method, object recognition device and robot
CN113840135A (en) Color cast detection method, device, equipment and storage medium
US6999621B2 (en) Text discrimination method and related apparatus
JP2010140315A (en) Object detection device
CN107704864B (en) Salient object detection method based on image object semantic detection
JP4967045B2 (en) Background discriminating apparatus, method and program
KR20060007901A (en) Apparatus and method for automatic extraction of salient object from an image

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ OM PH PL PT RO RU SC SD SE SG SI SK SL TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR IE IT LU MC NL PT SE SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2002785782

Country of ref document: EP

WWW Wipo information: withdrawn in national office

Ref document number: 2002785782

Country of ref document: EP

NENP Non-entry into the national phase in:

Ref country code: JP

WWW Wipo information: withdrawn in national office

Country of ref document: JP