WO2010073553A1 - 画像処理装置、画像処理方法および記憶媒体 - Google Patents
画像処理装置、画像処理方法および記憶媒体 Download PDFInfo
- Publication number
- WO2010073553A1 WO2010073553A1 PCT/JP2009/006951 JP2009006951W WO2010073553A1 WO 2010073553 A1 WO2010073553 A1 WO 2010073553A1 JP 2009006951 W JP2009006951 W JP 2009006951W WO 2010073553 A1 WO2010073553 A1 WO 2010073553A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- correction value
- gradation correction
- saturation
- input
- Prior art date
Links
- 238000012545 processing Methods 0.000 title claims abstract description 78
- 238000003672 processing method Methods 0.000 title claims description 39
- 238000012937 correction Methods 0.000 claims abstract description 289
- 238000004458 analytical method Methods 0.000 claims abstract description 89
- 238000000034 method Methods 0.000 claims description 36
- 230000014509 gene expression Effects 0.000 claims description 16
- 230000008569 process Effects 0.000 claims description 14
- 238000010219 correlation analysis Methods 0.000 claims description 12
- 238000010586 diagram Methods 0.000 description 11
- 230000002093 peripheral effect Effects 0.000 description 9
- 230000007423 decrease Effects 0.000 description 7
- 238000006243 chemical reaction Methods 0.000 description 6
- 238000004364 calculation method Methods 0.000 description 4
- 238000012986 modification Methods 0.000 description 4
- 230000004048 modification Effects 0.000 description 4
- 230000000875 corresponding effect Effects 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 239000003086 colorant Substances 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000007781 pre-processing Methods 0.000 description 2
- 230000008859 change Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000006866 deterioration Effects 0.000 description 1
- 238000010348 incorporation Methods 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000012805 post-processing Methods 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/46—Colour picture communication systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/77—Circuits for processing the brightness signal and the chrominance signal relative to each other, e.g. adjusting the phase of the brightness signal relative to the colour signal, correcting differential gain or differential phase
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/40—Picture signal circuits
- H04N1/407—Control or modification of tonal gradation or of extreme levels, e.g. background level
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/46—Colour picture communication systems
- H04N1/56—Processing of colour picture signals
- H04N1/60—Colour correction or control
- H04N1/6002—Corrections within particular colour systems
- H04N1/6005—Corrections within particular colour systems with luminance or chrominance signals, e.g. LC1C2, HSL or YUV
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/46—Colour picture communication systems
- H04N1/56—Processing of colour picture signals
- H04N1/60—Colour correction or control
- H04N1/6027—Correction or control of colour gradation or colour contrast
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4318—Generation of visual interfaces for content selection or interaction; Content or additional data rendering by altering the content in the rendering process, e.g. blanking, blurring or masking an image region
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/44008—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics in the video stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/57—Control of contrast or brightness
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/64—Circuits for processing colour signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/64—Circuits for processing colour signals
- H04N9/68—Circuits for processing colour signals for controlling the amplitude of colour signals, e.g. automatic chroma control circuits
Definitions
- the present invention relates to an image processing apparatus, an image processing method, and a storage medium.
- the present invention relates to an image processing apparatus, an image processing method, and a storage medium storing the program for executing gradation correction in a color space represented by luminance components and color difference components such as YUV and YCrCb.
- a color space represented by luminance components and color difference components such as a YUV color space and a YCrCb color space is used.
- tone correction in a color space in which such images and images are represented by luminance components and color difference components is not necessary, conversion processing to other color spaces and inverse conversion processing are unnecessary, so from the viewpoint of calculation cost. This is advantageous.
- the YUV (YCrCb) color space is composed of a luminance component (Y component) representing brightness and a color difference component (U component, V component or Cr component, Cb component) representing color tone, and gradation correction in the YUV color space Can be executed by correcting the Y component.
- a luminance component Y component
- a color difference component U component, V component or Cr component, Cb component
- gradation correction in the YUV color space can be executed by correcting the Y component.
- the vividness (saturation) for viewing deteriorates due to the change of the Y component because there is a shift between the axes of the YUV color space and the axes of brightness and color perceived by humans.
- various image processing apparatuses or image processing methods have been proposed so far.
- Patent Document 1 determines the correction amount of a target pixel based on a weighted average luminance value of a local region around the target pixel, thereby increasing the level of a highlight region or a shadow region included in the image. It is possible to correct the key appropriately.
- Patent Document 2 adjusts the gradation correction amount from the average luminance of the area determined to be a face in the image, so that the face of the person who is the most emphasized area when viewing the image Makes it possible to achieve optimal brightness.
- Patent Document 3 uses Y in (x, y) representing the Y component value before correction at the pixel position (x, y) and Y out (x, y) representing the Y component value after correction. Is multiplied by the U component value and the V component value to suppress the deterioration of the vividness of appearance.
- this correction is expressed by an equation, it corresponds to the equation (1).
- Patent Document 4 is such that when the a * component and the b * component are held in the CIE L * a * b * color space, the ratio of the Y value before and after the correction in each YCrCb value and the correction before and after the correction.
- the relationship of the saturation ratio in the YCrCb color space is obtained by linear approximation, and the Cr component and the Cb component are corrected.
- This correction can also be applied to the YUV color space. If the correction in this case is expressed by an equation, it corresponds to equation (2).
- Equation (2) p is a linear approximation of the relationship between Y out (x, y) / Y in (x, y) and the ratio of the U component before and after correction or the ratio of the V component before and after correction. This is an estimated value.
- p 1, it is the same as Formula (1).
- the above technique is devised to maintain the vividness (saturation) of colors perceived by humans by correcting the luminance component (Y component) as well as the color difference components (U component, V component).
- the analysis of the correlation between the luminance component correction value (tone correction value) and the color difference component correction value (saturation correction value) is not sufficient, there is a large error between the actual value and the approximate value locally. There was a case.
- the present invention has been made in view of the above circumstances, and an object of the present invention is to perform an appropriate saturation correction in the gradation correction of an image represented by a luminance component and a color difference component.
- a processing apparatus, an image processing method, and a storage medium are provided.
- an image input unit that receives an input image
- a tone correction value acquisition unit that acquires a tone correction value indicating a ratio between the luminance component of the input image and the luminance component of the output image.
- Saturation analysis means for obtaining a saturation correction value that minimizes the total sum in association with the gradation correction value, the gradation correction value acquired by the gradation correction value acquisition means, and the gradation correction value Image output means for outputting, as the output image, an image obtained by correcting the input image received by the image input means based on the saturation correction value associated with Image processing apparatus characterized by There is provided.
- an image input step for receiving an input image, and a gradation correction value for acquiring a gradation correction value indicating a ratio between the luminance component of the input image and the luminance component of the output image Saturation divergence between the acquisition step, the analysis image that is the same as or different from the input image, and the corrected image in which the luminance component of the analysis image is corrected based on one or more gradation correction values
- a saturation analysis step for obtaining a saturation correction value that minimizes the sum of degrees in association with the gradation correction value, the gradation correction value acquired in the gradation correction value acquisition step, and
- An image processing method characterized in that it comprises is provided.
- the computer-readable storage medium is an image input process for receiving an input image that is input, and the ratio between the luminance component of the input image and the luminance component of the output image that is output.
- a tone correction value acquisition process for acquiring a tone correction value, an analysis image that is the same as or different from the input image, and a luminance component of the analysis image is corrected based on one or more tone correction values
- the saturation analysis process for obtaining a saturation correction value that minimizes the sum of the degree of saturation divergence between the corrected image and the gradation correction value
- an analysis image and a correction image obtained by correcting the analysis image based on a specific gradation correction value can be analyzed to obtain a saturation correction value corresponding to the gradation correction value. Therefore, the present invention can perform saturation correction with relatively high accuracy.
- an image processing apparatus capable of executing appropriate saturation correction in gradation correction of an image represented by a luminance component and a color difference component are provided.
- FIG. 1 is a configuration diagram of an image processing apparatus according to a first embodiment of the present invention. It is a flowchart which shows the image processing method of the 1st Embodiment of this invention. It is a block diagram of the image processing apparatus of the 2nd Embodiment of this invention. It is a scatter diagram which shows the gradation correction value (alpha) and saturation correction value (beta) which a memory
- FIG. 5 is a diagram illustrating both the plot of FIG. 4 and an approximate curve for interpolating between the plots. It is a flowchart showing the pre-process in the image processing method of the 2nd Embodiment of this invention. It is a flowchart showing the post-process in the image processing method of the 2nd Embodiment of this invention. It is the figure which represented the straight line which the approximate expression used for patent document 4 represents on FIG.
- FIG. 1 is a configuration diagram of an image processing apparatus according to the first embodiment.
- the image processing apparatus 100 of the present embodiment may include an image input unit 11 that receives an input image P in input.
- the image processing apparatus 100 may comprise a gradation correction value acquiring unit 12 that acquires the gradation correction value ⁇ indicating the ratio of the luminance component of the output image P out output the luminance component of the input image P in .
- the image processing apparatus 100 the input image P in the same or different and analysis image P a, one or more corrected corrected image luminance component of the analysis image P a on the basis of the gradation correction value ⁇ and P b, the saturation correction value deviance sum is minimized respective saturation ⁇ between may be provided with a saturation analysis unit 13 for determining in association with the gradation correction value alpha.
- the image processing apparatus 100 includes an image input unit based on the gradation correction value ⁇ acquired by the gradation correction value acquisition unit 12 and the saturation correction value ⁇ associated with the gradation correction value ⁇ .
- the input image P in accepted at 11 may be provided with an image output unit 14 for outputting an image obtained by correcting the output image P out.
- the image input unit 11, the gradation correction value acquisition unit 12, and the saturation analysis unit 13 may be included in the data processing unit 10.
- the data processing unit 10 may store the processed data, variables, and the like in the storage unit 16 as appropriate. Further, the data processing unit 10 may appropriately read programs, data, and the like stored in the storage unit 16.
- All or part of the configuration included in the image processing apparatus 100 may be realized by hardware. Or you may implement
- the computer reads and executes the program from a recording medium such as a nonvolatile memory.
- the processor may be the data processing unit 10
- the storage medium may be the storage unit 16.
- the program may be executed an image input process of receiving an input image P in input to the computer.
- the program may be executed gradation correction value acquisition processing for acquiring gradation correction value ⁇ indicating the ratio of the luminance component of the output image P out output the luminance component of the input image Pi n to the computer .
- the program the input image P in the same or different and analysis image P a, one or more corrected image P b to the luminance component of the analysis image P a on the basis of the gradation correction value ⁇ is corrected
- the computer may be caused to execute a saturation analysis process for determining the saturation correction value ⁇ that minimizes the sum of the saturation divergence between the two and the gradation correction value ⁇ .
- this program performs image input processing based on the gradation correction value ⁇ acquired in the gradation correction value acquisition processing and the saturation correction value ⁇ associated with the gradation correction value ⁇ .
- image obtained by correcting the input image P in the accepted Te may be an image output processing is executed by a computer to output as an output image P out of.
- the image input unit 11, an external device that the image processing apparatus 100 is connected (not shown) or an external network may be input to the input image P in the (not shown).
- the image input unit 11 may transmit the request for inputting the input image P in the external device or external network.
- the external device or external network that has received the input request may input the input image Pin into the image input unit 11 in response to the input request.
- the image input unit 11 sends the input image P in the saturation analysis unit 13.
- the input image P in which the image input unit 11 is sent may be a YUV image, the luminance component (Y component) and color difference components (U component, V component) may be expressed in.
- the input image P in input to the image input unit 11 may not be an image represented by the luminance component and color difference components such as YUV images and YCrCb images.
- conversion processing to the image represented by the luminance component and color difference components is not required as described above, without executing the conversion process, the input image P in expressed by a luminance component and color difference components delivery Is possible.
- the image processing apparatus 100 may include an operation input unit 15 that receives user operation inputs.
- the operation input unit 15 may be a method of inputting by pressing a key or a touch panel method.
- the gradation correction value acquisition unit 12 may acquire the gradation correction value ⁇ determined by default from the storage unit 16.
- the gradation correction value acquisition unit 12 may acquire an arbitrary value input by the user's operation input received by the operation input unit 15 as the gradation correction value ⁇ .
- the gradation correction value ⁇ may be calculated and acquired based on an arbitrary value input by a user operation input received by the operation input unit 15 and data stored in the storage unit 16.
- the gradation correction value acquiring unit 12 may calculate the gradation correction value ⁇ from the luminance component of the output image P out defined by default and the luminance component of the input image P in.
- the input image P when representing the pixel position of the pixel which is included in the input image P in input to the image input unit 11 is (x, y), the input image P input luminance components of the in luminance values Y in ( x, y). Further, the color difference components of the input image P in are defined as input color difference values U in (x, y) and V in (x, y).
- the luminance component of the output image P out is represented by the output luminance value Y out (x, y).
- y the luminance value
- the color difference components of the output image P out are defined as output color difference values U out (x, y) and V out (x, y).
- the relationship between the input luminance value Y in (x, y) and the output luminance value Y out (x, y), the input color difference values U in (x, y), V in (x, y), and the output color difference value U out (x , Y) and V out (x, y) may be expressed as in Expression (3).
- the image output unit 14 uses a value obtained by multiplying the luminance component (Y in (x, y)) of the input image P in by the gradation correction value ⁇ as an output image.
- the luminance component of the P out (Y out (x, y)) may be corrected as.
- the image output unit 14 then multiplies the color difference components (U in (x, y), V in (x, y)) of the input image P in by the saturation correction value ⁇ to obtain the output image P out of the color difference components (U out (x, y) , V out (x, y)) may be corrected as.
- Saturation analysis unit 13 generates corrected the same or different analysis image P a and the input image P in, the analysis image P a on the basis of the gradation correction value ⁇ to the gradation correction value acquiring unit 12 has acquired
- the saturation correction value ⁇ that minimizes the total sum of the saturation divergences between the corrected image Pb and the corrected image Pb may be obtained in association with the gradation correction value ⁇ .
- the saturation analysis unit 13 receives the input image Pin received by the image input unit 11 as the analysis image Pa.
- the saturation analysis unit 13 selects the input image P in (for analysis) based on the saturation correction value ⁇ set with a temporary value and the gradation correction value ⁇ acquired by the gradation correction value acquisition unit 12. corrected by applying equation (3) for some or all of the pixels are included in the image Pa), it may generate a corrected image P b.
- all the pixels to be corrected may be included in the input image Pin, but in this case, the processing load of the saturation analysis unit 13 is increased.
- the saturation analysis unit 13 may select and correct only some representative pixels included in the input image Pin to reduce the processing load.
- the corrected image P b to be generated is a collection of some of pixel corrected image P b correction.
- the saturation analysis unit 13 converts the saturation S b (x, y) for each pixel of the corrected image P b into a chromatic index (a) when the corresponding pixel is converted into the L * a * b * color space. * And b *).
- the chroma analysis unit 13 also converts the chroma S a (x, y) for each pixel of the input image P in (analysis image P a ) into the L * a * b * color space. It may be determined by (a * and b *).
- the saturation analysis unit 13 of the present embodiment represents the saturation with a chromaticity index in the L * a * b * color space
- the present invention is not limited to this.
- the technique used here is preferably a technique that expresses saturation using a uniform color space in which the distance between colors in the color space and the difference in human sensational color approximate.
- the saturation analysis unit 13 represents the degree of saturation divergence for each pixel as (S b (x, y) ⁇ S a (x, y)) 2 , and the degree of saturation divergence in all corrected pixels.
- ⁇ (S b (x, y) ⁇ S a (x, y)) 2 may be calculated.
- the saturation analysis unit 13 changes the value of the saturation correction value ⁇ , and the total sum ⁇ (S b (x, y) ⁇ S a ( x, y)) 2 are calculated and compared to determine the value of ⁇ that minimizes the sum ⁇ (S b (x, y) ⁇ S a (x, y)) 2 of divergence.
- the saturation correction value ⁇ is determined in this way, it is necessary to make the number of pixels corrected by the saturation analysis unit 13 constant even if the saturation correction value ⁇ is changed.
- the saturation analysis unit 13 may send the saturation correction value ⁇ obtained by the above-described method to the image output unit 14 in association with the gradation correction value ⁇ .
- FIG. 2 is a flowchart showing the image processing method of the present embodiment.
- the image processing method of the present embodiment may include an image input step of accepting an input image P in input (step S101).
- the image processing method of the embodiment the input image P in the gradation correction value acquisition step of acquiring tone correction value ⁇ indicating the ratio of the luminance component of the output image P out output the luminance component of the (step S102 ) May be provided.
- the image processing method of this embodiment is corrected input image P in the same or different and analysis image P a, one or more of the luminance component of the analysis image P a on the basis of the gradation correction value ⁇
- the image processing method of the present embodiment includes the gradation correction value ⁇ acquired in the gradation correction value acquisition step (Step S102), and the saturation correction value ⁇ associated with the gradation correction value ⁇ .
- based on an image input step may include an image output step of outputting an image obtained by correcting the input image P in accepted at (step S101) as an output image P out (step S104).
- the present embodiment analyzes the input image P in (analysis image P a ) and the corrected image P b obtained by correcting the input image P in (analysis image P a ) based on a specific gradation correction value ⁇ .
- the saturation correction value ⁇ corresponding to the gradation correction value ⁇ can be obtained.
- Patent Document 4 it is found that there is a correlation between the brightness component increase width and the chroma increase width (color difference component increase width), and the saturation increase width (color difference component increase width) is estimated from the brightness component increase width. ing. However, paying attention to the entire region in which the value of the luminance component is 30 to 200, the approximate value is obtained by linear approximation.
- the present embodiment focuses on one specific gradation correction value ⁇ (brightness component increase width) to obtain the saturation correction value ⁇ (color difference component increase width), so that saturation correction is performed with relatively high accuracy. Can be executed.
- the image processing apparatus 100 of the present embodiment by using the input image P in which the image input unit 11 is inputted as the analysis image P a, so obtaining a color saturation correction value beta, have no statistical measurements However, appropriate saturation correction can also be performed.
- FIG. 3 is a configuration diagram of the image processing apparatus of the present embodiment.
- the image input unit 21, the gradation correction value acquisition unit 22, and the operation input unit 25 provided in the image processing apparatus 200 of the present embodiment are the image input unit 11 and the gradation correction value acquisition unit 12 described in the first embodiment. These are the same components as the operation input unit 15.
- the image input unit 21 and the gradation correction value acquisition unit 22 may be included in the data processing unit 20.
- the data processing unit 20 may store the processed data, variables, and the like in the storage unit 26 as appropriate. Furthermore, the data processing unit 20 may appropriately read programs, data, and the like stored in the storage unit 26.
- the saturation analysis unit 23 will be described when receiving an input image P in a different image by the image input unit 21 is accepted as the analysis image P a. Specifically, saturation analysis unit 23 receives the image for learning as the analysis image P a, if stored in advance in the storage unit 26 in association with the gradation correction value ⁇ and the chroma correction ⁇ Is described.
- the image processing apparatus 200 includes one or more gradation correction values ⁇ used for correcting the analysis image Pa, and the gradation correction values. May further include a correction value storage unit 27 for storing the saturation correction value ⁇ associated with the.
- the correction value storage unit 27 may be realized in a part of the storage area in the storage unit 26.
- the gradation correction value acquisition unit 22 may acquire any of the gradation correction values ⁇ stored in the correction value storage unit 27.
- the image output unit 24 may acquire the saturation correction value ⁇ associated with the gradation correction value ⁇ acquired by the gradation correction value acquisition unit 22 from the correction value storage unit 27. Then, the image output unit 24 based on the obtained and the chroma correction value ⁇ gradation correction value alpha, the image may be output as an output image P out a obtained by correcting the input image P in.
- FIG. 4 is a scatter diagram showing the gradation correction value ⁇ and the saturation correction value ⁇ stored in the correction value storage unit 27.
- the gradation correction value ⁇ is on the horizontal axis and the saturation correction value ⁇ is on the vertical axis.
- the gradation correction value ⁇ is set in increments of 0.05 from 0.5 to 3.5.
- the saturation correction value ⁇ is obtained by applying the technique used by the saturation analysis unit 13 of the first embodiment to each gradation correction value ⁇ .
- the gradation correction value ⁇ and the saturation correction value ⁇ associated with each other are represented as one point.
- the image processing apparatus 200 when the correction value storage unit 27 stores a plurality of gradation correction values ⁇ and a plurality of saturation correction values ⁇ in association with each other, the image processing apparatus 200 further includes the following components. May be provided.
- the image processing apparatus 200 may further include a correlation analysis unit 28 that analyzes a correlation between a plurality of sets of gradation correction values ⁇ and saturation correction values ⁇ stored in the correction value storage unit 27.
- the image processing apparatus 200 may further include a correlation storage unit 29 that stores the correlation.
- the image output unit 24 acquires the correlation from the correlation storage unit 29, applies the correlation to the gradation correction value ⁇ acquired by the gradation correction value acquisition unit 22, and applies the saturation correction value ⁇ . You may ask for.
- the “saturation correction value ⁇ associated with the gradation correction value ⁇ ” described above is correlated with the gradation correction value ⁇ acquired by the gradation correction value acquisition unit 22 as in the present embodiment.
- the saturation correction value ⁇ obtained by applying the equation may also be included.
- the correlation storage unit 29 may be realized in a part of the storage area in the storage unit 26.
- FIG. 5 is a diagram showing both the plot of FIG. 4 and an approximate curve for interpolating between the plots. This approximate curve is obtained by power approximation. Also, the following equation (4) is a correlation equation representing this approximate curve.
- q may be arbitrarily determined by the user within a range of 0 ⁇ q ⁇ 1. This q may be determined by an operation input received by the operation input unit 15.
- the above description explains the concept of analysis processing performed by the correlation analysis unit 28.
- the correlation analysis unit 28 does not need to actually create a diagram such as FIG. 4 or FIG.
- the correlation analysis unit 28 only needs to be able to calculate the correlation equation (formula (4)) representing the correlation by ⁇ approximation using ⁇ and ⁇ stored in the storage unit 26.
- the above-described saturation analysis unit 23, image output unit 24, and correlation analysis unit 28 may be included in the data processing unit 20.
- the correlation analysis unit 28 may be included in the saturation analysis unit 23.
- FIG. 6 is a flowchart showing preprocessing in the image processing method of the present embodiment.
- the pre-processing is processing for accepting a learning image as the analysis image Pa and storing the gradation correction value ⁇ and the saturation correction ⁇ in advance in the storage unit 26 in association with each other.
- the image processing method of the present embodiment in the saturation analysis step (step S201), receives an image (input image P in different image received by the image input step) for learning as the analysis image P a.
- processing for analyzing image P a is the same as the chroma analyzing step (step S103), obtains a saturation correction value ⁇ using the method described in the first embodiment May be.
- step S201 in the saturation analysis step (step S201), and one or more gradation correction value ⁇ used in the correction of the analysis image P a, to the gradation correction value ⁇
- a correction value storage step (step S202) for storing the associated saturation correction value ⁇ in the storage unit 26 may be further provided.
- the image processing method of the present embodiment is a correlation analysis that analyzes the correlation between a plurality of sets of gradation correction values ⁇ and saturation correction values ⁇ stored in the storage unit 26 (correction value storage unit 27).
- a step (step S203) and a correlation storage step (step S204) for storing the correlation in the storage unit 26 (correlation storage unit 29) may be further provided.
- FIG. 7 is a flowchart showing post-processing in the image processing method of the present embodiment.
- post-treatment includes a chroma correction ⁇ gradation correction value ⁇ stored in advance in the storage unit 26, or by correcting the input image P in using these correlations, and outputs the output image P out It is processing.
- the image processing method of the present embodiment may include an image input unit (step S301).
- the image input unit (step S301) is the same step as the image input unit (step S101) of the first embodiment.
- step S302 In the image processing method of this embodiment, in the gradation correction value acquisition step (step S302), one of the gradation correction values ⁇ stored in the storage unit 26 (correction value storage unit 27) is acquired. Well. In the image output step (step S305), the saturation correction value ⁇ associated with the gradation correction value ⁇ acquired in the gradation correction value acquisition step (step S302) is stored in the storage unit 26 (correction value storage unit). 27).
- the correlation is acquired from the storage unit 26 (correlation storage unit 29) in the image output step (step S305), and is acquired in the gradation correction value acquisition step (step S302).
- the saturation correction value ⁇ may be obtained by applying a correlation to the gradation correction value ⁇ .
- the gradation correction value ⁇ and the saturation correction value ⁇ can be obtained or calculated from the data stored in the storage unit 26 without performing saturation analysis. Therefore, when the saturation analysis is not performed (NO in step S303), the image output step (step S305) may be executed without executing the saturation analysis step (step S304).
- the saturation analysis step (step S304) may be executed.
- the saturation analysis step (step S304) and the image output step (step S305) are the same steps as the saturation analysis step (step S103) and the image output step (step S104) described in the first embodiment, respectively. It is.
- the image processing apparatus 200 accepts a learning image as an analysis image Pa, associates the acquired gradation correction value ⁇ with the saturation correction value ⁇ obtained from the learning image, and stores them in the storage unit. 26 can be stored in advance.
- the image output unit 24, by using the gradation correction value ⁇ and the saturation correction value storage unit 26 stores, after receiving an input image P in, the processing load until an output image P out It is reduced.
- the image processing apparatus 200 analyzes the correlation between a plurality of sets of gradation correction values ⁇ and saturation correction values ⁇ stored in the storage unit 26, and obtains the acquired gradation correction values ⁇ .
- the saturation correction value ⁇ can be obtained.
- the gradation correction that requires the gradation correction value ⁇ not stored in the storage unit 26 can also be executed.
- FIG. 8 is a diagram showing the straight line represented by the equation (2) on FIG. Referring to FIG. 8, it can be seen that the error increases when ⁇ is small or large.
- Expression (4) can be an expression that more appropriately represents the correlation between ⁇ and ⁇ by comparing Expression (1) and Expression (2).
- the image processing apparatus 200 of the present embodiment can execute saturation correction more appropriate than the technique described in Patent Document 3 or Patent Document 4 even when using the approximate expression (Expression (4)).
- the input image P in which the image input unit 11 sends only needs to be represented by a luminance component and color difference components.
- the input image P in which the image input unit 11 is transmitted has been described with YUV image but, as a variant, it may be a YCrCb image.
- the saturation analysis unit 13 sets the sum of the divergences of saturation so that ⁇ (S b (x, y) ⁇ S a (x, y)) 2 is minimized.
- the value of the saturation correction value ⁇ is set so that the root mean square ⁇ (S b (x, y) ⁇ S a (x, y)) 2 / N of the saturation deviation is minimized. It may be determined. However, N is the number of pixels in the corrected image P b.
- the image output unit 14 multiplies the gradation correction value ⁇ to the brightness component of the input image P in to give the value to be, as the luminance component of the output image P out, a value obtained by multiplying the chroma correction value ⁇ to the color difference component of the input image P in, has been described as the color difference component of the output image P out.
- the present invention is not limited to this.
- the image output unit 14 (or the image output unit 24), a value obtained by adding the gradation correction value ⁇ to the brightness component of the input image P in, or as the luminance component of the output image P out. Then, the image output unit 14 (or the image output unit 24), a value obtained by adding the chroma correction value ⁇ to the color difference component of the input image P in, or as a color difference component of the output image P out.
- This can be expressed by the following equation (5).
- q ′ may be arbitrarily determined by the user within a range of 0 ⁇ q ′ ⁇ 1. This q ′ may be determined by an operation input received by the operation input unit 15 (or the operation input unit 25). Furthermore, as a modification of the present invention, at least one of the correction value storage unit 27 and the correlation storage unit 29 may be realized as a storage device different from the storage unit 26.
- the image processing apparatus 100 (or the image processing apparatus 200) can also apply a retrograde correction technique in order to improve visibility. Details will be described below.
- the image processing apparatus 100 includes a peripheral luminance value Y ave (x, y) representing the brightness of the peripheral area of the pixel at each pixel position (x, y) included in the input image Pin input by the image input unit 11. May be calculated. Equations (6), (7), and (8) relating to this calculation are shown below.
- H (x, y, s, t) is an isotropic two-dimensional normal distribution centered on the pixel position (x, y) and having a deviation ⁇ , and satisfies Expressions (7) and (8). .
- the peripheral luminance value represented by (6) Y ave (x, y) is the weighted synthesized luminance component in the peripheral region around the pixel position (x, y), the peripheral luminance value Y ave (x, The larger y) means that the pixel position (x, y) is included in a bright area.
- the peripheral luminance value Y ave (x, y) is discretely calculated by Formula (6) with a certain interval, and the peripheral luminance at the pixel positions in between is calculated by linear interpolation or There is also a method of calculating by bicubic interpolation.
- the image output unit 14 the luminance component Y in (x, y) of the input image P in and Y ave (x, y) performs a gradation correction to input the luminance component Y out (x of the output image P out, y) is calculated.
- An example of a method for calculating Y out (x, y) is shown in Expression (9). However, Y max is the maximum value that Y out (x, y) can take.
- L (x, y) is a value calculated based on the peripheral luminance value Y ave (x, y) at the pixel position (x, y).
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Processing (AREA)
- Facsimile Image Signal Circuits (AREA)
- Color Image Communication Systems (AREA)
- Processing Of Color Television Signals (AREA)
Abstract
Description
上述した問題点を解決するため、これまでにも種々の画像処理装置または画像処理方法が提案されている。
〔第1の実施形態〕
続いて、本実施形態の画像処理方法について説明する。図2は、本実施形態の画像処理方法を示すフローチャートである。
〔第2の実施形態〕
以上、図面を参照して本発明の実施形態について述べたが、これらは本発明の例示であり、上記以外の様々な構成を採用することもできる。
さらに、本発明の変形例として、補正値記憶部27または相関関係記憶部29の少なくとも一方は、記憶部26とは異なる記憶装置として実現されてもよい。
Claims (17)
- 入力された入力画像を受け付ける画像入力手段と、
前記入力画像の輝度成分と出力された出力画像の輝度成分の比率を示す階調補正値を取得する階調補正値取得手段と、
前記入力画像と同一または異なる解析用画像と、一つまたは複数の階調補正値に基づいて前記解析用画像の輝度成分を補正された補正画像と、の間における彩度の乖離度の総和がそれぞれ最小となる彩度補正値を、当該階調補正値に対応付けて求める彩度解析手段と、
前記階調補正値取得手段が取得した前記階調補正値と、当該階調補正値に対応付けられた前記彩度補正値と、に基づいて、前記画像入力手段にて受け付けられた前記入力画像を補正して得られた画像を前記出力画像として出力する画像出力手段と、
を備えることを特徴とする画像処理装置。 - 前記彩度解析手段は、前記画像入力手段が受け付けた前記入力画像と異なる画像を前記解析用画像として受け付ける請求項1に記載の画像処理装置であって、
前記解析用画像の補正に用いた一つまたは複数の前記階調補正値と、当該階調補正値に対応付けられた前記彩度補正値と、を記憶する補正値記憶手段をさらに備えることを特徴とする画像処理装置。 - 請求項2に記載の画像処理装置において、
前記階調補正値取得手段は、前記補正値記憶手段が記憶している前記階調補正値のうちいずれかを取得するとともに、
前記画像出力手段は、前記階調補正値取得手段が取得した前記階調補正値に対応付けられた前記彩度補正値を前記補正値記憶手段から取得することを特徴とする画像処理装置。 - 請求項3に記載の画像処理装置において、
前記補正値記憶手段が記憶している複数組の前記階調補正値と前記彩度補正値との相関関係を解析する相関関係解析手段と、
前記相関関係を記憶する相関関係記憶手段と、をさらに備え、
前記画像出力手段は、
前記相関関係記憶手段から前記相関関係を取得し、前記階調補正値取得手段が取得した前記階調補正値に対して前記相関関係を適用して前記彩度補正値を求めることを特徴とする画像処理装置。 - 請求項4に記載の画像処理装置において、
前記相関関係解析手段は、前記相関関係を表す相関式を累乗近似により算出することを特徴とする画像処理装置。 - 請求項1乃至5いずれかに記載の画像処理装置において、
前記彩度解析手段は、
前記画像入力手段が受け付けた前記入力画像を前記解析用画像として受け付けることを特徴とする画像処理装置。 - 請求項1乃至6いずれかに記載の画像処理装置において、
前記画像出力手段は、
前記入力画像の輝度成分に前記階調補正値を乗算して得られる値を、前記出力画像の輝度成分とするとともに、
前記入力画像の色差成分に前記彩度補正値を乗算して得られる値を、前記出力画像の色差成分とすることを特徴とする画像処理装置。 - 請求項1乃至6いずれかに記載の画像処理装置において、
前記画像出力手段は、
前記入力画像の輝度成分に前記階調補正値を加算して得られる値を、前記出力画像の輝度成分とするとともに、
前記入力画像の色差成分に前記彩度補正値を加算して得られる値を、前記出力画像の色差成分とすることを特徴とする画像処理装置。 - 入力された入力画像を受け付ける画像入力ステップと、
前記入力画像の輝度成分と出力された出力画像の輝度成分の比率を示す階調補正値を取得する階調補正値取得ステップと、
前記入力画像と同一または異なる解析用画像と、一つまたは複数の階調補正値に基づいて前記解析用画像の輝度成分を補正された補正画像と、の間における彩度の乖離度の総和がそれぞれ最小となる彩度補正値を、当該階調補正値に対応付けて求める彩度解析ステップと、
前記階調補正値取得ステップにて取得された前記階調補正値と、当該階調補正値に対応付けられた前記彩度補正値と、に基づいて、前記画像入力ステップにて受け付けられた前記入力画像を補正して得られた画像を前記出力画像として出力する画像出力ステップと、
を備えることを特徴とする画像処理方法。 - 前記彩度解析ステップは、前記画像入力ステップで受け付けた前記入力画像と異なる画像を前記解析用画像として受け付ける請求項9に記載の画像処理方法であって、
前記解析用画像の補正に用いた一つまたは複数の前記階調補正値と、当該階調補正値に対応付けられた前記彩度補正値と、を記憶手段に記憶させる補正値記憶ステップをさらに備えることを特徴とする画像処理方法。 - 請求項10に記載の画像処理方法において、
前記階調補正値取得ステップは、前記記憶手段が記憶している前記階調補正値のうちいずれかを取得するとともに、
前記画像出力ステップは、前記階調補正値取得ステップにて取得された前記階調補正値に対応付けられた前記彩度補正値を前記記憶手段から取得することを特徴とする画像処理方法。 - 請求項11に記載の画像処理方法において、
前記記憶手段が記憶している複数組の前記階調補正値と前記彩度補正値との相関関係を解析する相関関係解析ステップと、
前記相関関係を前記記憶手段に記憶させる相関関係記憶ステップと、をさらに備え、
前記画像出力ステップは、
前記記憶手段から前記相関関係を取得し、前記階調補正値取得ステップにて取得された前記階調補正値に対して前記相関関係を適用して前記彩度補正値を求めることを特徴とする画像処理方法。 - 請求項12に記載の画像処理方法において、
前記相関関係解析ステップは、前記相関関係を表す相関式を累乗近似により算出することを特徴とする画像処理方法。 - 請求項9乃至13いずれかに記載の画像処理方法において、
前記彩度解析ステップは、
前記画像入力ステップで受け付けた前記入力画像を前記解析用画像として受け付けることを特徴とする画像処理方法。 - 請求項9乃至14いずれかに記載の画像処理方法において、
前記画像出力ステップは、
前記入力画像の輝度成分に前記階調補正値を乗算して得られる値を、前記出力画像の輝度成分とするとともに、
前記入力画像の色差成分に前記彩度補正値を乗算して得られる値を、前記出力画像の色差成分とすることを特徴とする画像処理方法。 - 請求項9乃至14いずれかに記載の画像処理方法において、
前記画像出力ステップは、
前記入力画像の輝度成分に前記階調補正値を加算して得られる値を、前記出力画像の輝度成分とするとともに、
前記入力画像の色差成分に前記彩度補正値を加算して得られる値を、前記出力画像の色差成分とすることを特徴とする画像処理方法。 - コンピュータが読み出し可能な記憶媒体であって、
入力された入力画像を受け付ける画像入力処理と、
前記入力画像の輝度成分と出力された出力画像の輝度成分の比率を示す階調補正値を取得する階調補正値取得処理と、
前記入力画像と同一または異なる解析用画像と、一つまたは複数の階調補正値に基づいて前記解析用画像の輝度成分を補正された補正画像と、の間における彩度の乖離度の総和がそれぞれ最小となる彩度補正値を、当該階調補正値に対応付けて求める彩度解析処理と、
前記階調補正値取得処理にて取得された前記階調補正値と、当該階調補正値に対応付けられた前記彩度補正値と、に基づいて、前記画像入力処理にて受け付けられた前記入力画像を補正して得られた画像を前記出力画像として出力する画像出力処理と、
を前記コンピュータに実行させるプログラムを記憶していることを特徴とする記憶媒体。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010543814A JP5212481B2 (ja) | 2008-12-26 | 2009-12-17 | 画像処理装置、画像処理方法および記憶媒体 |
US13/142,211 US8891865B2 (en) | 2008-12-26 | 2009-12-17 | Image processing device, image processing method, and storage medium for performing a gradation correction operation in a color space |
CN200980151950.1A CN102265621B (zh) | 2008-12-26 | 2009-12-17 | 图像处理设备、图像处理方法以及存储介质 |
EP09834365.0A EP2384006B1 (en) | 2008-12-26 | 2009-12-17 | Image processing device, image processing method, and storage medium |
KR1020117014719A KR101279773B1 (ko) | 2008-12-26 | 2009-12-17 | 화상 처리 장치, 화상 처리 방법 및 기억 매체 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2008334911 | 2008-12-26 | ||
JP2008-334911 | 2008-12-26 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010073553A1 true WO2010073553A1 (ja) | 2010-07-01 |
Family
ID=42287206
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2009/006951 WO2010073553A1 (ja) | 2008-12-26 | 2009-12-17 | 画像処理装置、画像処理方法および記憶媒体 |
Country Status (6)
Country | Link |
---|---|
US (1) | US8891865B2 (ja) |
EP (1) | EP2384006B1 (ja) |
JP (1) | JP5212481B2 (ja) |
KR (1) | KR101279773B1 (ja) |
CN (1) | CN102265621B (ja) |
WO (1) | WO2010073553A1 (ja) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7657833B2 (en) | 2000-09-15 | 2010-02-02 | Hartford Fire Insurance Company | Real-time single entry multiple carrier interface (SEMCI) |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0355078B2 (ja) * | 1986-08-29 | 1991-08-22 | ||
JPH03201697A (ja) * | 1989-12-27 | 1991-09-03 | Matsushita Electric Ind Co Ltd | 映像信号補正方法及び映像信号補正装置 |
JP2748678B2 (ja) * | 1990-10-09 | 1998-05-13 | 松下電器産業株式会社 | 階調補正方法および階調補正装置 |
JPH11252584A (ja) * | 1998-02-26 | 1999-09-17 | Victor Co Of Japan Ltd | テレビ受像機の信号処理回路 |
JP2002204373A (ja) * | 2000-10-23 | 2002-07-19 | Sony Corp | 画像処理装置および方法、並びに記録媒体 |
JP2006018465A (ja) | 2004-06-30 | 2006-01-19 | Canon Inc | 画像処理方法および画像処理装置およびコンピュータプログラムおよび記憶媒体 |
JP2006031255A (ja) | 2004-07-14 | 2006-02-02 | Mitsubishi Electric Corp | Dsrc車載器 |
WO2006025486A1 (ja) | 2004-09-01 | 2006-03-09 | Nec Corporation | 画像補正処理システム及び画像補正処理方法 |
JP2007142500A (ja) | 2005-11-14 | 2007-06-07 | Pioneer Electronic Corp | 表示装置、信号処理回路、プログラム及び表示方法 |
JP2007259242A (ja) * | 2006-03-24 | 2007-10-04 | Seiko Epson Corp | 映像信号処理 |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5489998A (en) * | 1991-03-04 | 1996-02-06 | Canon Kabushiki Kaisha | Color image processing apparatus having transmitter and receiver which perform color correction in accordance with a common standard color image and method for same |
JP3208814B2 (ja) * | 1992-01-09 | 2001-09-17 | 松下電器産業株式会社 | 映像信号補正装置 |
JP3626966B2 (ja) * | 1996-03-28 | 2005-03-09 | コニカミノルタビジネステクノロジーズ株式会社 | 画像処理装置 |
JPH10145582A (ja) * | 1996-11-05 | 1998-05-29 | Canon Inc | 画像処理方法及び装置 |
JP3596677B2 (ja) * | 2000-10-06 | 2004-12-02 | セイコーエプソン株式会社 | 画像処理装置、画像処理方法、プログラムおよび記録媒体 |
JP2003162715A (ja) * | 2001-11-22 | 2003-06-06 | Sharp Corp | 画像処理装置、画像処理方法および画像処理プログラムを記録した記録媒体、並びに画像入力装置および画像出力装置 |
KR100943274B1 (ko) * | 2002-12-20 | 2010-02-23 | 삼성전자주식회사 | 색신호 보정장치 및 그 방법, 및 그것을 이용한영상신호처리 시스템 및 그 방법 |
US7298892B2 (en) * | 2003-04-09 | 2007-11-20 | Eastman Kodak Company | Producing a balanced digital color image having minimal color errors |
JP2004336656A (ja) * | 2003-05-12 | 2004-11-25 | Konica Minolta Photo Imaging Inc | 輝度及び彩度調整方法、輝度及び彩度調整装置並びに画像出力装置 |
JP2005051739A (ja) * | 2003-07-16 | 2005-02-24 | Ricoh Co Ltd | 画像処理装置、画像処理方法および該画像処理方法による画像処理プログラムならびに該画像処理プログラムを収容した記録媒体 |
JP4189328B2 (ja) * | 2004-01-16 | 2008-12-03 | セイコーエプソン株式会社 | 画像処理装置、画像表示装置、画像処理方法および画像処理プログラム |
JP2006318255A (ja) | 2005-05-13 | 2006-11-24 | Konica Minolta Photo Imaging Inc | 画像処理方法、画像処理装置及び画像処理プログラム |
US8374458B2 (en) * | 2005-12-21 | 2013-02-12 | Nec Corporation | Tone correcting method, tone correcting apparatus, tone correcting program, and image equipment |
US7865031B2 (en) * | 2006-04-18 | 2011-01-04 | Tandent Vision Science, Inc. | Method and system for automatic correction of chromatic aberration |
KR100809346B1 (ko) * | 2006-07-03 | 2008-03-05 | 삼성전자주식회사 | 엣지 보정 장치 및 방법 |
CN101009851B (zh) * | 2007-01-19 | 2010-07-07 | 北京中星微电子有限公司 | 一种图像处理方法及其装置 |
TW200847799A (en) * | 2007-05-18 | 2008-12-01 | Mstar Semiconductor Inc | Image adjustment device and method thereof |
KR101274050B1 (ko) * | 2007-06-25 | 2013-06-12 | 엘지디스플레이 주식회사 | 선호색 보정 방법 및 장치와, 그를 이용한 액정 표시 장치 |
TWI356631B (en) * | 2008-04-02 | 2012-01-11 | Primax Electronics Ltd | Illumination-compensating method for image data an |
-
2009
- 2009-12-17 CN CN200980151950.1A patent/CN102265621B/zh not_active Expired - Fee Related
- 2009-12-17 US US13/142,211 patent/US8891865B2/en not_active Expired - Fee Related
- 2009-12-17 EP EP09834365.0A patent/EP2384006B1/en not_active Not-in-force
- 2009-12-17 JP JP2010543814A patent/JP5212481B2/ja not_active Expired - Fee Related
- 2009-12-17 WO PCT/JP2009/006951 patent/WO2010073553A1/ja active Application Filing
- 2009-12-17 KR KR1020117014719A patent/KR101279773B1/ko active IP Right Grant
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0355078B2 (ja) * | 1986-08-29 | 1991-08-22 | ||
JPH03201697A (ja) * | 1989-12-27 | 1991-09-03 | Matsushita Electric Ind Co Ltd | 映像信号補正方法及び映像信号補正装置 |
JP2748678B2 (ja) * | 1990-10-09 | 1998-05-13 | 松下電器産業株式会社 | 階調補正方法および階調補正装置 |
JPH11252584A (ja) * | 1998-02-26 | 1999-09-17 | Victor Co Of Japan Ltd | テレビ受像機の信号処理回路 |
JP2002204373A (ja) * | 2000-10-23 | 2002-07-19 | Sony Corp | 画像処理装置および方法、並びに記録媒体 |
JP2006018465A (ja) | 2004-06-30 | 2006-01-19 | Canon Inc | 画像処理方法および画像処理装置およびコンピュータプログラムおよび記憶媒体 |
JP2006031255A (ja) | 2004-07-14 | 2006-02-02 | Mitsubishi Electric Corp | Dsrc車載器 |
WO2006025486A1 (ja) | 2004-09-01 | 2006-03-09 | Nec Corporation | 画像補正処理システム及び画像補正処理方法 |
JP2007142500A (ja) | 2005-11-14 | 2007-06-07 | Pioneer Electronic Corp | 表示装置、信号処理回路、プログラム及び表示方法 |
JP2007259242A (ja) * | 2006-03-24 | 2007-10-04 | Seiko Epson Corp | 映像信号処理 |
Non-Patent Citations (1)
Title |
---|
See also references of EP2384006A4 |
Also Published As
Publication number | Publication date |
---|---|
US8891865B2 (en) | 2014-11-18 |
EP2384006A1 (en) | 2011-11-02 |
JPWO2010073553A1 (ja) | 2012-06-07 |
CN102265621A (zh) | 2011-11-30 |
EP2384006A4 (en) | 2015-12-16 |
EP2384006B1 (en) | 2017-06-28 |
KR20110099273A (ko) | 2011-09-07 |
CN102265621B (zh) | 2014-07-30 |
US20110293178A1 (en) | 2011-12-01 |
KR101279773B1 (ko) | 2013-07-04 |
JP5212481B2 (ja) | 2013-06-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4313032B2 (ja) | 映像の輝度制御装置及び方法 | |
US8103119B2 (en) | Image processing device, image processing method, image processing program product, and image-capturing device | |
US7755670B2 (en) | Tone-conversion device for image, program, electronic camera, and tone-conversion method | |
JP2014506757A (ja) | 映像の輝度変換方法及び装置 | |
JP4687673B2 (ja) | カラー画像のモノトーン化処理 | |
JP5235759B2 (ja) | 画像処理装置、画像処理方法、及びプログラム | |
KR101080846B1 (ko) | 보정 행렬의 추정에 의한 영상의 색상 손상 보정장치 및 방법 | |
US8345967B2 (en) | Apparatus and method of processing image, and record medium for the method | |
JP4321064B2 (ja) | 画像処理装置および画像処理プログラム | |
JP4001079B2 (ja) | カラー画像のモノトーン化処理 | |
JP2008048264A (ja) | 画像処理プログラムおよび画像処理装置 | |
JP5212481B2 (ja) | 画像処理装置、画像処理方法および記憶媒体 | |
JP5234127B2 (ja) | 階調変換装置、プログラム、電子カメラ、及びその方法 | |
JP2006114006A (ja) | 階調変換装置、プログラム、電子カメラ、およびその方法 | |
WO2007097125A1 (ja) | 画像の質感を補正する画像処理装置、画像処理プログラム、画像処理方法、および電子カメラ | |
JP5365881B2 (ja) | 画像処理装置、画像処理方法 | |
US20180063380A1 (en) | Image processing device | |
JP7022696B2 (ja) | 画像処理装置、画像処理方法およびプログラム | |
WO2019087403A1 (ja) | 画像処理装置、画像処理方法、及び画像処理プログラム | |
JP2007158446A (ja) | 画像処理装置、画像処理方法、プログラム、記憶媒体 | |
JP2019126025A (ja) | 色変換処理装置およびその制御方法 | |
JP2008530952A (ja) | 色プロファイルのための色補正テクニック | |
US9014472B2 (en) | Color balanced compression in dynamic range mismatched color systems | |
JP7005570B2 (ja) | 装置、方法、及びプログラム | |
JP6311345B2 (ja) | 画像処理装置及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200980151950.1 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09834365 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2010543814 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20117014719 Country of ref document: KR Kind code of ref document: A |
|
REEP | Request for entry into the european phase |
Ref document number: 2009834365 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13142211 Country of ref document: US Ref document number: 2009834365 Country of ref document: EP |
|
NENP | Non-entry into the national phase |
Ref country code: DE |