WO2006025486A1 - 画像補正処理システム及び画像補正処理方法 - Google Patents
画像補正処理システム及び画像補正処理方法 Download PDFInfo
- Publication number
- WO2006025486A1 WO2006025486A1 PCT/JP2005/015995 JP2005015995W WO2006025486A1 WO 2006025486 A1 WO2006025486 A1 WO 2006025486A1 JP 2005015995 W JP2005015995 W JP 2005015995W WO 2006025486 A1 WO2006025486 A1 WO 2006025486A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- gradation
- image processing
- local
- image
- value
- Prior art date
Links
- 238000000034 method Methods 0.000 title description 30
- 238000003702 image correction Methods 0.000 title description 19
- 238000012937 correction Methods 0.000 claims abstract description 393
- 230000002093 peripheral effect Effects 0.000 claims abstract description 118
- 238000012545 processing Methods 0.000 claims abstract description 112
- 238000004364 calculation method Methods 0.000 claims abstract description 89
- 238000006243 chemical reaction Methods 0.000 claims abstract description 22
- 238000003672 processing method Methods 0.000 claims description 49
- 230000003595 spectral effect Effects 0.000 claims description 12
- 230000000295 complement effect Effects 0.000 claims description 4
- 229910052900 illite Inorganic materials 0.000 claims description 2
- 238000012886 linear function Methods 0.000 claims description 2
- VGIBGUSAECPPNB-UHFFFAOYSA-L nonaaluminum;magnesium;tripotassium;1,3-dioxido-2,4,5-trioxa-1,3-disilabicyclo[1.1.1]pentane;iron(2+);oxygen(2-);fluoride;hydroxide Chemical compound [OH-].[O-2].[O-2].[O-2].[O-2].[O-2].[F-].[Mg+2].[Al+3].[Al+3].[Al+3].[Al+3].[Al+3].[Al+3].[Al+3].[Al+3].[Al+3].[K+].[K+].[K+].[Fe+2].O1[Si]2([O-])O[Si]1([O-])O2.O1[Si]2([O-])O[Si]1([O-])O2.O1[Si]2([O-])O[Si]1([O-])O2.O1[Si]2([O-])O[Si]1([O-])O2.O1[Si]2([O-])O[Si]1([O-])O2.O1[Si]2([O-])O[Si]1([O-])O2.O1[Si]2([O-])O[Si]1([O-])O2 VGIBGUSAECPPNB-UHFFFAOYSA-L 0.000 claims description 2
- 238000010586 diagram Methods 0.000 description 20
- 230000006870 function Effects 0.000 description 17
- 238000007796 conventional method Methods 0.000 description 2
- 230000005043 peripheral vision Effects 0.000 description 2
- 239000003086 colorant Substances 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 238000004590 computer program Methods 0.000 description 1
- 239000006185 dispersion Substances 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000001771 impaired effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/40—Image enhancement or restoration using histogram techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/20—Image enhancement or restoration using local operators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
- G06T5/94—Dynamic range modification of images or parts thereof based on local image properties, e.g. for local contrast enhancement
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N1/00—Scanning, transmission or reproduction of documents or the like, e.g. facsimile transmission; Details thereof
- H04N1/40—Picture signal circuits
- H04N1/407—Control or modification of tonal gradation or of extreme levels, e.g. background level
- H04N1/4072—Control or modification of tonal gradation or of extreme levels, e.g. background level dependent on the contents of the original
Definitions
- the present invention relates to an image correction processing system and an image correction processing method, and more particularly to a technique for improving an intermediate gradation of an image.
- a method for improving an intermediate gradation of an image is used to improve the gradation of an image in an apparatus that handles an image such as a camera, a scanner, or a printer.
- a conventional method for improving the intermediate gradation of an image the gradation value of an input original image is converted using a gradation conversion curve called a tone curve.
- a gradation conversion curve a global process is required in which the gradation conversion processing, which is fixed to the input gradation value and the output gradation value force, is applied to all pixels in the image.
- a gradation conversion curve When using a gradation conversion curve, a global process is required in which the gradation conversion processing, which is fixed to the input gradation value and the output gradation value force, is applied to all pixels in the image.
- the gradation conversion processing which is fixed to the input gradation value and the output gradation value force, is applied to all pixels in the image.
- adjustments can be made to obtain an image with a balanced tone in
- the halftone is improved by the processing flow shown in FIG. That is, in this conventional method, an image to be corrected is received as an input image (S801). At this time, a tone mask image is generated from the input image (S802). For each pixel value of the input image, a set of corrected pixel values is obtained (S803). The pixel value of the input image has a corresponding pixel mask value, and the corrected pixel value is obtained by a non-linear combination of the pixel value of the input image and the corresponding pixel mask value of the current pixel value.
- an automatic exposure control device is disclosed in Japanese Patent Laid-Open No. 5-176220.
- the average luminance of the entire screen of the image is detected, and the number of pixels having a luminance lower than a predetermined luminance is calculated.
- a correction amount is calculated based on the number of low-luminance pixels, and a target luminance for the entire screen is calculated based on the correction amount.
- the iris is controlled so that the average brightness follows the target brightness.
- An image processing apparatus is disclosed in Japanese Patent Laid-Open No. 11-355578.
- the correction amount of the granularity for each predetermined area is calculated based on each density of a plurality of predetermined areas obtained by dividing the area of the image data.
- the granularity of the image data is corrected based on the correction amount.
- 11-1 adjusts the intensity value for each position in each i-th spectral band.
- the adjusted intensity values for each position in each i th spectral band are filtered with a common function and adjusted for each i th spectral band so filtered for each position.
- An improved digital image based on the intensity value is displayed.
- an image processing method is disclosed in Japanese Patent Application Laid-Open No. 2003-333331.
- the signal is corrected so that nonlinear distortion of the input image data is reduced, and the relative ratio between the pixel value of the pixel of interest and the gray value of the surrounding distribution area is calculated with respect to the corrected signal. It is. From this relative ratio, the pixel value of the processing target pixel corresponding to the target pixel is determined.
- the relative ratio between the pixel value of the pixel of interest and the gray value of the surrounding distribution area is calculated by changing the target area, and the gain coefficient is set according to the size of the target surrounding distribution area. Calculated. Each relative ratio obtained is multiplied by a predetermined weighting factor and gain factor, respectively, and a combined value is calculated. The pixel value of the processing target pixel corresponding to the pixel of interest is determined from this composite value.
- An object of the present invention is to provide an image correction processing system, an image correction processing method, and an image correction processing program that can be processed with a small storage capacity.
- an image processing system includes a local region selection unit that selects a local region of at least one correction target with an input image force, and the local region includes at least one pixel.
- a correction amount calculation unit for calculating a gradation correction amount for the correction target local region, and a gradation correction for correcting the gradation value of the local region based on the gradation correction amount.
- a local gradation correction unit that generates a gradation correction image.
- the correction amount calculation unit calculates a gradation correction amount using a peripheral region gradation value calculation unit that calculates a gradation value of a peripheral region of the local region, and a calculated gradation value of the peripheral region. You may comprise the local correction amount conversion part to calculate.
- the image processing system further includes a partial region luminance calculation unit that calculates a representative gradation value for each partial region of the input image, and the correction amount calculation unit includes a peripheral region of the local region to be corrected.
- a partial area reference type peripheral area that calculates a gradation value using the representative gradation value of at least one partial area located in the vicinity of the local area to be corrected.
- a local luminance calculation unit; and a local correction amount conversion unit that calculates the gradation correction amount using the calculated gradation value of the peripheral region.
- the image processing system includes a global correction amount calculation unit that calculates a global correction amount of the local gradation correction image, and a gradation correction that corrects a global gradation in the local gradation correction image based on the global correction amount. And a global correction unit that generates a global gradation complementary image.
- the image processing system may further include an image input unit that provides the input image, and an image output unit that outputs the local tone correction image or the global tone correction image.
- the gradation value is preferably a luminance value or a gradation value expressed by a spectral component.
- the gradation correction may be repeated a plurality of times with respect to the correction target local region while changing the width of the correction target local region or maintaining the width of the correction target local region. Good.
- the gradation correction may be performed on the plurality of selected local regions.
- the peripheral area gradation value calculation unit may calculate the gradation value of the peripheral area and a weighted average luminance value of Gaussian weight, or the average luminance value of the peripheral area as the gradation value of the peripheral area. Good.
- one or more feature route lines connecting local regions and feature points are set, and one or more route feature points are obtained for each of the feature route lines, and the tone value of the route feature point or the route feature is obtained.
- a weighted average luminance value of points may be calculated as the gradation value of the surrounding area.
- the partial region reference type peripheral region luminance calculation unit is a product of a ratio of a distance between the local region to be corrected and at least one partial region located in the vicinity thereof and a representative gradation of the partial region. May be used to calculate the gradation of the surrounding area.
- the representative gradation value is preferably an average value of all or some of the pixels included in the partial area, or a median value of all or some of the pixels.
- the correction amount calculation unit is adapted to perform correction stronger than the shadow area and illite area.
- the gradation correction amount may be calculated, or the gradation correction amount of the local region may be calculated using the region luminance around the local region and the luminance of the highlight region of the image.
- the local gradation correction unit preferably performs the gradation correction using a nonlinear function.
- the global tone correction preferably includes contrast enhancement.
- an image processing method includes selecting at least one local area to be corrected for input image force, and including at least one pixel in the local area, and a gradation for the local area to be corrected. This is achieved by calculating a correction amount and generating a local tone correction image by performing tone correction for correcting the tone value of the local region based on the tone correction amount.
- the calculation of the gradation correction amount means calculating the gradation value of the peripheral region of the local region, and calculating the gradation correction amount using the calculated gradation value of the peripheral region. To be achieved by doing so.
- the image processing method further includes calculating a representative gradation value for each partial area of the input image, and calculating the gradation correction amount is positioned in the vicinity of the local area to be corrected. Calculating a gradation value of a peripheral area of the local area to be corrected using the representative gradation value of the at least one partial area, and using the calculated gradation value of the peripheral area It may be achieved by calculating a tone correction amount.
- the image processing method further calculates a global correction amount of the local gradation correction image, and performs gradation correction for correcting the global gradation on the local gradation correction image based on the global correction amount! Which can be achieved by generating a global complementary image,
- the image processing method may be further achieved by providing the input image and outputting the local tone correction image or the global tone correction image.
- the gradation value is preferably a luminance value or a gradation value expressed by a spectral component.
- the image processing method further applies the gradation correction to the correction target local area while changing the area of the correction target local area or maintaining the area of the correction target local area. On the other hand, it may be achieved by repeating a plurality of times.
- selecting the local region includes selecting a plurality of local regions
- the image processing method may be achieved by performing the gradation correction on the plurality of selected local regions.
- the gradation correction amount is calculated by calculating the gradation value of the peripheral area and a Gaussian weighted weighted average luminance value, or the average luminance value of the peripheral area as the gradation value of the peripheral area. May be achieved.
- the gradation correction amount is calculated by setting one or more feature route lines connecting a local region and feature points, obtaining one or more route feature points for each of the feature route lines, and calculating the route feature points. Or a weighted average luminance value of the route feature point may be calculated as the gradation value of the surrounding area.
- calculating the gradation value of the peripheral area means a product of a ratio of a distance between the local area to be corrected and at least one partial area located in the vicinity thereof and a representative gradation of the partial area. It may be achieved by calculating the gradation of the peripheral region using.
- the representative gradation value may be an average value of all or some of the pixels included in the partial area, or a median value of all or some of the pixels.
- the calculation of the gradation correction amount may be achieved by calculating the gradation correction amount so that the shadow area is corrected more strongly than the highlight area.
- the calculation of the gradation correction amount is achieved by calculating the gradation correction amount of the local region using the luminance of the area around the local region and the luminance of the highlight region of the image. May be.
- the calculation of the gradation correction amount may be achieved by performing the gradation correction using a nonlinear function.
- the global tone correction preferably includes contrast enhancement.
- FIG. 1 is a flowchart showing the operation of a conventional halftone improvement method.
- FIG. 2 is a block diagram showing a configuration of an image processing system according to the first embodiment of the present invention.
- FIG. 3 is a block diagram showing the configuration of the image processing apparatus of the image processing system according to the first embodiment of the present invention.
- FIG. 4 is a flowchart showing the operation of the image processing system according to the first embodiment of the present invention.
- FIG. 5 is a diagram showing an example of characteristic paths in the first embodiment.
- FIG. 6 is a diagram showing a conversion relationship in luminance calculation of a peripheral region in the first example.
- FIG. 7 is a diagram showing an example of local gradation correction in the first embodiment.
- FIG. 8 is a block diagram showing a configuration of an image processing system according to a second embodiment of the present invention.
- FIG. 9 is a flowchart showing the operation of the image processing system in the second embodiment.
- FIG. 10A is a diagram showing a calculation method of a 4-neighbor weighted average luminance value exemplified as a partial region reference type peripheral region luminance calculation method in the second embodiment.
- FIG. 10B is a diagram showing a 4-neighbor weighted average luminance value calculation method exemplified as a partial region reference type peripheral region luminance calculation method in the second embodiment.
- FIG. 10C is a diagram showing a 4-neighbor weighted average luminance value calculation method exemplified as a partial region reference type peripheral region luminance calculation method in the second embodiment.
- FIG. 10D is a diagram showing a 4-neighbor weighted average luminance value calculation method exemplified as a partial region reference type peripheral region luminance calculation method in the second embodiment.
- FIG. 11 is a block diagram showing a configuration of an image processing system according to a third embodiment of the present invention.
- FIG. 12 is a flowchart showing an operation of the image processing system according to the third embodiment.
- FIG. 13 is a diagram for explaining an example of global gradation correction according to the third embodiment.
- FIG. 14 is a block diagram showing a specific configuration of the image processing apparatus of the present invention.
- FIG. 15A is a diagram showing an example of image processing according to the first embodiment.
- FIG. 15B is a diagram illustrating an example of image processing according to the first embodiment.
- FIG. 15C is a diagram illustrating an example of image processing according to the first embodiment.
- FIG. 16A is a diagram illustrating an example of image processing according to the third embodiment.
- FIG. 16B is a diagram showing an example of image processing according to the third embodiment.
- FIG. 2 is a block diagram illustrating the configuration of the image processing system according to the first embodiment.
- the image processing system includes an image input unit 1, an image processing device 2, and an image output unit 3.
- image The input unit 1 is realized by an imaging device such as a camera or a scanner, an image database in which image data is captured and stored, or a network to which they are connected.
- Original image data is input to the image processing device 2 from the image input unit 1.
- the image processing device 2 performs a correction process for improving the intermediate gradation of the input original image, and outputs the corrected image to the image output unit 3.
- the image output unit 3 is realized by a display, a printer, a storage medium such as a hard disk or a memory card that holds image data, or a network to which these are connected, and displays and stores images and transmits them.
- the image processing apparatus 2 includes a local region selection unit 21, a correction amount calculation unit 22, a local gradation correction unit 23, and a control unit 24.
- the local region selection unit 21 selects at least one local region from the original image data input from the image input unit 1 as a correction target.
- the local area selection unit 21 outputs the image data to the correction amount calculation unit 22 and the local gradation correction unit 23.
- the local area may be selected manually by the user or automatically. In the case of automatic selection, the average luminance of the image may be obtained, and a pixel region that is brighter or darker than the average luminance by a predetermined threshold value may be selected as the local region.
- a local area is an image area composed of one or more pixels.
- the local region may be a set of a plurality of pixels. Further, the image may be divided into a plurality of divided areas, and each of the divided areas may be a local area. When a plurality of local regions are present, the local regions may include regions that overlap each other. Data specifying the selected local area is sent to the control unit 24.
- Image statistics include, for example, pixel dispersion in the local area and average luminance.
- the control unit 24 extracts these image statistics, and sets a region that exceeds the threshold value or a value that falls below the threshold value as a local region to be corrected.
- the control unit 24 holds local area designation data to be corrected.
- the correction amount calculation unit 22 calculates a correction amount based on the peripheral area image data.
- the correction amount calculation unit 22 includes a peripheral region luminance calculation unit 221 and a local correction amount conversion unit 222.
- the peripheral region luminance calculation unit 221 calculates the luminance value of the peripheral region of the local region that is the correction target selected by the local region selection unit 21 in accordance with the instruction from the control unit 24.
- Local correction variable The conversion unit 222 obtains the correction amount of the local region based on the luminance value of the peripheral region calculated by the peripheral region luminance calculation unit 221.
- the peripheral area luminance calculation unit 221 determines image data of the peripheral area of the selected local area from the image data based on the local area specification data supplied from the control unit 24. Then, the luminance value of the image data of the determined peripheral area is calculated.
- the local correction amount conversion unit 222 calculates the correction amount of the local region based on the calculated luminance value of the peripheral region.
- a peripheral area is determined around the local area.
- the peripheral region is a portion within a certain distance from the local region.
- the center of the local area is the origin
- the area of M pixels vertically and N pixels horizontally is the peripheral area.
- M and N indicating the size of the region are preferably set based on the size of the image.
- the peripheral area may not be a rectangular area but may be a circular (elliptical) area.
- the peripheral area luminance calculation unit 221 uses the median or average value of the luminance histogram of the peripheral area, the weighted average luminance value obtained by weighted average of the luminance of the peripheral area, and the feature path as the luminance value of the peripheral area.
- a weighted average luminance value of a reference point can be used.
- the calculation method of the weighted average brightness AS of the area centered on the local area and the weighted average brightness FS of the reference point on the feature path will be described.
- the weighted average luminance value AS of the region centered on the local region is calculated as follows. If the input original image is I (x, y) and the weighting factor is isotropic two-dimensional normal distribution H (x, y), the weight of the surrounding area centered on the center of the local area (X, y) Average brightness value AS is expressed by equation (2).
- the reference point is a point that is on the feature path and is included as an element of the weighted average among the points that are characteristic in the surrounding area.
- the peripheral areas areas with higher brightness than the surrounding area, areas with high edge strength, areas with characteristics such as RGB gradation values or luminance maximum and minimum values are set as feature points. .
- the image processing apparatus 2 selects P feature points from among the feature points, and the feature path connecting these P feature points and the local regions. Set the line. This feature path line is set so as to connect P feature points and other unselected feature points between local regions as feature path points.
- the characteristic paths and characteristic points set in this way are as shown in FIG. 5, for example.
- the weighted average luminance FS is calculated using the feature point on the path P1 shown in Fig. 5 as a reference point.
- the load average luminance value FS of the reference point on the feature path is obtained by the following equation (5).
- FS , P n ⁇ ( ⁇ ⁇ , ⁇ , P , i)... ( ⁇ )
- P (> 0) is the number of feature paths
- Np (> 0) is the number of reference points on the feature path ⁇
- a (x, y) is a coefficient.
- a (x, y) for example, the two-dimensional normal distribution H (x, y) shown in Eqs. (3) and (4) is used.
- a (x, y) may be a constant.
- the feature path is an arbitrary link between the feature points given around the local area shown in path P1 in Fig. 5 and the local area. This is a route that connects the feature points on the line.
- the local correction amount conversion unit 222 calculates the correction amount of the local region based on the peripheral region luminance value.
- the local correction amount conversion unit 222 performs linear conversion or non-linear conversion on the luminance value of the peripheral region in the conversion process from the luminance value of the peripheral region to the correction amount of the local region, and the correction amount to the local gradation correction unit 23. Is output.
- the local gradation correction unit 23 performs gamma correction
- the luminance of the surrounding area can be converted into a correction amount as follows. That is, when the luminance value S of the peripheral region is obtained, the local correction amount conversion unit 222 calculates the correction amount Z of the local region based on the luminance value S of the peripheral region by the following equation (6).
- the slope of the correction amount Z is shown.
- b is a predetermined constant (0 ⁇ b ⁇ l. 0) and indicates the minimum value of the correction amount Z.
- c is a predetermined constant (0 ⁇ c ⁇ l. 0), which indicates the constant of the saturation part of the correction amount.
- the index ⁇ is a predetermined constant ( ⁇ 0), and the value is determined based on the image to be corrected. Therefore, if the brightness value S in the surrounding area is less than the brightness value S in the highlight area, the correction amount ⁇
- the correction amount Z is a continuous value before and after the luminance value S of the peripheral region, and the value range of the correction amount Z
- the highlight area is not corrected to be darker or corrected to be excessively bright. Therefore, it is possible to maintain good gradation in the highlight area.
- S indicates the luminance value of the shadow area in the image.
- a is a predetermined constant (> 0)
- b is a predetermined constant (0 ⁇ b ⁇ l. 0).
- c is a predetermined constant (0 ⁇ c ⁇ l. 0).
- the index ⁇ is a predetermined constant ( ⁇ 0), and the value is determined based on the image to be corrected.
- the correction quantity Z is determined by a function with intercept b (0 ⁇ b ⁇ l. 0), constant c (0 ⁇ c ⁇ l. 0), exponent a ⁇ 0).
- the correction amount Z is fixed.
- a highlight area is a set of pixels having gradation values or luminance values that are equal to or higher than a specific luminance.
- a shadow area is a set of pixels having a gradation value or luminance value lower than a specific luminance.
- a histogram of gradation values (luminance values) of an image is generated.
- a pixel having a gradation value equal to or higher than the gradation value of the upper 768th pixel is the highlight area.
- pixels having gradation values equal to or lower than the gradation value of the lower 768th pixel are shadow areas.
- the ratio of the highlight area and the shadow area has been described as the same value, but may be different values.
- the local gradation correction unit 23 selects a local region selected from the image data based on the image data supplied from the local region selection unit 21 and the correction amount supplied from the correction amount calculation unit 22. The tone value of is corrected.
- the corrected image data that has been corrected is output to the image output unit 3.
- the gradation value may be a luminance value or a gradation value expressed by a spectral component.
- the local gradation correction processing can be realized using a nonlinear function such as a linear correction function or a sigmoid function.
- a nonlinear function such as a linear correction function or a sigmoid function.
- Equation 10 Imax is the maximum gradation value
- ⁇ is the gamma value calculated by the correction amount calculation unit 22.
- Equation (10) can also correct only the force / luminance component, which is an equation that processes R, G, and ⁇ independently.
- the gradation value in the RGB space is first divided into a luminance component and a chromaticity component.
- Equation (10) It is corrected as a value.
- the converted luminance value and original chromaticity component are converted back to RGB space gradation values. By processing in this way, it is also possible to correct only the luminance component.
- the processing of the correction amount calculation unit 22 and the local gradation correction unit 23 uses a gradation correction table obtained in advance corresponding to the luminance value S of the peripheral region calculated by the peripheral region luminance calculation unit 221. It can also be realized. The calculation time can be shortened by using the gradation correction table.
- the original image input from the image input unit 1 is supplied to the local region selection unit 21.
- the local region selection unit 21 selects a local region from the input original image.
- the control unit 24 selects a correction target local region from the selected local regions (step S21).
- the correction amount calculation unit 22 calculates the gradation correction amount of the selected local region of the original image (step S22).
- the peripheral region luminance calculation unit 221 first calculates a weighted average luminance value obtained by using a median or average value of the luminance histogram of the peripheral region, a weighted average of the luminance of the peripheral region, The brightness value of the surrounding area is calculated based on the weighted average brightness value of the reference points on the feature path.
- the peripheral correction amount conversion unit 222 the calculated luminance value of the peripheral region is converted into the gradation correction amount of the local region.
- the gradation correction amount of this local area is determined by image quality factors such as local brightness and contrast of the input original image, and the gradation is adjusted so that the image quality of the local area of the image is optimized. It is a value that determines the intensity of gradation correction for correction.
- This local gradation correction is sequentially performed on one or more local areas.
- the control unit 24 determines whether or not the gradation correction of the local area has been performed for all the correction target local areas (step S24). If an unprocessed local area remains (NO in step S24), a correction amount is acquired and gradation correction is performed on the area. All local regions If the processing has been completed (step S24—YES), the gradation correction processing is terminated and the corrected image is supplied to the image output unit 3. The output of the corrected image to the image output unit 3 may be performed every time the correction process of each local region is completed.
- the local gradation correction is performed once for each of the local regions.
- this local gradation correction is performed. Tonal correction may be repeated multiple times.
- gradation correction may be performed while maintaining the area of the local area, but if gradation correction is performed while buying the area of the local area, it is affected by local areas such as singular points. It is possible to prevent the correction from proceeding excessively.
- the control unit 24 may supply the threshold for selecting the local region to the local region selecting unit 21 and perform correction while selecting a new local region to be corrected. Also, you can divide the entire image into multiple areas and use them as local areas for tone correction.
- the system includes a CCD camera as the image input unit 1, a personal computer as the image processing device 2, and a display as the image output unit 3.
- the personal computer includes a central processing unit (CPU: Central Processing Unit) 31, an input unit 32, an output unit 33, a memory unit 34, and an external storage unit 35.
- the central processing unit 31 sequentially retrieves and executes the program stored in the memory unit 34, and performs image tone correction processing based on the image data and work data stored in the memory unit 34 or the external storage unit 35. I do.
- the input unit 32 also captures image data using the CCD camera power.
- the output unit 33 outputs the corrected image data to the display.
- the memory unit 34 stores the captured image data, holds the progress of calculation, and stores a program for calculation processing.
- the external storage unit 35 stores programs to be stored in the memory unit 34 and saves the contents of the memory unit 34.
- the personal computer includes a local area selection unit 21, a correction amount calculation unit 22, and a local gradation correction unit 23 that are realized by software as the image processing apparatus 2.
- Dedicated hardware may be implemented for parts with a large amount of repeated computation.
- the correction amount calculation unit 22 calculates a gamma value in the case of gamma correction, and the local gradation correction unit performs gamma correction using the calculated gamma value.
- the power sigmo If correction is performed using an id function, parameters such as the slope and shift amount of the sigmoid function are calculated.
- the central processing unit 31 selects a local region, for example, one pixel in the input original image, and calculates the luminance value of the peripheral region of the local region as shown in FIG. 15B.
- a local region for example, one pixel in the input original image
- the luminance value of the peripheral region of the local region is indicated in FIG. 15B.
- two forces are indicated as the local region
- the peripheral region is indicated by a circle with a radius R centered on the local region.
- the brightness value of this peripheral area is calculated.
- a correction amount is calculated according to Equation (6) based on the luminance value of the peripheral area, and a gamma value for correcting the local area is obtained.
- the gamma value for correcting the local area is close to the b value in Equation (6) (for example, 0.5).
- the gradation value in the local area is corrected brightly by gamma correction.
- the brightness value of the surrounding area is high, such as the local area in the upper left of Fig. 15B,
- the processing ends by outputting the processing result to the display.
- the personal computer executes the image correction processing program to determine the correction amount according to the image quality of the peripheral area and correct the image quality of the local area.
- the tone correction function of the local area changes according to the brightness of the local area of the image as shown in FIG. . Therefore, a process is realized in which the dark area is bright and the bright area retains the gradation of the original image as it is. As a result, it is possible to improve the image quality of the dark portion while maintaining the image quality of the bright portion in the image.
- the above gamma correction and Equation (8) it is possible to improve the image quality of the bright area while maintaining the image quality of the dark area in the image.
- the configuration is such that the acquisition of the gradation correction amount of the local area from the image and the gradation correction are sequentially processed for each local area. For this reason, the amount of data held in the memory unit 34 and the external storage unit 35 out of the correction amount and various data necessary for obtaining the correction amount It can be made smaller than the data size of the input original image.
- FIG. 8 is a block diagram showing a configuration of an image processing system according to the second embodiment of the present invention.
- the image processing system according to the second embodiment of the present invention is different from the configuration of the first embodiment in that a storage device 5 having a partial region luminance value storage unit 51 and a partial region luminance calculation unit 25 are added. Is different.
- the correction amount calculation unit 22 of the first embodiment is replaced with a correction amount calculation unit 26.
- a partial region reference type peripheral region luminance calculation unit 261 is provided instead of the peripheral region luminance calculation unit 221. Therefore, the different points are explained below.
- the partial area luminance calculation unit 25 divides the input image area into a horizontal m area and a vertical n area, calculates a representative luminance value of each area, and stores the representative luminance value in the partial area luminance value storage unit 51. Store the value.
- the correction amount calculation unit 26 calculates a local region correction amount using the representative luminance value stored in the partial region luminance value storage unit 51. Specifically, the partial region reference type peripheral region luminance calculation unit 261 calculates the peripheral luminance value using the representative luminance value of the partial region located in the vicinity of the local region, and the local correction amount conversion unit 222 calculates the peripheral luminance. The local correction amount is determined using the value.
- steps S21, S23 and S24 in FIG. 9 are the same as steps S21, S23 and S24 in the first embodiment shown in FIG.
- the partial region luminance calculation unit 25 divides the region of the input image into a horizontal m region and a vertical n region (m, n> 0) (FIG. 10A). ), Representative brightness of each area The degree value is obtained (FIG. 10B), and the representative luminance value is stored in the partial area luminance value storage unit 51 (step S25).
- the local area selection unit 21 selects a local area to be corrected in the same manner as in the first embodiment (step S21).
- the partial region reference type peripheral region luminance calculation unit 261 of the correction amount calculation unit 26 calculates the peripheral luminance value of the local region using the representative luminance value calculated in step S25, and the local correction amount conversion unit 222 is selected.
- the correction amount for the local area is calculated (step S26).
- the partial area reference type peripheral area luminance calculating unit 261 calculates, for example, the 4-neighbor weighted average value IS as the luminance value of the peripheral area using the representative luminance value of the partial area.
- the neighborhood weighted average value IS is linearly compensated using the representative luminance value S of the nearest neighbor region of the region of interest.
- Region (1) Calculated from 4 neighboring regions
- peripheral luminance value When the peripheral luminance value is calculated based on the above formula, it is possible to obtain a peripheral luminance value close to the peripheral luminance value obtained from the weighted average luminance value AS as shown in FIG. 10D.
- a storage capacity of about several bytes is required to store the representative luminance value of the partial area. It is not necessary to refer to the calculation to calculate the route, and it is possible to operate at a high speed with fewer points to refer to the calculation compared to other methods of calculating the brightness value of the surrounding area.
- gradation correction is performed in two stages, local gradation correction and global gradation correction.
- the configuration of the image processing system has the same parts as those of the first embodiment shown in FIG. 2, and the configuration of the image processing apparatus 2 is as shown in FIG.
- the image processing apparatus includes a local region selection unit 21, a correction amount calculation unit 22, a local tone correction unit 23, a global tone correction amount calculation unit 27, And a global tone correction unit 28.
- the local area selection unit 21, the correction amount calculation unit 22, and the local gradation correction unit 23 are the same as those in the first embodiment.
- the global tone correction amount calculation unit 27 and the global tone correction unit 28 perform global tone correction processing. Since the local gradation correction process is the same as the process in the first embodiment, the description thereof is omitted here.
- the local tone corrected image from the local tone correcting unit 23 is output to the global tone correction amount calculating unit 27 and the global tone correcting unit 28.
- the global tone correction amount calculation unit 27 calculates a global tone correction amount for improving the global image quality from the image locally corrected by the local tone correction unit 23.
- the global gradation correction unit 28 corrects the gradation of the entire image based on the global gradation correction amount calculated by the global gradation correction amount calculation unit 27.
- Image data that has been subjected to local gradation correction and global gradation correction is output from the global gradation correction unit 28 to the image output unit 3.
- the global tone correction amount calculation unit 27 performs correction processing for obtaining an image quality that is globally optimized for an image whose image quality is locally optimized by the local tone correction unit 23. Determine the correction amount.
- Global correction includes overall image quality such as contrast correction and gamma correction. It is possible to use a gradation correction method for adjusting the image quality.
- the global tone correction amount calculation unit 27 calculates the correction amount of these global image quality correction methods. For example, when contrast correction is used, correction amounts a and b, which are coefficients of a contrast correction conversion formula, are calculated by the following formulas (14) and (15).
- HL-SH Here, HLopt and SHopt indicate the optimum gradation values in the highlight and shadow areas of the image.
- HL and SH are the luminance values of the highlight area and shadow area in the original image.
- the brightness value HL of the highlight area is obtained by applying a low-pass filter LPF (x, y) to the input original image force and the created brightness image I (x, y) as shown in the following equation (16). This is the maximum gradation value in the image when applied.
- the brightness value SH of the shadow area is expressed by the low pass filter LPF (X, y) for the brightness image I (x, y) created from the input original image as shown in the following equation (17). Is the minimum gradation value in the image when.
- SH mm (/ ⁇ , ⁇ ) * LPF (x, ')) (17)
- the operator * represents a convolution operation.
- the function max () and the function min () are functions having the maximum value and the minimum value of the elements indicated in parentheses as function values, respectively.
- the local gradation correction unit 28 then performs local gradation correction.
- the corrected image data is subjected to global gradation correction based on the correction amount calculated by the global gradation correction amount calculation unit 27.
- global correction a gradation correction method for adjusting the image quality of the entire image, such as contrast correction and gamma correction, can be used.
- contrast correction is used, the gradation of each pixel is corrected by the following equation (18) using the correction amounts a and b calculated by the global gradation correction amount calculation unit 25.
- the gradation values of the input original image are distributed up to the minimum gradation value SH power and the maximum gradation value HL.
- a correction process is performed to convert this according to Expression (18) so that it is distributed from the minimum gradation value SHopt to the maximum gradation value HLopt, which is the optimum gradation value for the image.
- the equation (18) can also correct only the force / luminance component, which is an equation for processing each primary color R, G, B independently.
- the gradation value in the RGB space is divided into a luminance component and a chromaticity component. Of these, only the luminance component is corrected as the input value of equation (18). The luminance value after correction and the original chromaticity component are converted back to RGB space gradation values. In this way, it is possible to correct only the luminance component.
- Steps S21 to S24 are the same as the overall operation in the first embodiment.
- the original image input from the image input unit 1 is supplied to the local region selection unit 21.
- the local area selection unit 21 selects a local area to be corrected from the input original image (step S21).
- the correction amount calculation unit 22 calculates the gradation correction amount of the local region of the input original image (step S22).
- the gradation correction amount for the local area is determined by first calculating the median or average value of the luminance histogram of the peripheral area, a weighted average luminance value obtained by weighted average of the luminance of the peripheral area, Based on the weighted average luminance value of the reference points on the collection route, the luminance value of the surrounding area is calculated.
- the calculated luminance value of the peripheral area is converted into the gradation correction amount of the local area.
- This local area gradation correction amount is determined by image quality factors such as local brightness and contrast of the input original image, and gradation correction is performed so that the image quality of the local area of the image is optimized. It is a value that determines the intensity of gradation correction for performing.
- the gradation correction amount of the local region is calculated, the gradation of the corresponding local region is corrected based on the obtained local gradation correction amount (step S23).
- the correction amount S for correcting the local region is calculated as the S gamma value
- This local gradation correction is sequentially performed on one or more local regions. It is determined whether or not the local area tone correction is performed for all local areas to be corrected (step S24). If an unprocessed local area remains (step S24—NO), a correction amount is acquired and gradation correction is performed on the area. If the processing has been completed for all local regions (step S24—YES), the local tone correction processing is terminated, and the image after local correction is supplied to the global tone correction processing. The output of the local correction image to the global tone correction processing may be performed after the correction processing of each local region is completed.
- global gradation correction is performed.
- global tone correction first, the global image quality of an image is determined.
- contrast correction global image quality is usually obtained by generating a luminance histogram of the entire image. That is, as shown in FIG. 13, the minimum luminance SH and the maximum luminance HL are obtained from the luminance histogram of the input image.
- the correction amount a is calculated according to equations (14) and (15) using the optimal brightness value HLopt in the highlight area and the optimal brightness value SHopt in the shadow area. , B are calculated (step S25).
- the global tone correction unit 28 performs global tone correction.
- global tone correction is performed based on equation (18).
- the image that has been subjected to the global tone correction is supplied to the image output unit 3, and the image correction process is completed.
- the gradation correction processing in the local region and the global region is performed.
- the applied system is the same as in the first example.
- the personal computer program of the first embodiment is additionally equipped with a program for correcting the global gradation, that is, a program that becomes the global gradation correction amount calculation unit 27 and the global gradation correction unit 28.
- the storage device 5 and the partial region luminance value storage unit 51 are added to this configuration, and the partial region luminance is calculated between the image input unit 1 and the local region selection unit 21.
- the correction amount calculation unit 22 may be replaced with the correction amount calculation unit 26 by adding the unit 25.
- the contrast of the image is corrected based on the contrast correction formula of Expression (18).
- the result is an image with improved global image quality, as shown in Figure 16B.
- This corrected image has improved visibility in dark areas as compared with the input original image. That is, it can be seen that the local image quality is also improved.
- the image after processing as described above is displayed on the display, and the processing ends.
- the global tone correction after performing the local tone correction, the global image quality of the image whose local image quality is improved is improved, and the image quality of the entire image is impaired.
- the image quality in the local area can be improved without any problem.
- an image processing device an image correction processing method, and an image correction that can be processed with a small storage capacity because they are sequentially processed and do not need to store intermediate results of the processing.
- a processing program can be provided.
- the corrected image since the global gradation correction is performed on the image obtained as a result of the local gradation correction, the corrected image has an appropriate image quality locally and globally.
- An image processing device, an image correction processing method, and an image correction processing program can be provided. wear.
- a computer readable software product that records code for a software product comprising:
- the step of calculating the gradation correction amount includes:
- the image processing method includes:
- the image processing method includes:
- a software product further comprising:
- the image processing method includes:
- the image processing method includes:
- a software product further comprising:
- the step of selecting the local region includes a step of selecting a plurality of local regions
- the image processing method includes:
- a software product further comprising the step of performing the tone correction on the plurality of selected local regions.
- the step of calculating the gradation correction amount includes calculating a gradation value of the peripheral area and a weighted average luminance value of a Gaussian weight or an average luminance value of the peripheral area as the gradation value of the peripheral area.
- a software product comprising:
- the step of calculating the gradation correction amount includes:
- One or more feature route lines connecting local regions and feature points are set, one or more route feature points are obtained for each of the feature route lines, and the tone value of the route feature point or the route feature point is obtained.
- a software product comprising a step of calculating a weighted average luminance value of as the gradation value of the peripheral region.
- the step of calculating the gradation value of the peripheral area includes:
- a software product comprising:
- the representative gradation value is an average value of all or some of the pixels included in the partial area, or a median value of all or some of the pixels.
- the step of calculating the gradation correction amount includes:
- a software product comprising:
- the step of calculating the gradation correction amount includes:
- a software product comprising:
- the step of calculating the gradation correction amount includes:
- a software product comprising:
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Image Processing (AREA)
- Facsimile Image Signal Circuits (AREA)
Abstract
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP05777032A EP1788795A4 (en) | 2004-09-01 | 2005-09-01 | IMAGE CORRECTION DEVICE AND IMAGE CORRECTION |
US11/574,516 US7949201B2 (en) | 2004-09-01 | 2005-09-01 | Image correction processing system and image correction processing method |
JP2006531980A JP4577621B2 (ja) | 2004-09-01 | 2005-09-01 | 画像補正処理システム及び画像補正処理方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2004254901 | 2004-09-01 | ||
JP2004-254901 | 2004-09-01 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2006025486A1 true WO2006025486A1 (ja) | 2006-03-09 |
Family
ID=36000135
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2005/015995 WO2006025486A1 (ja) | 2004-09-01 | 2005-09-01 | 画像補正処理システム及び画像補正処理方法 |
Country Status (5)
Country | Link |
---|---|
US (1) | US7949201B2 (ja) |
EP (1) | EP1788795A4 (ja) |
JP (1) | JP4577621B2 (ja) |
CN (1) | CN101010937A (ja) |
WO (1) | WO2006025486A1 (ja) |
Cited By (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007272477A (ja) * | 2006-03-30 | 2007-10-18 | Denso It Laboratory Inc | 画像補正方法 |
WO2008111180A1 (ja) * | 2007-03-13 | 2008-09-18 | Olympus Corporation | 画像信号処理装置、画像信号処理プログラム、画像信号処理方法 |
WO2010073553A1 (ja) | 2008-12-26 | 2010-07-01 | 日本電気株式会社 | 画像処理装置、画像処理方法および記憶媒体 |
JP2010258502A (ja) * | 2009-04-21 | 2010-11-11 | Canon Inc | 画像処理装置および方法 |
WO2011013745A1 (ja) * | 2009-07-31 | 2011-02-03 | シャープ株式会社 | 階調調整装置、画像表示装置、テレビ受像機、プログラム、及び、プログラムが記録されたコンピュータ読み取り可能な記憶媒体 |
WO2011030383A1 (ja) * | 2009-09-10 | 2011-03-17 | 株式会社 東芝 | 画像処理装置 |
WO2012017946A1 (ja) * | 2010-08-04 | 2012-02-09 | 日本電気株式会社 | 画像処理方法、画像処理装置及び画像処理プログラム |
EP1884892A3 (en) * | 2006-07-31 | 2017-11-29 | Samsung Electronics Co., Ltd. | Method, medium, and system compensating shadow areas |
US10438323B2 (en) | 2014-09-29 | 2019-10-08 | Nec Corporation | Image brightness correction and noise suppression method, device, and recording medium for storing image processing program |
US20210090221A1 (en) * | 2019-09-20 | 2021-03-25 | Realtek Semiconductor Corp. | Image processing circuit and associated image processing method |
JP2022520264A (ja) * | 2019-05-05 | 2022-03-29 | ▲騰▼▲訊▼科技(深▲セン▼)有限公司 | 画像輝度の調整方法及び装置、電子機器及びコンピュータプログラム |
Families Citing this family (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8103119B2 (en) * | 2005-06-20 | 2012-01-24 | Nikon Corporation | Image processing device, image processing method, image processing program product, and image-capturing device |
US7477777B2 (en) * | 2005-10-28 | 2009-01-13 | Aepx Animation, Inc. | Automatic compositing of 3D objects in a still frame or series of frames |
JP4816569B2 (ja) * | 2006-07-06 | 2011-11-16 | セイコーエプソン株式会社 | 画像表示システム |
JP2008118383A (ja) * | 2006-11-02 | 2008-05-22 | Matsushita Electric Ind Co Ltd | デジタルカメラ |
WO2008136629A1 (en) * | 2007-05-03 | 2008-11-13 | Mtekvision Co., Ltd. | Image brightness controlling apparatus and method thereof |
KR100901353B1 (ko) * | 2007-05-25 | 2009-06-05 | 주식회사 코아로직 | 영상 처리 장치 및 그 방법 |
JP4973372B2 (ja) * | 2007-08-06 | 2012-07-11 | 株式会社ニコン | 画像処理装置、撮像装置および画像処理プログラム |
CN100578179C (zh) * | 2007-08-21 | 2010-01-06 | 友达光电(苏州)有限公司 | 测量发光画面亮度均匀性的方法 |
JP4525719B2 (ja) * | 2007-08-31 | 2010-08-18 | カシオ計算機株式会社 | 階調補正装置、階調補正方法、及び、プログラム |
JP5053802B2 (ja) * | 2007-11-05 | 2012-10-24 | オリンパス株式会社 | 信号処理装置及び信号処理プログラム |
CN101431637B (zh) * | 2007-11-06 | 2013-06-19 | 奇美电子股份有限公司 | 帧数据处理的装置及方法 |
EP2299695A4 (en) * | 2008-07-17 | 2012-08-29 | Nikon Corp | IMAGING DEVICE, IMAGE PROCESSING PROGRAM, IMAGE PROCESSING DEVICE, AND IMAGE PROCESSING METHOD |
US8391634B1 (en) * | 2009-04-28 | 2013-03-05 | Google Inc. | Illumination estimation for images |
JP5299867B2 (ja) * | 2009-06-30 | 2013-09-25 | 日立コンシューマエレクトロニクス株式会社 | 画像信号処理装置 |
CN102129678A (zh) * | 2010-01-12 | 2011-07-20 | 鸿富锦精密工业(深圳)有限公司 | 影像特征模型建立系统及方法以及应用其的影像处理系统 |
JP2011228807A (ja) * | 2010-04-15 | 2011-11-10 | Nikon Corp | 画像処理プログラム、画像処理装置、および画像処理方法 |
JP2012019392A (ja) * | 2010-07-08 | 2012-01-26 | Nikon Corp | 画像処理装置、電子カメラおよび画像処理プログラム |
US8798393B2 (en) | 2010-12-01 | 2014-08-05 | Google Inc. | Removing illumination variation from images |
JP5488530B2 (ja) * | 2011-05-23 | 2014-05-14 | 富士ゼロックス株式会社 | 画像処理装置及び画像処理プログラム |
CN102231206B (zh) * | 2011-07-14 | 2012-11-28 | 浙江理工大学 | 适用于汽车辅助驾驶系统的彩色夜视图像亮度增强方法 |
JP2013045316A (ja) * | 2011-08-25 | 2013-03-04 | Sanyo Electric Co Ltd | 画像処理装置及び画像処理方法 |
JP5948203B2 (ja) * | 2011-10-12 | 2016-07-06 | 富士フイルム株式会社 | 内視鏡システム及びその作動方法 |
CN102402918B (zh) * | 2011-12-20 | 2014-07-09 | 深圳Tcl新技术有限公司 | 改善画面画质的方法及lcd显示器 |
US9202281B2 (en) * | 2012-03-17 | 2015-12-01 | Sony Corporation | Integrated interactive segmentation with spatial constraint for digital image analysis |
US8873812B2 (en) * | 2012-08-06 | 2014-10-28 | Xerox Corporation | Image segmentation using hierarchical unsupervised segmentation and hierarchical classifiers |
US11113821B2 (en) * | 2017-12-20 | 2021-09-07 | Duelight Llc | System, method, and computer program for adjusting image contrast using parameterized cumulative distribution functions |
TWI476754B (zh) * | 2013-06-25 | 2015-03-11 | Mstar Semiconductor Inc | 應用於顯示裝置之校正系統及校正方法 |
US8958658B1 (en) * | 2013-09-10 | 2015-02-17 | Apple Inc. | Image tone adjustment using local tone curve computation |
US9251574B2 (en) * | 2013-12-17 | 2016-02-02 | Adobe Systems Incorporated | Image compensation value computation |
CN105760884B (zh) * | 2016-02-22 | 2019-09-10 | 北京小米移动软件有限公司 | 图片类型的识别方法及装置 |
EP3291173A1 (en) | 2016-09-02 | 2018-03-07 | Casio Computer Co., Ltd. | Diagnosis assisting device, image processing method in diagnosis assisting device, and program |
CN107820067B (zh) * | 2017-10-29 | 2019-09-20 | 苏州佳世达光电有限公司 | 多投影画面的拼接方法及拼接装置 |
CN112422941A (zh) * | 2020-11-11 | 2021-02-26 | 北京德为智慧科技有限公司 | 显示器的伽玛调校方法、装置、电子设备及存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH07296160A (ja) * | 1994-04-27 | 1995-11-10 | Matsushita Electron Corp | 画像補正方法 |
JP2000228747A (ja) * | 1998-12-03 | 2000-08-15 | Olympus Optical Co Ltd | 画像処理装置 |
JP2002290707A (ja) * | 2001-03-26 | 2002-10-04 | Olympus Optical Co Ltd | 画像処理装置 |
JP2003209857A (ja) * | 2002-01-11 | 2003-07-25 | Fuji Photo Film Co Ltd | 画像処理装置 |
Family Cites Families (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH05176220A (ja) | 1991-12-25 | 1993-07-13 | Matsushita Electric Ind Co Ltd | 自動露光制御装置 |
EP0613294B1 (en) * | 1993-02-24 | 1998-10-28 | Matsushita Electric Industrial Co., Ltd. | Gradation correction device and image sensing device therewith |
JP3501252B2 (ja) * | 1995-06-16 | 2004-03-02 | 三菱電機株式会社 | 階調補正装置 |
US5991456A (en) | 1996-05-29 | 1999-11-23 | Science And Technology Corporation | Method of improving a digital image |
JP2951909B2 (ja) * | 1997-03-17 | 1999-09-20 | 松下電器産業株式会社 | 撮像装置の階調補正装置及び階調補正方法 |
NL1006371C2 (nl) | 1997-06-20 | 1998-12-22 | Koninkl Kpn Nv | Inrichting en werkwijze voor het testen van de bereikbaarheid van ten minste een gebruiker van een telecommunicatienetwerk. |
JP3576812B2 (ja) | 1998-06-11 | 2004-10-13 | 富士写真フイルム株式会社 | 画像処理装置及び画像処理方法 |
JP3714657B2 (ja) * | 1999-05-12 | 2005-11-09 | パイオニア株式会社 | 階調補正装置 |
US6650774B1 (en) * | 1999-10-01 | 2003-11-18 | Microsoft Corporation | Locally adapted histogram equalization |
JP3902894B2 (ja) | 1999-10-15 | 2007-04-11 | 理想科学工業株式会社 | 画像処理装置及び画像処理方法 |
US6813041B1 (en) | 2000-03-31 | 2004-11-02 | Hewlett-Packard Development Company, L.P. | Method and apparatus for performing local color correction |
US6915024B1 (en) | 2000-09-29 | 2005-07-05 | Hewlett-Packard Development Company, L.P. | Image sharpening by variable contrast mapping |
US7251056B2 (en) * | 2001-06-11 | 2007-07-31 | Ricoh Company, Ltd. | Image processing apparatus, image processing method and information recording medium |
JP3705180B2 (ja) * | 2001-09-27 | 2005-10-12 | セイコーエプソン株式会社 | 画像表示システム、プログラム、情報記憶媒体および画像処理方法 |
JP3900972B2 (ja) * | 2002-03-04 | 2007-04-04 | 三菱電機株式会社 | コントラスト強調方式 |
JP4096613B2 (ja) | 2002-05-10 | 2008-06-04 | 松下電器産業株式会社 | 画像処理方法及び画像処理装置 |
WO2005027041A1 (ja) * | 2003-09-11 | 2005-03-24 | Matsushita Electric Industrial Co., Ltd. | 視覚処理装置、視覚処理方法、視覚処理プログラムおよび半導体装置 |
JP4551836B2 (ja) * | 2005-07-29 | 2010-09-29 | 株式会社東芝 | 映像信号処理装置及び映像信号処理方法 |
-
2005
- 2005-09-01 WO PCT/JP2005/015995 patent/WO2006025486A1/ja active Application Filing
- 2005-09-01 JP JP2006531980A patent/JP4577621B2/ja not_active Expired - Fee Related
- 2005-09-01 EP EP05777032A patent/EP1788795A4/en not_active Withdrawn
- 2005-09-01 CN CNA200580029343XA patent/CN101010937A/zh active Pending
- 2005-09-01 US US11/574,516 patent/US7949201B2/en not_active Expired - Fee Related
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH07296160A (ja) * | 1994-04-27 | 1995-11-10 | Matsushita Electron Corp | 画像補正方法 |
JP2000228747A (ja) * | 1998-12-03 | 2000-08-15 | Olympus Optical Co Ltd | 画像処理装置 |
JP2002290707A (ja) * | 2001-03-26 | 2002-10-04 | Olympus Optical Co Ltd | 画像処理装置 |
JP2003209857A (ja) * | 2002-01-11 | 2003-07-25 | Fuji Photo Film Co Ltd | 画像処理装置 |
Non-Patent Citations (1)
Title |
---|
See also references of EP1788795A4 * |
Cited By (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2007272477A (ja) * | 2006-03-30 | 2007-10-18 | Denso It Laboratory Inc | 画像補正方法 |
EP1884892A3 (en) * | 2006-07-31 | 2017-11-29 | Samsung Electronics Co., Ltd. | Method, medium, and system compensating shadow areas |
US8655098B2 (en) | 2007-03-13 | 2014-02-18 | Olympus Corporation | Image signal processing apparatus and computer-readable recording medium recording image signal processing program |
WO2008111180A1 (ja) * | 2007-03-13 | 2008-09-18 | Olympus Corporation | 画像信号処理装置、画像信号処理プログラム、画像信号処理方法 |
JPWO2008111180A1 (ja) * | 2007-03-13 | 2010-06-24 | オリンパス株式会社 | 画像信号処理装置、画像信号処理プログラム |
JP4850281B2 (ja) * | 2007-03-13 | 2012-01-11 | オリンパス株式会社 | 画像信号処理装置、画像信号処理プログラム |
WO2010073553A1 (ja) | 2008-12-26 | 2010-07-01 | 日本電気株式会社 | 画像処理装置、画像処理方法および記憶媒体 |
US8891865B2 (en) | 2008-12-26 | 2014-11-18 | Nec Corporation | Image processing device, image processing method, and storage medium for performing a gradation correction operation in a color space |
JP2010258502A (ja) * | 2009-04-21 | 2010-11-11 | Canon Inc | 画像処理装置および方法 |
JP5502868B2 (ja) * | 2009-07-31 | 2014-05-28 | シャープ株式会社 | 階調調整装置、画像表示装置、テレビ受像機、プログラム、及び、プログラムが記録されたコンピュータ読み取り可能な記憶媒体 |
WO2011013745A1 (ja) * | 2009-07-31 | 2011-02-03 | シャープ株式会社 | 階調調整装置、画像表示装置、テレビ受像機、プログラム、及び、プログラムが記録されたコンピュータ読み取り可能な記憶媒体 |
WO2011030383A1 (ja) * | 2009-09-10 | 2011-03-17 | 株式会社 東芝 | 画像処理装置 |
WO2012017946A1 (ja) * | 2010-08-04 | 2012-02-09 | 日本電気株式会社 | 画像処理方法、画像処理装置及び画像処理プログラム |
US8831345B2 (en) | 2010-08-04 | 2014-09-09 | Nec Corporation | Image processing method, image processing apparatus, and image processing program |
JP5991486B2 (ja) * | 2010-08-04 | 2016-09-14 | 日本電気株式会社 | 画像処理方法、画像処理装置及び画像処理プログラム |
US10438323B2 (en) | 2014-09-29 | 2019-10-08 | Nec Corporation | Image brightness correction and noise suppression method, device, and recording medium for storing image processing program |
JP2022520264A (ja) * | 2019-05-05 | 2022-03-29 | ▲騰▼▲訊▼科技(深▲セン▼)有限公司 | 画像輝度の調整方法及び装置、電子機器及びコンピュータプログラム |
JP7226893B2 (ja) | 2019-05-05 | 2023-02-21 | ▲騰▼▲訊▼科技(深▲セン▼)有限公司 | 画像輝度の調整方法及び装置、電子機器及びコンピュータプログラム |
US20210090221A1 (en) * | 2019-09-20 | 2021-03-25 | Realtek Semiconductor Corp. | Image processing circuit and associated image processing method |
US11922608B2 (en) * | 2019-09-20 | 2024-03-05 | Realtek Semiconductor Corp. | Image processing circuit and associated image processing method |
Also Published As
Publication number | Publication date |
---|---|
JPWO2006025486A1 (ja) | 2008-05-08 |
US20070248282A1 (en) | 2007-10-25 |
CN101010937A (zh) | 2007-08-01 |
EP1788795A4 (en) | 2010-08-04 |
JP4577621B2 (ja) | 2010-11-10 |
EP1788795A1 (en) | 2007-05-23 |
US7949201B2 (en) | 2011-05-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2006025486A1 (ja) | 画像補正処理システム及び画像補正処理方法 | |
US8411991B2 (en) | Image processing apparatus, image processing method, and program | |
JP4210577B2 (ja) | 選択的空間フィルタを使用するディジタル画像の階調及び空間特性の向上方法 | |
JP4894595B2 (ja) | 画像処理装置および方法、並びに、プログラム | |
US7599578B2 (en) | Apparatus, program, and method for image tone transformation, and electronic camera | |
JP4998287B2 (ja) | 画像処理装置および方法、並びにプログラム | |
US8103119B2 (en) | Image processing device, image processing method, image processing program product, and image-capturing device | |
US20050226526A1 (en) | Image processing device and method | |
EP1111907A2 (en) | A method for enhancing a digital image with noise-dependant control of texture | |
JP4214457B2 (ja) | 画像処理装置および方法、記録媒体、並びにプログラム | |
US20070115369A1 (en) | Tone-conversion device for image, program, electronic camera, and tone-conversion method | |
US8831346B2 (en) | Image processing apparatus and method, and program | |
JP5392560B2 (ja) | 画像処理装置および画像処理方法 | |
EP1111906A2 (en) | A method for enhancing the edge contrast of a digital image independently from the texture | |
JP2006114005A (ja) | 階調変換装置、プログラム、電子カメラ、およびその方法 | |
JP2010244360A (ja) | 画像処理装置、画像処理方法、及びコンピュータプログラム | |
KR100601967B1 (ko) | 영상의 다이나믹 레인지 압축 장치 및 방법 | |
JP5410378B2 (ja) | 映像信号補正装置および映像信号補正プログラム | |
JP2002281312A (ja) | 画像処理のための装置、方法及びプログラム | |
JP5234127B2 (ja) | 階調変換装置、プログラム、電子カメラ、及びその方法 | |
JP6335614B2 (ja) | 画像処理装置、その制御方法、及びプログラム | |
JP2006114006A (ja) | 階調変換装置、プログラム、電子カメラ、およびその方法 | |
JP4353233B2 (ja) | 画像処理プログラムおよび画像処理装置 | |
JP4120928B2 (ja) | 画像処理装置、画像処理方法、およびそのプログラム | |
JPH09275496A (ja) | 画像の輪郭強調処理装置および方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KM KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NG NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SM SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): BW GH GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IS IT LT LU LV MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 2006531980 Country of ref document: JP |
|
REEP | Request for entry into the european phase |
Ref document number: 2005777032 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2005777032 Country of ref document: EP Ref document number: 11574516 Country of ref document: US Ref document number: 200580029343.X Country of ref document: CN |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWP | Wipo information: published in national office |
Ref document number: 2005777032 Country of ref document: EP |
|
WWP | Wipo information: published in national office |
Ref document number: 11574516 Country of ref document: US |