US11017564B2 - Image processing apparatus and non-transitory computer readable medium storing image processing program - Google Patents
Image processing apparatus and non-transitory computer readable medium storing image processing program Download PDFInfo
- Publication number
- US11017564B2 US11017564B2 US16/731,054 US201916731054A US11017564B2 US 11017564 B2 US11017564 B2 US 11017564B2 US 201916731054 A US201916731054 A US 201916731054A US 11017564 B2 US11017564 B2 US 11017564B2
- Authority
- US
- United States
- Prior art keywords
- image
- metal gloss
- gloss level
- reflection image
- diffuse reflection
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/90—Determination of colour characteristics
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/50—Lighting effects
- G06T15/506—Illumination models
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/001—Texturing; Colouring; Generation of texture or colour
-
- G06T11/10—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/60—Editing figures and text; Combining figures or text
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/40—Analysis of texture
- G06T7/41—Analysis of texture based on statistical description of texture
- G06T7/44—Analysis of texture based on statistical description of texture using image operators, e.g. filters, edge density metrics or local histograms
Definitions
- the present invention relates to an image processing apparatus and a non-transitory computer readable medium storing an image processing program.
- BRDF bidirectional reflectance distribution function
- JP2015-049691A describes a method for acquiring a BRDF and estimating a model coefficient to be fitted to the acquired BRDF by a least square method or the like as a method for determining a BRDF model coefficient.
- JP2005-115645A describes a method for emitting light from a large number of angles, capturing an image with a camera from a large number of angles, preparing a conversion table between the luminance of the captured image and the emission angle or the imaging angle, and calculating the luminance of the target position by performing conversion table interpolation processing.
- Non-limiting embodiments of the present disclosure relate to an image processing apparatus and a non-transitory computer readable medium storing an image processing program capable of expressing the texture of an object surface with less image data and calculation than the case of expressing the texture of an object surface using a large number of image data obtained by capturing an image with a camera from a large number of angles.
- aspects of certain non-limiting embodiments of the present disclosure overcome the above disadvantages and/or other disadvantages not described above.
- aspects of the non-limiting embodiments are not required to overcome the disadvantages described above, and aspects of the non-limiting embodiments of the present disclosure may not overcome any of the disadvantages described above.
- an image processing apparatus including an acquisition unit that acquires a diffuse reflection image of an object surface and a specular reflection image of the object surface; and a reproduction unit that reproduces the object surface such that illumination color is darker as saturation of the diffuse reflection image or a value representing the saturation is higher than of the specular reflection image, by using the diffuse reflection image and the specular reflection image.
- FIG. 1 is a diagram illustrating a schematic configuration of a texture reading device
- FIG. 2 is a block diagram illustrating a configuration of an image processing apparatus according to a first exemplary embodiment
- FIG. 3 is a diagram for explaining an example of determination of a metal gloss level by a metal gloss level determination unit in the first exemplary embodiment
- FIG. 4 is a flowchart illustrating an example of a flow of a process performed by the image processing apparatus according to the first exemplary embodiment
- FIG. 5 is a block diagram illustrating a configuration of an image processing apparatus according to a second exemplary embodiment
- FIG. 6 is a diagram for explaining an example of determination of a metal gloss level by the metal gloss level determination unit according to the second exemplary embodiment
- FIG. 7 is a flowchart illustrating an example of a flow of a process performed by the image processing apparatus according to the second exemplary embodiment
- FIG. 8 is a block diagram illustrating a configuration of a display device according to a third exemplary embodiment
- FIG. 9 is a flowchart illustrating an example of a flow of a process performed by the display device according to the third exemplary embodiment.
- FIG. 10 is a block diagram illustrating a configuration of an image forming apparatus according to a fourth exemplary embodiment.
- FIG. 11 is a flowchart illustrating an example of a flow of a process performed by the image forming apparatus according to the fourth exemplary embodiment.
- FIG. 1 is a diagram illustrating a schematic configuration of a texture reading device.
- the texture reading device 10 optically reads the surface characteristics of the object 12 and generates image information representing the reading result.
- the object 12 is described as a planar object in the present exemplary embodiment, it is not limited to a plane.
- the image information generated by the texture reading device 10 includes image information based on diffuse reflection light and image information based on specular reflection light.
- the texture reading device 10 includes a platen glass 14 , a carriage 16 , light sources 18 , 20 , and 22 , an imaging optical system 24 , and a sensor 26 .
- the texture reading device 10 may read a three-dimensional object.
- the components shown in FIG. 1 are disposed with a predetermined width in a direction perpendicular to the paper surface. This direction is the main-scanning direction of the texture reading device 10 , and the direction indicated by the arrow in FIG. 1 is the sub-scanning direction of the texture reading device 10 .
- the platen glass 14 is composed of a transparent glass plate that supports the object 12 to be read.
- the platen glass 14 is not limited to a glass plate, but may be an acrylic plate, for example.
- a platen cover which covers the platen glass 14 so as to block external light and sandwiches the object 12 may be provided.
- the carriage 16 moves in the sub-scanning direction at a predetermined speed in a case of reading the object 12 .
- the carriage 16 includes light sources 18 , 20 , and 22 inside.
- the light source 18 emits light for reading diffuse reflection light from the object 12 by irradiating the object 12 with light at an incident angle of 45° from the front side in the moving direction of the carriage 16 with respect to the normal direction of the object 12 .
- the light source 20 emits light for reading diffuse reflection light from the object 12 by irradiating the object 12 with light at an incident angle of 45° from the rear side in the moving direction of the carriage 16 with respect to the normal direction of the object 12 .
- the light source 22 emits light for reading specular reflection light from the object 12 by irradiating the object 12 with light at an incident angle of 10° with respect to the normal direction of the object 12 .
- the light source 22 is provided at a position that does not block the principal ray of the reflected light.
- the incident angle of light emitted from the light source 22 is 10° in the exemplary embodiment, but is not limited to 10°, and may be, for example, about 5° to 10°.
- the reflection light of the light emitted from the light source 22 which travels in the normal direction of the object 12 is read.
- the light source 22 has a narrow angle of light to be emitted. In a case where the angle of light emitted from the light source 22 is relatively large, a cover or the like for limiting the angle of light emitted from the light source 22 may be provided. Further, since the light source 22 is for reading gloss information of the object 12 , for example, it is desirable that the luminance in the main-scanning direction be as uniform and continuous as possible, compared with the light sources 18 , 20 .
- the light source 22 white light such as a fluorescent lamp or a rare gas fluorescent lamp (such as a xenon fluorescent lamp) is applied. Further, the light source 22 may be formed in which a plurality of white LEDs are arranged in the main-scanning direction and the luminance distribution in the main-scanning direction is made uniform using a diffusion plate or the like.
- the carriage 16 further includes an imaging optical system 24 and a sensor 26 inside.
- the imaging optical system 24 includes a reflection mirror and an imaging lens, and causes the sensor 26 to form an image of diffuse reflection light and specular reflection light from the object 12 .
- the sensor 26 receives the diffuse reflection light and the specular reflection light imaged by the imaging optical system 24 and generates an image signal corresponding to the received light.
- the sensor 26 is composed of a light receiving element such as a charge coupled device (CCD) linear image sensor or a complementary metal oxide semiconductor (CMOS) image sensor, and converts the received light into a signal representing the intensity thereof.
- the sensor 26 includes a color filter, and generates image information representing the color of the object 12 .
- the sensor 26 outputs diffuse reflection image information obtained by receiving diffuse reflection light and specular reflection image information obtained by receiving specular reflection light to an external device or the like.
- a normal image reading device is configured to read diffuse reflection light from the object 12 by emitting light from the light source 18 or 20 at an incident angle of 45° with respect to the normal direction of the object 12 .
- the texture reading device 10 of the present exemplary embodiment is configured to read specular reflection light from the object 12 by emitting light from the light source 22 at an incident angle of 10° with respect to the normal direction of the object 12 .
- the incident angle and the light receiving angle are constant in each pixel on a two-dimensional plane. Therefore, by calculating the difference between an image acquired at a light source incident angle of 45°, which is a diffuse reflection condition, and an image acquired at a light source incident angle of 10°, which is a specular reflection condition, the metal gloss information is accurately extracted. That is, specular reflectance of each of an area with metal gloss and the two-dimensional plane is acquired at a time by simple difference calculation. Note that for the diffuse reflection condition and the specular reflection condition, calibration is performed with the identical white calibration plate, so gloss information is extracted by a simple difference calculation.
- FIG. 2 is a block diagram illustrating a configuration of the image processing apparatus according to the present exemplary embodiment.
- An image processing apparatus 30 acquires and processes the diffuse reflection image and the specular reflection image obtained by the texture reading device 10 shown in FIG. 1 , and outputs information representing the metal gloss level of the surface of the object 12 .
- the diffuse reflection image is subtracted from the specular reflection image as the difference calculation.
- the difference value of red (R), green (G), and blue (B) or the difference value of L* (lightness) is positive (greater than 0)
- the block diagrams and flowcharts described below are based on the precondition that the object surface is determined to be glossy.
- the gloss of the object surface includes a metal gloss in which the specular reflection is colored and a non-metal gloss in which the specular reflection is not colored.
- the observed glossiness differs between metal gloss and non-metal gloss such as resin other than metal.
- metal in a case where illumination light is emitted, light in which the color of the illumination light is changed according to the color of the metal or the color of the layer including the metal present on the metal surface is observed.
- the color of the illumination light is observed in the case of a non-metal.
- the metal gloss level represents how much the metal gloss is, and the image processing apparatus 30 according to the present exemplary embodiment derives the metal gloss level.
- the image processing apparatus 30 includes functions of a diffuse reflection image acquisition unit 32 , a specular reflection image acquisition unit 34 , an RGB ⁇ L* C* H* conversion unit 36 , a saturation difference image acquisition unit 38 , and a metal gloss level determination unit 40 .
- the image processing apparatus 30 includes a computer in which a central processing unit (CPU), a read only memory (ROM), a random access memory (RAM), and the like are connected through a bus so as to communicate with each other.
- the CPU is a central processing unit, and executes various programs or controls each unit. That is, the CPU reads a program from the ROM, and executes the program using the RAM as a work area.
- the CPU controls the above-described components and performs various arithmetic processes in accordance with the program stored in the ROM.
- the ROM stores a program for executing the above-described respective functions.
- the diffuse reflection image acquisition unit 32 and the specular reflection image acquisition unit 34 correspond to the acquisition unit.
- the RGB ⁇ L* C* H* conversion unit 36 , the saturation difference image acquisition unit 38 , and the metal gloss level determination unit 40 correspond to the output unit.
- the diffuse reflection image acquisition unit 32 and the specular reflection image acquisition unit 34 acquire the diffuse reflection image and the specular reflection image obtained by the texture reading device 10 , respectively.
- the diffuse reflection image acquisition unit 32 and the specular reflection image acquisition unit 34 may be connected to the texture reading device 10 , respectively, and acquire these images from the texture reading device 10 . Alternatively, these images may be acquired through a server or the like connected to the texture reading device 10 through a network.
- the RGB ⁇ L* C* H* conversion unit 36 converts the diffuse reflection image acquired by the diffuse reflection image acquisition unit 32 and the specular reflection image acquired by the specular reflection image acquisition unit 34 from image data of the color components of red (R), green (G), and blue (B) into image data of the color components of L* (lightness), C* (saturation), and H* (hue angle). Note that the conversion method varies by applying a known color conversion technique, and thus detailed description thereof is omitted.
- the saturation difference image acquisition unit 38 acquires a saturation difference image, by using the image data of each of the diffuse reflection image and the specular reflection image converted by the RGB ⁇ L* C* H* conversion unit 36 . That is, a saturation difference image is acquired by obtaining a difference image of the C* color component.
- a saturation difference image there are a case where a diffuse reflection image is subtracted from a specular reflection image and a case where a specular reflection image is subtracted from a diffuse reflection image, and either one may be used, but in the present exemplary embodiment, an example in which a diffuse reflection image is subtracted from a specular reflection image will be described.
- the metal gloss level determination unit 40 determines the metal gloss level using the saturation difference image acquired by the saturation difference image acquisition unit 38 and outputs the determination result.
- the magnitude of the saturation difference is output as object surface information representing the metal gloss level. Further, the metal gloss level determination unit 40 determines as metal gloss in a case where the saturation difference is positive (in a case where the specular reflection image is larger than the diffuse reflection image), and determines as non-metal gloss in a case where the saturation difference is 0, or is negative.
- FIG. 3 is a diagram for explaining an example of determination of a metal gloss level by the metal gloss level determination unit 40 according to the present exemplary embodiment.
- three types A to C as metal gloss levels are compared and shown. Note that the diffuse reflection images and specular reflection images of A to C are magenta color images.
- the metal gloss level determination unit 40 determines the metal gloss level using the magnitude of the saturation difference as the metal gloss level.
- the metal gloss level determination unit 40 may determine whether or not there is metal gloss level by binarizing the saturation difference image with any set value. For example, in the following description, as an example, in a case where the saturation difference (specular reflection image-diffuse reflection image) in FIG. 3 is positive, it is determined that there is metal gloss, and in a case of negative, it is determined that there is no metal gloss. In a case where the saturation difference (diffuse reflection image-specular reflection image) is negative, it is determined that there is metal gloss, and in a case of positive, it is determined that there is no metal gloss.
- FIG. 4 is a flowchart illustrating an example of a flow of a process performed by the image processing apparatus 30 according to the present exemplary embodiment.
- step S 100 the diffuse reflection image acquisition unit 32 acquires a diffuse reflection image, and the process proceeds to step S 102 .
- step S 102 the specular reflection image acquisition unit 34 acquires a specular reflection image, and the process proceeds to step S 104 .
- step S 104 the RGB ⁇ L* C* H* conversion unit 36 converts each of the diffuse reflection image and the specular reflection image from the RGB image into image data of the color components of L* C* H*, and the process proceeds to step S 106 .
- step S 106 the saturation difference image acquisition unit 38 acquires a saturation difference image between the diffuse reflection image and the specular reflection image, and the process proceeds to step S 108 .
- a description will be given as an example of acquiring a saturation difference image obtained by subtracting a diffuse reflection image from a specular reflection image.
- step S 108 the metal gloss level determination unit 40 determines whether or not there is metal gloss.
- step S 110 the metal gloss level determination unit 40 outputs the presence of metal gloss as a determination result.
- step S 112 the metal gloss level determination unit 40 outputs the absence of metal gloss as a determination result.
- FIG. 5 is a block diagram illustrating a configuration of the image processing apparatus according to the present exemplary embodiment.
- the identical components to the components in FIG. 2 may be denoted by the identical reference numerals and detailed description thereof may be omitted.
- the image processing apparatus 31 has functions of a diffuse reflection image acquisition unit 32 , a specular reflection image acquisition unit 34 , a maximum difference color acquisition unit 35 , a difference image acquisition unit 39 of a maximum difference color, and a metal gloss level determination unit 40 .
- the image processing apparatus 31 includes a computer in which a central processing unit (CPU), a read only memory (ROM), a random access memory (RAM), and the like are connected through a bus so as to communicate with each other.
- the CPU is a central processing unit, and executes various programs or controls each unit. That is, the CPU reads a program from the ROM, and executes the program using the RAM as a work area.
- the CPU controls the above-described components and performs various arithmetic processes in accordance with the program stored in the ROM.
- the ROM stores a program for executing the above-described respective functions.
- the maximum difference color acquisition unit 35 the difference image acquisition unit 39 of a maximum difference color, and the metal gloss level determination unit 40 correspond to an output unit.
- the diffuse reflection image acquisition unit 32 and the specular reflection image acquisition unit 34 acquire the diffuse reflection image and the specular reflection image obtained by the texture reading device 10 , respectively.
- the diffuse reflection image acquisition unit 32 and the specular reflection image acquisition unit 34 may be connected to the texture reading device 10 , respectively, and acquire these images from the texture reading device 10 . Alternatively, these images may be acquired through a server or the like connected to the texture reading device 10 through a network.
- the maximum difference color acquisition unit 35 extracts, as the maximum difference color, the maximum color having the maximum value and the minimum color having the minimum value from among the RGB colors, in the diffuse reflection image acquired by the diffuse reflection image acquisition unit 32 , and outputs notification of the extraction result to the difference image acquisition unit 39 of the maximum difference color.
- the difference image acquisition unit 39 of a maximum difference color acquires a difference image for maximum difference color of each of the diffuse reflection image and the specular reflection image, based on the acquisition result of the maximum difference color acquisition unit 35 . Since the saturation is determined by a value (difference) obtained by subtracting the minimum color value from the maximum color value of RGB, the difference image of the maximum difference color of each of the diffuse reflection image and the specular reflection image is equivalent to a saturation difference image in the first exemplary embodiment.
- the metal gloss level determination unit 40 determines the metal gloss level.
- the metal gloss level determination unit 40 determines the metal gloss level using the difference image of a maximum difference color acquired by the difference image acquisition unit 39 of a maximum difference color, and outputs the determination result.
- the difference between the maximum difference colors for example, the R value-G value in FIG. 6 ) is output as information representing the metal gloss level.
- FIG. 6 is a diagram for explaining an example of determination of a metal gloss level by the metal gloss level determination unit 40 according to the present exemplary embodiment.
- FIG. 6 as in FIG. 3 , three types A to C as metal gloss levels are compared and shown. Note that the diffuse reflection images and specular reflection images of A to C are magenta color images.
- the maximum difference color acquisition unit 35 acquires the R color as the maximum color, and the G color as the minimum color, from the RGB image of the diffuse reflection image, as being surrounded by dotted lines.
- the difference image acquisition unit 39 of a maximum difference color acquires a difference image between the diffuse reflection image and the specular reflection image, for the R color and the G color as the maximum difference colors, as being surrounded by one-dot chain lines.
- the example of FIG. 6 shows an example obtained by subtracting the diffuse reflection image from the specular reflection image for the maximum difference color.
- the metal gloss level determination unit 40 subtracts the G color from the R color of the difference image corresponding to the saturation, and outputs the difference as information representing the metal gloss level as in the first exemplary embodiment.
- FIG. 7 is a flowchart illustrating an example of a flow of a process performed by the image processing apparatus 31 according to the present exemplary embodiment.
- the identical processes to the processes in the first exemplary embodiment will be described with the identical reference numerals.
- step S 100 the diffuse reflection image acquisition unit 32 acquires a diffuse reflection image, and the process proceeds to step S 102 .
- step S 102 the specular reflection image acquisition unit 34 acquires a specular reflection image, and the process proceeds to step S 103 .
- step S 103 the maximum difference color acquisition unit 35 acquires the maximum color and the minimum color among the RGB colors as the maximum difference color, from the diffuse reflection image acquired by the diffuse reflection image acquisition unit 32 , and the process proceeds to step S 105 .
- step S 105 the difference image acquisition unit 39 of the maximum difference color acquires a difference image for the maximum difference color acquired by the difference image acquisition unit 39 of the maximum difference color of each of the diffuse reflection image and the specular reflection image, and the process proceeds to step S 107 .
- a difference image obtained by subtracting a diffuse reflection image from a specular reflection image for the maximum difference color is acquired.
- step S 107 the metal gloss level determination unit 40 determines whether or not there is metal gloss. For example, in this determination, since the difference image of the maximum difference color corresponds to the saturation, it is determined whether or not there is metal gloss by the identical method to the method in the first exemplary embodiment. In a case where the determination is positive, the process proceeds to step S 110 , and in a case where the determination is negative, the process proceeds to step S 112 .
- step S 110 the metal gloss level determination unit 40 outputs the presence of metal gloss as a determination result.
- step S 112 the metal gloss level determination unit 40 outputs the absence of metal gloss as a determination result.
- FIG. 8 is a block diagram illustrating a configuration of a display device according to the present exemplary embodiment. Note that, since a display device 50 according to the present exemplary embodiment includes the image processing apparatus 30 according to the first exemplary embodiment, the identical components to the components in FIG. 2 may be denoted by the identical reference numerals and detailed description thereof may be omitted.
- the display device 50 includes the image processing apparatus 30 according to the first exemplary embodiment. That is, the display device 50 has functions of a diffuse reflection image acquisition unit 32 , a specular reflection image acquisition unit 34 , an RGB ⁇ L* C* H* conversion unit 36 , a saturation difference image acquisition unit 38 , and a metal gloss level determination unit 40 .
- the display device 50 has functions of a difference image acquisition unit 42 as an image acquisition unit, a diffuse reflectance distribution function calculation unit 44 , a specular reflectance distribution function calculation unit 46 , a reflectance distribution function calculation unit 52 , a parameter adjustment unit 54 , a light source information acquisition unit 56 , a camera information acquisition unit 58 , a rendering unit 60 , and a display unit 62 .
- the rendering unit 60 and the display unit 62 correspond to a reproduction unit.
- difference image acquisition unit 42 calculates at least one of these difference images.
- the diffuse reflectance distribution function calculation unit 44 calculates the diffuse reflectance distribution function of the object 12 by using the diffuse reflection image. For example, the diffuse reflectance distribution function calculation unit 44 calculates the diffuse reflectance ⁇ d as a parameter from the diffuse reflection image, with ⁇ d as the diffuse reflectance for incident light, ⁇ i as the incident angle, and the diffuse reflectance distribution function as ⁇ d ⁇ cos ⁇ i, according to the Lambert reflection model.
- the specular reflectance distribution function calculation unit 46 calculates the specular reflectance distribution function of the object 12 using the RGB difference image in a case where there is metal gloss, and calculates the specular reflectance distribution function of the object 12 using the gray scale difference image in a case where there is no metal gloss.
- the specular reflectance distribution function calculation unit 46 calculates the specular reflectance ⁇ s, n as parameters from the difference image, with ⁇ s as a specular reflectance, ⁇ as the angle formed by the direction of specular reflection and the line-of-sight direction, n as the specular reflection index, and ⁇ s ⁇ cos n ⁇ as the specular reflectance distribution function, according to the Phong reflection model.
- two difference images are acquired by the difference image acquisition unit 42 .
- the specular reflectance distribution function calculation unit defines a specular reflectance distribution function as ⁇ s 1 ⁇ cos n 1 ⁇ for the difference image (specular reflection image-diffuse reflection image), and defines a specular reflectance distribution function as ⁇ s 2 ⁇ cos n 2 ⁇ for the difference image (diffuse reflection image-specular reflection image), and calculates ⁇ s 1 , ⁇ s 2 , n 1 , and n 2 as parameters, from the respective difference images.
- the reflectance distribution function calculation unit 52 calculates the reflectance distribution function for each pixel of the object 12 by using the diffuse reflectance distribution function calculated by the diffuse reflectance distribution function calculation unit 44 and the specular reflectance distribution function calculated by the specular reflectance distribution function calculation unit 46 .
- the parameter adjustment unit 54 sets various parameters, the light source information acquisition unit 56 acquires light source information such as the light source direction, and the camera information acquisition unit 58 acquires camera information such as the line-of-sight direction.
- the parameter adjustment unit 54 sets parameters of the reflection model used when the specular reflectance distribution function calculation unit 46 calculates the specular reflectance from the difference image and the specular reflection weight coefficient.
- the parameter adjustment unit 54 sets each of parameters such as the diffuse reflection weight coefficient wd, the specular reflection weight coefficient ws, and the specular reflection index n to a predetermined fixed value, used in the reflection model, according to the reflection characteristics of the target object and the output characteristics of the display device.
- the rendering unit 60 renders a three-dimensional model on a virtual screen set in a virtual three-dimensional space, based on the reflectance distribution function calculated by the reflectance distribution function calculation unit 52 , the various parameters set by the parameter adjustment unit 54 , the light source information (light source direction) acquired by the light source information acquisition unit 56 , and the camera information (line-of-sight direction) acquired by camera information acquisition unit 58 .
- the rendering process is known, and for example, the rendering process may be performed using a radiosity method or a ray tracing method considering mutual reflection.
- the display unit 62 reproduces and displays the texture of the object 12 as CG by using image data obtained by the rendering process.
- the object surface is reproduced and displayed such that the color of the illumination is darker as the saturation of the diffuse reflection image is higher than of the specular reflection image.
- the second exemplary embodiment may be used to make a determination by the metal gloss level determination unit 40 to reproduce and display the texture of the object 12 as CG.
- the object surface is reproduced and displayed such that the color of the illumination is darker as the value representing the saturation of the diffuse reflection image is higher than of the specular reflection image.
- FIG. 9 is a flowchart illustrating an example of a flow of a process performed by the display device 50 according to the present exemplary embodiment.
- the identical processes to the processes performed in the image processing apparatus 30 according to the first exemplary embodiment will be described with the identical reference numerals.
- step S 100 the diffuse reflection image acquisition unit 32 acquires a diffuse reflection image, and the process proceeds to step S 102 .
- step S 102 the specular reflection image acquisition unit 34 acquires a specular reflection image, and the process proceeds to step S 104 .
- step S 104 the RGB ⁇ L* C* H* conversion unit 36 converts each of the diffuse reflection image and the specular reflection image from the RGB image into image data of the color components of L* C* H*, and the process proceeds to step S 106 .
- step S 106 the saturation difference image acquisition unit 38 acquires a saturation difference image between the diffuse reflection image and the specular reflection image, and the process proceeds to step S 108 .
- a description will be given as an example of acquiring a saturation difference image obtained by subtracting a diffuse reflection image from a specular reflection image.
- step S 108 the metal gloss level determination unit 40 determines whether or not there is metal gloss.
- step S 109 the difference image acquisition unit 42 acquires a difference image having color components and the process proceeds to step S 114 . That is, a difference image is acquired by calculating a difference between a diffuse reflection image and a specular reflection image in an RGB image having color components.
- step S 111 the difference image acquisition unit 42 acquires a difference image having no color component and the process proceeds to step S 114 . That is, the difference image acquisition unit 42 acquires the difference image by converting the RGB image of each of the diffuse reflection image and the specular reflection image into a gray scale, and calculating a difference between the diffuse reflection image and the specular reflection image in the gray scale having no color component.
- step S 114 the diffuse reflectance distribution function calculation unit 44 calculates the diffuse reflectance distribution function of the object 12 by using the diffuse reflection image, and the process proceeds to step S 116 .
- the diffuse reflectance is calculated from the diffuse reflection image and the diffuse reflection weight coefficient.
- the parameter adjustment unit 54 sets a fixed value for the diffusion weight coefficient in advance according to the reflection characteristics of the target object and display device output characteristics.
- step S 116 the specular reflectance distribution function calculation unit 46 calculates the specular reflectance distribution function of the object 12 by using the difference image acquired by the difference image acquisition unit 42 , and the process proceeds to step S 118 .
- the specular reflectance is calculated from the difference image and the specular reflection weight coefficient.
- the parameter adjustment unit 54 sets a fixed value for the specular reflection weight coefficient in advance according to the reflection characteristics of the target object and display device output characteristics.
- the specular reflectance distribution function of the object 12 is calculated using the RGB difference image when the metal gloss is present, and in a case where the metal gloss is absent, the specular reflectance distribution function of the object is calculated using the gray scale difference image.
- step S 118 the reflectance distribution function calculation unit 52 calculates the reflectance distribution function for each pixel of the object 12 by using the diffuse reflectance distribution function calculated in step S 114 and the specular reflectance distribution function calculated in step S 116 , and the process proceeds to step S 120 .
- n specular reflection index
- the parameter adjustment unit 54 sets a fixed value for the specular reflection index in advance according to the reflection characteristics of the target object and display device output characteristics. Further, the difference image is (specular reflection image-diffuse reflection image), and the number of pixels having a negative difference value is 0.
- the reflection light intensity I(x, y) is calculated separately for each of the R, G, and B components of the diffuse reflection image and the specular reflection image. Thereby, the reflected color of the object 12 is calculated.
- step S 120 the display unit 62 displays an image in which the texture of the object surface is reproduced, and the series of processes end. That is, the rendering unit 60 renders a three-dimensional model on a virtual screen set in a virtual three-dimensional space, based on the reflectance distribution function calculated in step S 116 , the various parameters set by the parameter adjustment unit 54 , the light source information (light source direction) acquired by the light source information acquisition unit 56 , and the camera information (line-of-sight direction) acquired by camera information acquisition unit 58 . Then, the display unit 62 reproduces and displays the texture of the object 12 as CG by using image data obtained by the rendering process.
- the rendering unit 60 renders a three-dimensional model on a virtual screen set in a virtual three-dimensional space, based on the reflectance distribution function calculated in step S 116 , the various parameters set by the parameter adjustment unit 54 , the light source information (light source direction) acquired by the light source information acquisition unit 56 , and the camera information (line-of-sight direction) acquired by
- the diffuse reflectance distribution function is calculated from the diffuse reflection image, and the specular reflectance distribution function is calculated from the difference image.
- the difference image only the glossy part of the object is accurately extracted, so the specular reflectance distribution function for each pixel on the two-dimensional plane is calculated with high accuracy.
- the texture of the object surface is displayed by a simple calculation using a small amount of image data without acquiring a bidirectional reflectance distribution function (BRDF).
- BRDF bidirectional reflectance distribution function
- the presence or absence of metal gloss is determined from the saturation difference image, and the image is displayed by calculating the reflectance distribution function corresponding to the presence or absence of metal gloss, in a case where the texture of the object surface has metal gloss, the metal gloss is reproduced, and in a case where there is no metal gloss, the non-metal gloss is reproduced.
- FIG. 10 is a block diagram illustrating a configuration of the image forming apparatus according to the present exemplary embodiment. Note that, since an image forming apparatus 70 according to the present exemplary embodiment includes the image processing apparatus 30 according to the first exemplary embodiment, the identical components to the components in FIGS. 2 and 8 may be denoted by the identical reference numerals and detailed description thereof may be omitted.
- the image forming apparatus 70 includes the image processing apparatus 30 according to the first exemplary embodiment, and also has functions of a diffuse reflection image acquisition unit 32 , a specular reflection image acquisition unit 34 , an RGB ⁇ L* C* H* conversion unit 36 , a saturation difference image acquisition unit 38 , and a metal gloss level determination unit 40 .
- the image forming apparatus 70 has functions of a difference image acquisition unit 42 as an image acquisition unit, a data conversion unit 64 as a conversion unit, and an image forming unit 66 .
- the data conversion unit 64 and the image forming unit 66 correspond to a reproduction unit.
- the difference image acquisition unit 42 acquires the difference image by calculating the difference between the diffuse reflection image and the specular reflection image in the RGB image having the color components. In a case where the metal gloss level determination unit 40 determines that there is no metal gloss, the difference image acquisition unit 42 acquires the difference image by converting the RGB image of each of the diffuse reflection image and the specular reflection image into a gray scale, and calculating a difference between the diffuse reflection image and the specular reflection image in the gray scale having no color component. Note that there are two difference images: (specular reflection image-diffuse reflection image) and (diffuse reflection image-specular reflection image), and the difference image acquisition unit 42 calculates at least one of these difference images.
- the data conversion unit 64 performs conversion from the diffuse reflection image (R, G, B) and the RGB difference image ( ⁇ R, ⁇ G, ⁇ B) with color components into recording material data including metal components (for example, cyan (C), magenta (M), yellow (Y), black (K), and silver recording material (Si)). Further, in a case where the metal gloss level determination unit 40 determines that there is no metal gloss, conversion from the diffuse reflection image (R, G, B) and the gray scale difference image ( ⁇ Gray) without color components into recording material data including no metal components (for example, C, M, Y, K, Ct (clear recording material)) is performed.
- the data conversion unit 64 performs data conversion using the glossy texture reproduction profile.
- the glossy texture reproduction profile is generated by obtaining the correspondence between (C, M, Y, K, Si) color space of recording material data and (R, G, B, ⁇ R, ⁇ G, ⁇ B) or (R, G, B, ⁇ Gray) color space of texture information.
- the correspondence is generated using the technique described in Japanese Patent Application No. 2018-189223 proposed in the present application.
- the silver recording material may be replaced with a metallic film or metallic paper instead of metallic ink or metallic toner. Further, RGB may be converted into L* a* b*, and ⁇ Gray may be converted into Gloss.
- the image forming unit 66 forms an image on a recording medium, using the recording material data converted by the data conversion unit 64 .
- an image is formed in which the texture of the object surface is reproduced such that the color of the illumination is darker as the saturation of the diffuse reflection image is higher than of the specular reflection image.
- an image in which the texture of the object 12 is reproduced may be formed by the metal gloss level determination unit 40 performing the determination by using the second exemplary embodiment. In this case, an image is formed in which the object surface is reproduced such that the color of the illumination is darker as the value representing the saturation of the diffuse reflection image is higher than of the specular reflection image.
- FIG. 11 is a flowchart illustrating an example of a flow of a process performed by the image forming apparatus 70 according to the present exemplary embodiment.
- the identical processes to the processes performed in the image processing apparatus 30 according to the first exemplary embodiment will be described with the identical reference numerals.
- step S 100 the diffuse reflection image acquisition unit 32 acquires a diffuse reflection image, and the process proceeds to step S 102 .
- step S 102 the specular reflection image acquisition unit 34 acquires a specular reflection image, and the process proceeds to step S 104 .
- step S 104 the RGB ⁇ L* C* H* conversion unit 36 converts each of the diffuse reflection image and the specular reflection image from the RGB image into image data of the color components of L* C* H*, and the process proceeds to step S 106 .
- step S 106 the saturation difference image acquisition unit 38 acquires a saturation difference image between the diffuse reflection image and the specular reflection image, and the process proceeds to step S 108 .
- a description will be given as an example of acquiring a saturation difference image obtained by subtracting a diffuse reflection image from a specular reflection image.
- step S 108 the metal gloss level determination unit 40 determines whether or not there is metal gloss.
- step S 109 the difference image acquisition unit 42 acquires a difference image having color components and the process proceeds to step S 113 . That is, a difference image is acquired by calculating a difference between a diffuse reflection image and a specular reflection image in an RGB image having color components.
- step S 113 the data conversion unit 64 performs data conversion from the RGB color space into the color space of the recording material data, and the process proceeds to step S 115 . That is, in a case where a difference image having color components is acquired in step S 109 , (R, G, B, ⁇ R, ⁇ G, ⁇ B) is converted into (C, M, Y, K, Si). In a case where a difference image having no color component is acquired in step S 111 , (R, G, B, ⁇ Gray) is converted into (C, M, Y, K, Ct).
- step S 115 the image forming unit 66 forms an image on a recording medium such as paper using the converted recording material data, and the series of processing ends.
- the texture of the object surface is displayed by a simple calculation using a small amount of image data without acquiring a bidirectional reflectance distribution function (BRDF).
- BRDF bidirectional reflectance distribution function
- the processes performed by the image processing apparatus 30 , 31 , the display device 50 , and the image forming apparatus 70 may be processes performed by software, processes performed by hardware, or a combination of both.
- the processes performed by the image processing apparatus 30 , 31 , the display device 50 , and the image forming apparatus 70 may be stored as a program in a storage medium and be distributed.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Probability & Statistics with Applications (AREA)
- Computer Graphics (AREA)
- Color Image Communication Systems (AREA)
- Image Processing (AREA)
- Facsimile Image Signal Circuits (AREA)
- Facsimile Scanning Arrangements (AREA)
- Image Generation (AREA)
Abstract
Description
Gray Scale=((0.298912*R)+(0.586611*G)+(0.114478* B)) (1)
I(x,y)=I i ·{w d ·G d(x,y)·cos θi }+I i ·{w s ·{G s(x,y)−G d(x,y)}·cosnγ} [Expression 1]
Claims (20)
Applications Claiming Priority (3)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP2019083183A JP7334458B2 (en) | 2019-04-24 | 2019-04-24 | Image processing device and image processing program |
| JPJP2019-083183 | 2019-04-24 | ||
| JP2019-083183 | 2019-04-24 |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20200342629A1 US20200342629A1 (en) | 2020-10-29 |
| US11017564B2 true US11017564B2 (en) | 2021-05-25 |
Family
ID=72917221
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/731,054 Active US11017564B2 (en) | 2019-04-24 | 2019-12-31 | Image processing apparatus and non-transitory computer readable medium storing image processing program |
Country Status (3)
| Country | Link |
|---|---|
| US (1) | US11017564B2 (en) |
| JP (1) | JP7334458B2 (en) |
| CN (1) | CN111862293B (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20220207856A1 (en) * | 2020-03-30 | 2022-06-30 | Shenzhen Hypernano Optics Technology Co., Ltd. | Method for extracting spectral information of a substance under test |
Families Citing this family (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JP7643007B2 (en) * | 2020-10-19 | 2025-03-11 | 富士フイルムビジネスイノベーション株式会社 | Measuring device and program |
| JP2023040925A (en) * | 2021-09-10 | 2023-03-23 | 富士フイルムビジネスイノベーション株式会社 | Surface inspection device and program |
| US12073544B2 (en) * | 2021-12-23 | 2024-08-27 | Dell Products L.P. | Method and system of identifying and correcting environmental illumination light sources reflecting onto display surface |
| JP7790216B2 (en) * | 2022-03-09 | 2025-12-23 | 富士フイルムビジネスイノベーション株式会社 | Information processing device and program |
Citations (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20050073939A1 (en) | 2003-10-07 | 2005-04-07 | Digital Fashion Ltd. | Shading-information acquisition device and image processing apparatus |
| US20110274351A1 (en) * | 2009-01-27 | 2011-11-10 | Masato Tsukada | Color image processing method, color image processing device, and color image processing program |
| US20130301908A1 (en) * | 2012-05-10 | 2013-11-14 | Samsung Electronics Co., Ltd. | Method and apparatus for acquiring geometry of specular object based on depth sensor |
| US20140092227A1 (en) * | 2012-05-22 | 2014-04-03 | Panasonic Corporation | Image capturing processor and endoscope |
| JP2015049691A (en) | 2013-08-30 | 2015-03-16 | 本田技研工業株式会社 | Design layer data creation device and method, and design simulation device |
| US20150253487A1 (en) * | 2010-04-16 | 2015-09-10 | Flex Lighting Ii, Llc | Reflective display comprising a frontlight with extraction features and a light redirecting optical element |
| US20180152649A1 (en) * | 2016-11-30 | 2018-05-31 | Canon Kabushiki Kaisha | Image processing apparatus, image pickup apparatus, image processing method, and storage medium |
| JP2018189223A (en) | 2017-05-11 | 2018-11-29 | 又一 児島 | Rolling pivot device using a plurality of spherical bodies |
| US20190109986A1 (en) * | 2017-10-06 | 2019-04-11 | Canon Kabushiki Kaisha | Image processing apparatus, image pickup apparatus, image processing method, and non-transitory computer-readable storage medium |
| US20190266788A1 (en) * | 2018-02-28 | 2019-08-29 | Canon Kabushiki Kaisha | System and method of rendering a surface |
Family Cites Families (9)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPH06236440A (en) * | 1993-02-08 | 1994-08-23 | Hitachi Ltd | Image processing method |
| JP4827697B2 (en) * | 2006-11-15 | 2011-11-30 | 関西ペイント株式会社 | Paint color texture map, its creation method, creation program and creation system |
| CN101146233A (en) * | 2007-09-26 | 2008-03-19 | 东南大学 | A Light Source Color Calculation and Image Correction Method |
| JP5293355B2 (en) * | 2009-04-07 | 2013-09-18 | 株式会社リコー | Glossiness evaluation method, glossiness evaluation apparatus, image evaluation apparatus having the apparatus, image evaluation method, and program for executing the method |
| JP5884362B2 (en) * | 2011-09-26 | 2016-03-15 | コニカミノルタ株式会社 | Image processing method, image reading apparatus, and image processing program |
| JP6363916B2 (en) * | 2014-09-02 | 2018-07-25 | キヤノン株式会社 | Image processing apparatus and method |
| CN106127818B (en) * | 2016-06-30 | 2019-10-11 | 珠海金山网络游戏科技有限公司 | A kind of material appearance acquisition system and method based on single image |
| JP2019029826A (en) * | 2017-07-31 | 2019-02-21 | キヤノン株式会社 | Image processing apparatus, image processing method, and program |
| CN108320272A (en) * | 2018-02-05 | 2018-07-24 | 电子科技大学 | The method that image delusters |
-
2019
- 2019-04-24 JP JP2019083183A patent/JP7334458B2/en active Active
- 2019-12-31 US US16/731,054 patent/US11017564B2/en active Active
-
2020
- 2020-02-05 CN CN202010080885.6A patent/CN111862293B/en active Active
Patent Citations (12)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20050073939A1 (en) | 2003-10-07 | 2005-04-07 | Digital Fashion Ltd. | Shading-information acquisition device and image processing apparatus |
| JP2005115645A (en) | 2003-10-07 | 2005-04-28 | Digital Fashion Ltd | Shading information acquisition device and image processor |
| US20110274351A1 (en) * | 2009-01-27 | 2011-11-10 | Masato Tsukada | Color image processing method, color image processing device, and color image processing program |
| US20150253487A1 (en) * | 2010-04-16 | 2015-09-10 | Flex Lighting Ii, Llc | Reflective display comprising a frontlight with extraction features and a light redirecting optical element |
| US20130301908A1 (en) * | 2012-05-10 | 2013-11-14 | Samsung Electronics Co., Ltd. | Method and apparatus for acquiring geometry of specular object based on depth sensor |
| US20140092227A1 (en) * | 2012-05-22 | 2014-04-03 | Panasonic Corporation | Image capturing processor and endoscope |
| JP2015049691A (en) | 2013-08-30 | 2015-03-16 | 本田技研工業株式会社 | Design layer data creation device and method, and design simulation device |
| US10269173B2 (en) | 2013-08-30 | 2019-04-23 | Honda Motor Co., Ltd. | Layer data creation device and method, and design simulation device |
| US20180152649A1 (en) * | 2016-11-30 | 2018-05-31 | Canon Kabushiki Kaisha | Image processing apparatus, image pickup apparatus, image processing method, and storage medium |
| JP2018189223A (en) | 2017-05-11 | 2018-11-29 | 又一 児島 | Rolling pivot device using a plurality of spherical bodies |
| US20190109986A1 (en) * | 2017-10-06 | 2019-04-11 | Canon Kabushiki Kaisha | Image processing apparatus, image pickup apparatus, image processing method, and non-transitory computer-readable storage medium |
| US20190266788A1 (en) * | 2018-02-28 | 2019-08-29 | Canon Kabushiki Kaisha | System and method of rendering a surface |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20220207856A1 (en) * | 2020-03-30 | 2022-06-30 | Shenzhen Hypernano Optics Technology Co., Ltd. | Method for extracting spectral information of a substance under test |
| US12106543B2 (en) * | 2020-03-30 | 2024-10-01 | Shenzhen Hypernano Optics Technology Co., Ltd. | Method for extracting spectral information of a substance under test |
Also Published As
| Publication number | Publication date |
|---|---|
| JP2020182074A (en) | 2020-11-05 |
| CN111862293B (en) | 2025-01-21 |
| JP7334458B2 (en) | 2023-08-29 |
| US20200342629A1 (en) | 2020-10-29 |
| CN111862293A (en) | 2020-10-30 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US11017564B2 (en) | Image processing apparatus and non-transitory computer readable medium storing image processing program | |
| US10606545B2 (en) | Display apparatus, scanner, and non-transitory computer readable medium | |
| JP6004481B2 (en) | Color image processing method, color image processing apparatus, and color image processing program | |
| JP7687559B2 (en) | Imaging System | |
| IL132728A (en) | Method for image processing | |
| CN102301391A (en) | Color image processing method, color image processing device and recording medium | |
| JP2000261686A (en) | Image processor, its method and recording medium | |
| JP7271888B2 (en) | Image processing device and program | |
| JP2022048321A (en) | Image processing apparatus, image processing method, and program | |
| JP2023007097A (en) | Image processing device, image processing method, and program | |
| Menk et al. | Visualisation techniques for using spatial augmented reality in the design process of a car | |
| US20050073719A1 (en) | Color image processor | |
| US11935255B2 (en) | Display apparatus | |
| US20240394953A1 (en) | Systems and Methods for Color Correcting Three-Dimensional Objects Formed by Point Cloud Data Points | |
| JP7309415B2 (en) | Image processing device, image processing method and program | |
| JP6917796B2 (en) | Image processing equipment, imaging equipment, image processing methods, and programs | |
| JP7139854B2 (en) | Information processing device and program | |
| CN118329391A (en) | AR optical waveguide stray light testing method and system | |
| JP6502275B2 (en) | Depth feeling evaluation apparatus, depth feeling evaluation method and depth feeling evaluation program | |
| US7381443B2 (en) | Method for forming print with surface textures corresponding to printed image | |
| RU2288453C2 (en) | Method for measuring color of surface and device for measuring color of surface | |
| US12430843B2 (en) | Generating lighting map using ambient light sensor and tracking camera | |
| Guarnera et al. | DIY absolute tele-colorimeter using a camera-projector system | |
| KR20200045264A (en) | Inspection area determination method and visual inspection apparatus using the same | |
| Bärz et al. | Validating photometric and colorimetric consistency of physically-based image synthesis |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: FUJI XEROX CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KUWADA, YOSHITAKA;HARIGAI, JUNGO;REEL/FRAME:051417/0717 Effective date: 20190823 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: AWAITING TC RESP., ISSUE FEE NOT PAID |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| AS | Assignment |
Owner name: FUJIFILM BUSINESS INNOVATION CORP., JAPAN Free format text: CHANGE OF NAME;ASSIGNOR:FUJI XEROX CO., LTD.;REEL/FRAME:056237/0501 Effective date: 20210401 |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |