US20130077869A1 - Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns - Google Patents
Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns Download PDFInfo
- Publication number
- US20130077869A1 US20130077869A1 US13/615,521 US201213615521A US2013077869A1 US 20130077869 A1 US20130077869 A1 US 20130077869A1 US 201213615521 A US201213615521 A US 201213615521A US 2013077869 A1 US2013077869 A1 US 2013077869A1
- Authority
- US
- United States
- Prior art keywords
- importance
- image
- region
- unit
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/001—Texturing; Colouring; Generation of texture or colour
Definitions
- the present invention relates to an image processing apparatus and an image processing method as well as a storage medium.
- Japanese Unexamined Patent Application, Publication No. 2011-139329 discloses painterly conversion processing, which is executed on data of an original image to be converted into data of a painterly image such as an image similar to an oil painting or a colored pencil drawing.
- Japanese Unexamined Patent Application, Publication No. 2006-114024 discloses image processing, in which an edge is detected in data of an original image that includes a person's face as a subject, thereby converting the data of the original image into data of an image that is similar to an ink-wash painting (suiboku-ga) (hereinafter referred to as an “ink-wash-painting-like image”) that is a type of an image with high artistry.
- ink-wash painting suiboku-ga
- a converted image may be an image being remote from a real ink-wash painting in some cases.
- a first conversion unit that converts data of an original image into data of a painterly image
- a characteristic region detection unit that detects a characteristic region of the original image, from the data of the original image
- a second conversion unit that executes gradation processing of gradating the characteristic region detected by the characteristic region detection unit, and margin setting processing of setting a margin region to be added to the painterly image, as image processing of further converting the data of the painterly image that was converted by the first conversion unit.
- an image processing method performed by an image processing apparatus to execute image processing on an original image according to one aspect of the present invention including:
- a storage medium having stored therein a computer readable program for controlling an image processing apparatus that executes image processing on an original image according to one aspect of the present invention, the program causing a computer to implement functions of:
- a first converting unit that converts data of the original image into data of a painterly image
- a characteristic region detection unit that detects a characteristic region of the original image, from the data of the original image
- a second conversion unit that executes gradation processing of gradating the characteristic region detected by the characteristic region detection unit, and margin setting processing of setting a margin region to be added to the painterly image, as image processing of further converting the data of the painterly image that was converted by the first conversion unit.
- FIG. 1 is a block diagram showing a hardware configuration of an image processing apparatus according to the first embodiment of the present invention
- FIG. 2 is a functional block diagram showing a functional configuration for executing ink-wash-painting-like image generation processing, overall gradation processing, and margin region setting processing, in a functional configuration of the image processing apparatus shown in FIG. 1 ;
- FIG. 3 shows an example of data of an original image
- FIG. 4A is a diagram showing an example of a brushstroke pattern
- FIG. 4B is a diagram showing an example of a brushstroke pattern
- FIG. 5 is a diagram showing examples of edge images
- FIG. 6 a diagram showing examples of ink-wash-painting-like images that are converted from the edge images shown in FIG. 5 ;
- FIG. 7A shows an example of an image after setting a margin region to be added to an ink-wash-painting-like image, by way of an ink-wash-painting-like image and a white background image;
- FIG. 7B shows an example of an image after setting a margin region to be added to an ink-wash-painting-like image, by way of an ink-wash-painting-like image and a white background image;
- FIG. 7C shows an example of an image after setting a margin region to be added to an ink-wash-painting-like image, by way of an ink-wash-painting-like image and a white background image;
- FIG. 8 is a flowchart showing an example of a flow of ink-wash-painting-like image processing that is executed by the image processing apparatus shown in FIG. 1 having the functional configuration shown in FIG. 2 ;
- FIG. 9 is a flowchart illustrating importance region calculation processing, regarding the ink-wash-painting-like image processing shown in FIG. 8 ;
- FIG. 10 is a flowchart illustrating ink-wash-painting-like image generation processing, regarding the ink-wash-painting-like image processing shown in FIG. 8 ;
- FIG. 11 is a flowchart illustrating overall gradation processing, regarding the ink-wash-painting-like image processing shown in FIG. 8 ;
- FIG. 12 is a flowchart illustrating margin region setting processing, regarding the ink-wash-painting-like image processing shown in FIG. 8 ;
- FIG. 13 is a functional block diagram showing a functional configuration for executing overall gradation processing according to a second embodiment
- FIG. 14 shows an example of data of an original image of the second embodiment
- FIG. 15 is a flowchart illustrating the overall gradation processing, regarding the ink-wash-painting-like image processing of the second embodiment.
- FIG. 1 is a block diagram showing a hardware configuration of an image processing apparatus according to the first embodiment of the present invention.
- the image processing apparatus 1 is provided with a CPU (Central Processing Unit) 11 , ROM (Read Only Memory) 12 , RAM (Random Access Memory) 13 , an image processing unit 14 , a bus 15 , an input/output interface 16 , an input unit 17 , an output unit 18 , a storage unit 19 , a communication unit 20 , and a drive 21 .
- CPU Central Processing Unit
- ROM Read Only Memory
- RAM Random Access Memory
- the CPU 11 executes various processing according to programs that are recorded in the ROM 12 , or programs that are loaded from the storage unit 19 to the RAM 13 .
- the RAM 13 also stores data and the like necessary for the CPU 11 to execute the various processing, as appropriate.
- the image processing unit 14 is configured by a DSP (Digital Signal Processor), VRAM (Video Random Access Memory) and the like, and collaborates with the CPU 11 to execute various image processing on image data.
- the image processing unit 14 also includes a function of executing processing on data of an original image to be converted into data of a painterly image such as an image similar to an oil painting or a colored pencil drawing.
- the CPU 11 , the ROM 12 , the RAM 13 , and the image processing unit 14 are connected to one another via the bus 15 .
- the bus 15 is also connected with the input/output interface 16 .
- the input/output interface 16 is connected to the input unit 17 , the output unit 18 , the storage unit 19 , the communication unit 20 , and the drive 21 .
- the input unit 17 is configured by a keyboard, a mouse and the like, and inputs various information in accordance with instruction operations by a user.
- the output unit 18 is configured by a display, a speaker and the like, and outputs images and sound.
- the storage unit 49 is configured by DRAM (Dynamic Random Access Memory) or the like, and stores data of various images.
- DRAM Dynamic Random Access Memory
- the communication unit 20 controls communication with other devices (not shown) via a network, which includes the Internet.
- Removable media 31 made of a magnetic disk, an optical disk, a magneto-optical disk, semiconductor memory or the like is installed in the drive 21 , as appropriate. Programs that are read via the drive 21 from the removable media 31 are installed in the storage unit 19 , as necessary. In addition, similarly to the storage unit 19 , the removable media 31 can also store various data such as the image data stored in the storage unit 19 .
- FIG. 2 is a functional block diagram showing a functional configuration for executing ink-wash-painting-like image generation processing, overall gradation processing, and margin region setting processing, in the functional configuration of such an image processing apparatus 1 .
- the ink-wash-painting-like image generation processing refers to a sequence of processing of generating data of an ink-wash-painting-like image from data of an original image that is input as an object of image processing.
- the overall gradation processing refers to a sequence of processing of gradating a characteristic region in the ink-wash-painting-like image, based on importance information calculated from the data of the original image.
- the margin region setting processing refers to a sequence of processing of setting a margin region to be added to the ink-wash-painting-like image, based on the importance information.
- an original image acquisition unit 41 In a case in which the ink-wash-painting-like image generation processing is executed according to control by the CPU 11 , an original image acquisition unit 41 , a characteristic region detection unit 42 , and an ink-wash painting conversion unit 44 function in the image processing unit 14 . In a case in which the overall gradation processing is executed according to control by the CPU 11 , the original image acquisition unit 41 , an importance information calculation unit 43 , and a conversion unit 45 function in the image processing unit 14 .
- the importance information calculation unit 43 and the conversion unit 45 function in the image processing unit 14 .
- a brushstroke pattern storage unit (not shown) is provided as an area of the storage unit 19 .
- the original image acquisition unit 41 acquires data of an image transmitted from another device and received by the communication unit 20 , data of an image that is read from the removable media 31 via the drive 21 and the like as data of the original image, and supplies such data to the characteristic region detection unit 42 and the importance information calculation unit 43 .
- the characteristic region detection unit 42 executes edge detection processing on data of the original image, and generates data of a binary image, in which, for example, the detected edge is black, and the other portions are white (such a binary image is hereinafter referred to as an “edge image”).
- the characteristic region detection unit 42 detects a characteristic region for data of the edge image thus generated.
- the object detected as a characteristic region is not limited in particular, it is an edge region in the present embodiment.
- the data of the edge image thus generated by the characteristic region detection unit 42 is supplied to the ink-wash painting conversion unit 44 .
- the importance information calculation unit 43 detects importance of each pixel composing an original image, by using a color, brightness, an edge direction or the like as an indicator of importance, from the data of the original image acquired by the original image acquisition unit 41 . In addition, the importance information calculation unit 43 generates an importance map indicating spatial distribution of the importance detected for each pixel, thereby executing importance region calculation processing of calculating importance information of the original image.
- the importance information calculation unit 43 calculates importance information by executing the importance region calculation processing on data of the original image as shown in FIG. 3 .
- the importance information thus calculated by the importance information calculation unit 43 is supplied to the ink-wash painting conversion unit 44 and the conversion unit 45 .
- FIG. 3 shows an example of data of an original image 100 .
- the importance region 102 is a gray region, among each of a plurality of pixels 101 composing the original image 100 .
- an importance barycenter 111 calculated based on the importance region 102 is shown in a substantially central portion of the importance region 102 .
- the importance information calculation unit 43 of the present embodiment includes an importance calculation unit 51 , an importance region calculation unit 52 , and an importance barycenter calculation unit 53 .
- the importance calculation unit 51 calculates importance of each pixel composing the original image acquired by the original image acquisition unit 41 . For example, in the example shown in FIG. 3 , importance of each pixel composing the original image 100 is calculated.
- the calculation of importance is not limited in particular, but in the present embodiment, the importance calculation unit 51 calculates importance, based on color information, brightness (brightness information), or information of an edge direction, within a predetermined range around each pixel. In this case, since importance cannot be calculated based on only a single pixel, the importance calculation unit 51 calculates importance based on whether pixels as representative values are important for a peripheral region around a particular region with respect to an attention pixel.
- the importance calculation unit 51 when detecting an edge in a direction different from the direction of the edge around the attention pixel, the importance calculation unit 51 recognizes such a portion as an object having a color different from its peripheral colors, thereby recognizing that a characteristic object is drawn in the portion, and increasing the importance of the portion.
- the importance information calculated by the importance calculation unit 51 is supplied to the importance region calculation unit 52 and the importance barycenter calculation unit 53 .
- the importance region calculation unit 52 calculates an importance region, for example, the importance region 102 in the example shown in FIG. 3 .
- the calculation of the importance region is not limited in particular, but in the present embodiment, in a case in which an upper limit of an importance evaluation value calculated by the importance calculation unit 51 is assumed to be 100, a region with an importance evaluation value being 90 or higher can be calculated as an importance region.
- the information of the importance region calculated by the importance region calculation unit 52 is supplied to the ink-wash painting conversion unit 44 and the conversion unit 45 .
- the importance barycenter calculation unit 53 calculates an importance barycenter of the original image.
- the calculation of an importance barycenter is not limited in particular, but in the present embodiment, when the importance region calculation unit 52 calculates an importance region, the importance barycenter calculation unit 53 calculates a central position in such a region as an importance barycenter. More specifically, at first, the importance barycenter calculation unit 53 calculates an importance barycenter (x, y) in spatial distribution of the importance.
- the importance barycenter calculation unit 53 calculates the importance barycenter (x, y) by applying Equations (1) and (2) as follows to pixels having importance of at least a threshold value.
- the threshold value can be appropriately changed by the user as necessary, and in the present embodiment, the threshold value can be 90 in a case in which the upper limit of the importance is 100.
- n represents the number of pixels having importance of at least a threshold value
- sumw represents a sum of the importance
- w (x, y) represents importance in (x, y)
- the importance barycenter information calculated by the importance information calculation unit 43 is supplied to the conversion unit 45 .
- the ink-wash painting conversion unit 44 executes processing of converting the data of the edge image into data of an ink-wash-painting-like image. Such processing is hereinafter referred to as “ink-wash painting conversion processing”.
- a texture that imitates handwriting of a writing brush for calligraphy is applied to an original image.
- a pattern of the “texture that imitates handwriting of a writing brush for calligraphy” is referred to as a “brushstroke pattern” in the present specification.
- a shape, a size and the like of the texture employed as the brushstroke pattern are not limited in particular. However, two types of brushstroke patterns shown in FIG. 4 are employed in the present embodiment.
- FIG. 4 is a diagram showing examples of the brushstroke patterns.
- FIG. 4A shows a brushstroke pattern that imitates handwriting in a case of using a portion “the longest hair on the tip” of the writing brush (an extreme tip portion of the writing brush). It should be noted that the brushstroke pattern as shown in FIG. 4A is hereinafter referred to as an “upright-brush brushstroke pattern”.
- FIG. 4B shows a brushstroke pattern that imitates handwriting in a case of using a lateral portion (a part of the “head”) of the writing brush ranging from a “neck (portion adjacent to the extreme tip)” to a “ventral (middle portion)” thereof.
- the brushstroke pattern as shown in FIG. 4B is hereinafter referred to as an “oblique-brush brushstroke pattern”.
- the upright-brush brushstroke pattern is characterized in that a length in a direction orthogonal to a longitudinal direction (hereinafter referred to as a “width”) is narrow, as a result of which the variation in the gray levels is small.
- the oblique-brush brushstroke pattern is characterized in that the width is wide, as a result of which the variation in the gray levels is large.
- each data of the upright-brush brushstroke pattern and the oblique-brush brushstroke pattern as described above is stored in a brushstroke pattern storage unit (not shown).
- a technique for generating data of the brushstroke patterns is not limited in particular, and for example, a technique for generating data by using CG (Computer Graphics) may be employed.
- the present embodiment employs a technique, in which handwriting actually written on a paper medium with a writing brush soaked in ink (Japanese Sumi) is captured with a scanner or a digital camera to create data of brushstroke patterns.
- the ink-wash painting conversion unit 44 executes ink-wash painting conversion processing on data of an edge image, in which the upright-brush brushstroke pattern is applied to an edge region, the oblique-brush brushstroke pattern is applied to a surrounding region of the edge region, and in a case in which the edge region forms a closed region, the oblique-brush brushstroke pattern is applied to an inside of the closed region.
- the ink-wash painting conversion unit 44 executes the ink-wash painting conversion processing on data of an edge image as shown in FIG. 5 , thereby generating data of an ink-wash-painting-like image as shown in FIG. 6 .
- FIG. 5 shows examples of edge images.
- a black and narrow area is the edge region.
- an edge region E 1 shows a contour of a mountain
- an edge region E 2 shows a contour of a cloud
- an edge region E 3 shows a contour of a house.
- FIG. 6 shows examples of ink-wash-painting-like images that are converted from the edge images in FIG. 5 .
- An upright-brush brushstroke pattern D 1 is applied to the edge region E 1 showing the contour of the mountain, and an oblique-brush brushstroke pattern S 1 is applied to a surrounding region on the right side of the edge region E 1 .
- edge region E 2 showing the contour of the cloud forms a closed region
- an upright-brush brushstroke pattern D 2 is applied to the edge region E 2
- an oblique-brush brushstroke pattern S 2 is applied so as to paint out an inside of the closed region.
- an upright-brush brushstroke pattern D 3 is applied to the edge region E 3
- an oblique-brush brushstroke pattern S 3 is applied so as to paint out an inside of the closed region.
- the ink-wash painting conversion unit 44 of the present embodiment executes processing as follows.
- the ink-wash painting conversion unit 44 detects a characteristic region for data of the edge image.
- the object detected as a characteristic region is not limited in particular, it is an edge region in the present embodiment.
- the ink-wash painting conversion unit 44 determines a brushstroke pattern to be used from among brushstroke patterns, of which data is stored in the brushstroke pattern storage unit (not shown).
- the ink-wash painting conversion unit 44 determines an upright-brush brushstroke pattern as the brushstroke pattern to be used for the edge region that is the characteristic region. Moreover, the ink-wash painting conversion unit 44 determines an oblique-brush brushstroke pattern as the brushstroke pattern to be used for the surrounding region of the edge region that is the characteristic region, or for the region inside a closed curve of the edge region that forms the closed curve.
- the ink-wash painting conversion unit 44 converts the data of the edge image into data of an ink-wash-painting-like image by using data of the brushstroke pattern thus determined.
- the ink-wash painting conversion unit 44 converts the data of the edge region into data of the upright-brush brushstroke pattern, such that the upright-brush brushstroke pattern is applied to the edge region.
- the ink-wash painting conversion unit 44 converts the data of the surrounding region of the edge region into data of the oblique-brush brushstroke pattern, such that the oblique-brush brushstroke pattern is applied to the surrounding region of the edge region.
- the ink-wash painting conversion unit 44 converts data of the inside of the closed region into data of the oblique-brush brushstroke pattern, such that the oblique-brush brushstroke pattern is applied so as to paint out the inside of the closed region.
- the ink-wash painting conversion unit 44 Based on the importance map generated by the importance information calculation unit 43 , the ink-wash painting conversion unit 44 adjusts each color of the brushstroke pattern used in the ink-wash painting conversion unit 44 , i.e. adjusts the gray levels of the ink.
- the importance map shows the spatial distribution of the importance of each pixel.
- the gray level is determined for each brushstroke pattern occupying a region consisting of a plurality of pixel groups.
- the ink-wash painting conversion unit 44 extracts, from the importance map, each importance of the plurality of pixel groups included in the region occupying the brushstroke pattern of an adjustment object, and calculates integrated importance of the entire region, based on a plurality of pieces of importance thus extracted.
- a technique for calculating integrated importance of the entire region is not limited in particular as long as importance of a plurality of pixels composing the region is used, and for example, a technique for calculating a root mean square and a mean value can be employed.
- a technique for calculating a simple average of importance of a plurality of pixels is employed in the present embodiment.
- the ink-wash painting conversion unit 44 adjusts the gray levels, such that the brushstroke pattern occupying the region is darker (verges on black) as the integrated importance of the entire region is increased, whereas the brushstroke pattern occupying the region is lighter (verges on white) as the integrated importance of the entire region is decreased.
- the ink-wash painting conversion unit 44 executes image processing of gradating from the region of the processing object to its perimeter, for each data of the region that was converted into (applied as) a brushstroke pattern by the ink-wash painting conversion unit 44 (this image processing is hereinafter referred to as “gradation processing”).
- the amount (width) of gradation in the gradation processing is determined depending on the gray levels in the region of the processing object. More specifically, since the gray levels in the region of the processing object are adjusted based on the importance map as described above, the amount of gradation in the region of the processing object is also determined based on the importance map. In this case, as the importance is higher, the color is darker (verges on black), and the amount of gradation is smaller; conversely, as the importance is lower, the color is lighter (verges on white), and the amount of gradation is greater.
- the manner of gradating each pixel may be a technique that depends on a distance x from an edge of a region of a processing object (a brushstroke pattern), and is not limited in particular; however, the present embodiment employs a technique, in which the color is lighter as the distance x is increased. More specifically, the present embodiment employs a technique, in which the gray scales of an image (a range of brightness indicating the gray levels) are 256 gray scales, and the color gradation (brightness indicating the gray levels) of a pixel of a processing object is calculated according to Equation (3) as follows.
- B represents the color gradation (brightness indicating the gray levels) of a pixel of a processing object.
- L represents a color of the brushstroke pattern applied to the pixel of the processing object (brightness indicating the gray levels regarding the entire region of the processing object).
- f(D+n) represents an arbitrary function, of which output value is increased in accordance with an input parameter (D+n).
- D represents an amount of gradation in the brushstroke pattern (the region of the processing object) applied to the pixel of the processing object.
- n represents an arbitrary integer.
- the conversion unit 45 includes: a gradation unit 61 that executes overall gradation processing on data of the ink-wash-painting-like image in the characteristic region, based on the importance information calculated by the importance information calculation unit 43 ; and a margin region setting unit 62 that executes margin region setting processing of setting a margin region of the original image.
- the gradation unit 61 calculates an amount of gradation in each pixel. More specifically, the gradation unit 61 calculates an amount of gradation such that the amount of gradation in the edge is maximized, by gradually changing the brightness toward the white gray scale (256), based on primary approximation by a predetermined function and the Gaussian distribution (normal distribution). In other words, the gradation processing depends on a distance x from the edge of the brush, and the color is lighter as the distance from the edge of the brush is increased.
- a gradation color B is expressed by Equation (4) as follows.
- 256 gray scales are used as the gray scales of an image, and a function f represents an arbitrary function that is increased according to a value of D+n. Moreover, n represents an arbitrary integer.
- the gradation unit 61 executes gradation processing on the entire image of the ink-wash-painting-like image, based on the amount of gradation calculated for each pixel.
- the margin region setting unit 62 executes margin region setting processing on data of the ink-wash-painting-like image, thereby setting a margin region to be added to the ink-wash-painting-like image.
- the margin region setting unit 62 of the present embodiment includes an importance barycenter position ratio calculation unit 71 , a background image position ratio calculation unit 72 , and a synthesis unit 73 .
- the importance barycenter position ratio calculation unit 71 calculates a position ratio of the importance barycenter calculated by the importance barycenter calculation unit 53 .
- the background image position ratio calculation unit 72 calculates a barycenter position ratio of a white background image.
- the synthesis unit 73 executes synthesis processing such that the data of the ink-wash-painting-like image, on which the gradation processing was executed by the gradation unit 61 , is synthesized so as to be superimposed on the data of the white background image.
- the image processing unit 14 outputs data, on which the synthesis processing was executed, as data of a final output image.
- FIG. 7 shows an example of an image after setting a margin region 140 to be added to an ink-wash-painting-like image 120 , with the ink-wash-painting-like image 120 and a white background image 130 .
- the margin region 140 is set as a region between the periphery of the ink-wash-painting-like image 120 and the white background image 130 superimposed with the ink-wash-painting-like image 120 .
- FIG. 7A shows the ink-wash-painting-like image 120 having coordinates (x, y) composed of a width X and a height Y of the importance barycenter.
- the margin region setting unit 62 sets a width A and a height B for a size of the white background image 130 having an aspect ratio (A/B) that is identical to the aspect ratio (X/Y) of the ink-wash-painting-like image 120 .
- A/B aspect ratio
- X/Y aspect ratio of the ink-wash-painting-like image 120
- the margin degree ( ⁇ ) refers to a degree of a size (an area) of the margin region 140 that is set between the ink-wash-painting-like image 120 and the white background image 130 .
- the margin degree can be freely set as appropriate by the user operating the input unit (not shown). In a case in which the margin degree is set high based on the operation by the user, the margin region setting unit 62 sets the size (the area) of the margin region 140 large, and in a case in which the margin degree is set low, the margin region setting unit 62 sets the size (the area) of the margin region 140 small.
- (a, b) are calculated such that Equations (5) to (7) as follows are established.
- the background image position ratio calculation unit 72 calculates such that the margin region 140 on a side of such a portion is small, and the margin region 140 on a side far from such a center of importance is large. Therefore, the ink-wash-painting-like image 120 (the original image) can be displayed by respecting an object that is desired to receive attention. In addition, as shown in FIG.
- the margin region setting unit 62 executes the synthesis processing of synthesizing each data of the ink-wash-painting-like image 120 and the white background image 130 , such that the coordinates (x, y) of the importance barycenter 111 of the ink-wash-painting-like image 120 (the original image) coincide with the coordinates (a, b) of the barycenter 131 of the white background image 130 .
- FIG. 8 is a flowchart illustrating an example of a flow of the ink-wash-painting-like image processing.
- the original image acquisition unit 41 acquires data of an original image
- the ink-wash-painting-like image processing is initiated, and a sequence of processing is executed as follows.
- Step S 11 the original image acquisition unit 41 acquires data of the original image.
- Step S 12 the characteristic region detection unit 42 executes smoothing processing on the data of the original image.
- the smoothing processing refers to image processing, in which noise and fine texture included in the original image are considered to be unnecessary in an ink-wash-painting-like image, and are therefore removed or reduced from an original image, thereby smoothing the original image such that the edge representing the contour of the subject is preserved.
- Step S 13 the importance information calculation unit 43 executes importance region calculation processing to be described below with reference to FIG. 9 , thereby calculating importance information of the original image from the data of the original image acquired by the original image acquisition unit 41 .
- Step S 14 the image processing unit 14 executes monochromatic processing on the data of the original image, thereby generating data of the edge image that is made binary (black and white).
- Step S 15 the ink-wash painting conversion unit 44 executes ink-wash-painting-like image generation processing to be described below with reference to FIG. 10 , thereby converting the data of the edge image into data of the ink-wash-painting-like image.
- Step S 16 the gradation unit 61 executes overall gradation processing to be described below with reference to FIG. 11 , thereby calculating an amount of gradation in each pixel, and executing the gradation processing on the entire data of the ink-wash-painting-like image, based on the amount of gradation calculated for each pixel.
- Step S 17 the margin region setting unit 62 executes margin region setting processing to be described below with reference to FIG. 12 , thereby executing synthesis processing of synthesizing each data of the ink-wash-painting-like image, on which the gradation processing was executed, and the white background image, based on the importance barycenter of the ink-wash-painting-like image (the importance barycenter of the original image) and the barycenter of the white background image.
- Step S 18 the image processing unit 14 stores the data of the ink-wash-painting-like image with the margin region being set, i.e. the data of the image, in which the margin region was added to the ink-wash-painting-like image, on which the gradation processing was executed, into the storage unit 19 shown in FIG. 1 .
- the size of the image with the margin region added to the ink-wash-painting-like image i.e. the resolution
- the image processing unit 14 executes reduction processing on the data of the image with the margin region added to the ink-wash-painting-like image, such that the size coincides with the size of the original image.
- FIG. 9 is a flowchart illustrating the importance region calculation processing.
- Step S 31 the importance calculation unit 51 calculates importance of each pixel of the original image. For example, in the example shown in FIG. 3 , importance of each pixel 101 of the original image 100 is calculated.
- Step S 32 the importance barycenter calculation unit 53 calculates an importance barycenter of the original image, based on the importance of each pixel thus calculated. For example, in the example shown in FIG. 3 , an importance barycenter 11 of the original image 100 is calculated based on the importance of each pixel 101 .
- the importance region calculation unit 52 calculates an importance region of the original image, based on pixels having importance of at least a threshold value. For example, in the example shown in FIG. 3 , the importance region 102 of the original image 100 is calculated from the pixels 101 .
- Step S 13 shown in FIG. 8 the processing in Step S 13 shown in FIG. 8 is terminated, and the processing advances to Step S 14 .
- Step S 15 a detailed flow of the ink-wash-painting-like image generation processing in Step S 15 is described with reference to FIG. 10 .
- FIG. 10 is a flowchart illustrating the ink-wash-painting-like image generation processing.
- Step S 51 the characteristic region detection unit 42 searches the data of the edge image for an edge.
- Step S 52 the characteristic region detection unit 42 determines whether an edge exists, based on a result of the processing in Step S 51 .
- Step S 53 the characteristic region detection unit 42 traces the edge.
- the characteristic region detection unit 42 scans the edge image in a so-called raster sequence from the upper left, and searches for pixels belonging to the edge region (Step S 51 ). In a case in which a pixel belonging to such an edge region exists (Step S 52 : YES), the characteristic region detection unit 42 traces the edge so as to search for other pixels belonging to the edge region (Step S 53 ).
- Step 54 based on the edge region traced in this manner, the ink-wash painting conversion unit 44 determines a brushstroke pattern to be used in subsequent steps, from among brushstroke patterns, of which data is stored in the brushstroke pattern storage unit (not shown).
- the ink-wash painting conversion unit 44 determines an upright-brush brushstroke pattern as the brushstroke pattern to be used for the edge region that was traced in the processing in Step S 53 . Moreover, the ink-wash painting conversion unit 44 determines an oblique-brush brushstroke pattern as the brushstroke pattern to be used for the surrounding region of the edge region. Furthermore, in a case in which the edge region is a region of a closed curve, the ink-wash painting conversion unit 44 determines the oblique-brush brushstroke pattern as the brushstroke pattern to be used for the inside of the closed curve.
- Step S 55 the ink-wash painting conversion unit 44 executes processing of converting the data of the edge region into data of the upright-brush brushstroke pattern, such that the upright-brush brushstroke pattern is applied to the edge region that was traced in the processing in Step S 53 (hereinafter referred to as “application processing”).
- the ink-wash painting conversion unit 44 enlarges or reduces the data of the upright-brush brushstroke pattern that is read from the brushstroke pattern storage unit (not shown), in accordance with the length of the edge region.
- the ink-wash painting conversion unit 44 converts (applies) the data of the edge region into (to) the data of the upright-brush brushstroke pattern that was enlarged or reduced.
- Step S 56 based on the importance region calculated in the processing in Step S 33 ( FIG. 9 ), the ink-wash painting conversion unit 44 calculates a drawing color of the upright-brush brushstroke pattern, on which the application processing was executed in Step S 55 .
- the drawing color i.e. the gray level of the ink, is determined in accordance with the importance of the original image. Since the importance has a value for each pixel, when the drawing color is determined, the ink-wash painting conversion unit 44 calculates an average of the importance in a region to be drawn with a brush. The ink-wash painting conversion unit 44 draws in a darker color (a color closer to black) as the average importance thus calculated is higher, and draws in a lighter color (a color closer to white) as the average importance is lower. Also in processing in Steps S 59 and S 63 to be described below, similarly to the processing in Step S 56 , the ink-wash painting conversion unit 44 calculates a drawing color based on an average of the importance.
- Step S 57 based on the importance region calculated in the processing in Step S 33 ( FIG. 9 ), the ink-wash painting conversion unit 44 executes gradation processing on the upright-brush brushstroke pattern, on which the application processing was executed in Step S 55 .
- the ink-wash painting conversion unit 44 determines an amount (width) of gradation based on the importance of the image.
- the ink-wash painting conversion unit 44 executes gradation processing based on the importance.
- Step S 58 the ink-wash painting conversion unit 44 executes application processing, such that the oblique-brush brushstroke pattern is applied to the surrounding region of the edge region traced in the processing in Step S 53 .
- Step S 59 based on the importance region calculated in the processing in Step S 33 ( FIG. 9 ), the ink-wash painting conversion unit 44 calculates a drawing color of the oblique-brush brushstroke pattern, on which the application processing was executed in Step S 58 .
- Step S 60 based on the importance region calculated in the processing in Step S 33 ( FIG. 9 ), the ink-wash painting conversion unit 44 executes gradation processing on the oblique-brush brushstroke pattern, on which the application processing was executed in Step S 58 .
- Step S 61 the ink-wash painting conversion unit 44 determines whether the edge region traced in the processing in Step S 53 is a region of a closed curve.
- Step S 61 the determination in Step S 61 is NO, the processing returns to Step S 51 in which another edge is searched, and the processing in and after Step S 52 is repeated.
- Step S 61 determines whether the edge region is a region of a closed curve. If the edge region is a region of a closed curve, i.e. in a case in which the region is a region without a starting point and an ending point.
- Step S 62 the ink-wash painting conversion unit 44 executes application processing, such that the oblique-brush brushstroke pattern is applied to the inside of the closed curve of the edge region traced in the processing in Step S 53 .
- Step S 63 based on the importance region calculated in the processing in Step S 33 ( FIG. 9 ), the ink-wash painting conversion unit 44 calculates a drawing color of the oblique-brush brushstroke pattern, on which the application processing was executed in Step S 62 .
- Step S 64 based on the importance region calculated in the processing in Step S 33 ( FIG. 9 ), the ink-wash painting conversion unit 44 executes gradation processing on the oblique-brush brushstroke pattern, on which the application processing was executed in Step S 62 .
- Step S 51 in which another edge is searched, and the processing in and after Step S 52 is repeated.
- Steps S 51 to S 64 are repeatedly executed on each edge region included in the edge image.
- the determination in the next Step S 52 is NO, and the ink-wash-painting-like image generation processing is terminated. In other words, the processing in Step 15 shown in FIG. 8 is terminated.
- Step S 16 a detailed flow of the overall gradation processing in Step S 16 is described with reference to FIG. 11 .
- FIG. 11 is a flowchart illustrating the overall gradation processing.
- Step S 81 the gradation unit 61 acquires importance barycenter information calculated by the importance barycenter calculation unit 53 .
- Step S 82 the gradation unit 61 acquires information of the importance region calculated by the importance region calculation unit 52 .
- Step S 83 the gradation unit 61 sets one pixel in the importance region of the original image as an attention pixel.
- Step S 84 the gradation unit 61 calculates an amount of gradation, based on the importance of the attention pixel, and based on a distance between each pixel and the importance barycenter acquired in Step S 81 .
- Step S 85 the gradation unit 61 determines whether all pixels in the importance region have been set as attention pixels. In a case in which all pixels in the importance region have not been set as attention pixels in Step S 83 , the determination in Step S 85 is NO, and the processing returns to Step S 83 . In other words, the processing in Steps S 83 to S 85 is repeated until the amount of gradation is calculated for all pixels. On the other hand, in a case in which all pixels in the importance region have been set as attention pixels in Step S 83 , the determination in Step S 85 is YES, and the processing advances to Step S 86 .
- Step S 86 the gradation unit 61 executes gradation processing of gradating the entire image of the original image, based on the gradation amount calculated for each pixel in Step S 84 .
- the overall gradation processing is terminated, i.e. the processing in Step 16 shown in FIG. 8 is terminated, and the processing advances to Step S 17 .
- Step S 17 a detailed flow of the margin region setting processing in Step S 17 is described with reference to FIG. 12 .
- FIG. 12 is a flowchart illustrating the margin region setting processing.
- Step S 101 the importance barycenter position ratio calculation unit 71 calculates a position ratio of the importance barycenter of the ink-wash-painting-like image. For example, in the example shown in FIG. 7 , a position ratio of the importance barycenter 111 of the original image 100 is calculated.
- Step S 102 the margin region setting unit 62 sets a level of the margin degree of the margin region 140 shown in FIG. 7 , based on the operation by the user. For example, in the example shown in FIG. 7 , a level of the margin degree of the margin region 140 is set.
- Step S 103 the margin region setting unit 62 sets a size of the white background image. For example, in the example shown in FIG. 7 , a size of the white background image 130 is set.
- Step S 104 the background image position ratio calculation unit 72 calculates a barycenter of the white background image. For example, in the example shown in FIG. 7 , the barycenter 131 of the white background image 130 is calculated.
- Step S 105 the synthesis unit 73 executes synthesis processing of synthesis and superimposition such that a position of the coordinates of the barycenter of the white background image coincides with a position of the coordinates of the importance barycenter of the ink-wash-painting-like image, based on the barycenter of the white background image and the importance barycenter of the ink-wash-painting-like image.
- the synthesis processing of synthesis and superimposition is executed, such that the positions of the coordinates of the barycenter 131 and the importance barycenter 111 coincide with each other, based on the barycenter 131 of the white background image 130 and the importance barycenter 111 of the original image 100 .
- the margin region setting processing is terminated, i.e. the processing in Step 17 shown in FIG. 8 is terminated, and the processing advances to Step S 18 .
- the image processing apparatus of the first embodiment configured as above includes the original image acquisition unit 41 , the characteristic region detection unit 42 , the importance information calculation unit 43 , the ink-wash painting conversion unit 44 , and the conversion unit 45 .
- the original image acquisition unit 41 acquires data of the original image 100 .
- the ink-wash painting conversion unit 44 converts the data of the original image 100 acquired by the original image acquisition unit 41 into data of the ink-wash-painting-like image 120 .
- the characteristic region detection unit 42 detects a characteristic region (an edge region in the present embodiment) of the original image 100 from the data of the original image 100 acquired by the original image acquisition unit 41 .
- the importance information calculation unit 43 calculates importance information of the original image 100 , based on the data of the original image 100 acquired by the original image acquisition unit 41 .
- the conversion unit 45 executes the gradation processing of gradating the characteristic region detected by the characteristic region detection unit 42 , and the margin setting processing of setting the margin region 140 to be added to the ink-wash-painting-like image, based on the importance information calculated by the importance information calculation unit 43 .
- the importance information calculation unit 43 of the image processing apparatus of the present embodiment includes the importance calculation unit 51 and the importance barycenter calculation unit 53 .
- the importance calculation unit 51 calculates importance of each pixel 101 composing the original image 100 .
- the importance barycenter calculation unit 53 calculates the importance barycenter 111 of the original image 100 .
- the conversion unit 45 executes conversion of the data of the ink-wash-painting-like image 120 .
- the conversion unit 45 of the image processing apparatus of the present embodiment includes the gradation unit 61 .
- the gradation unit 61 executes the gradation processing on the characteristic region detected by the characteristic region detection unit 42 , and an object(s) in at least one region in the characteristic region, based on the importance barycenter calculated by the importance barycenter calculation unit 53 .
- the conversion unit 45 of the image processing apparatus of the present embodiment includes the importance barycenter position ratio calculation unit 71 and the background image position ratio calculation unit 72 .
- the importance barycenter position ratio calculation unit 71 calculates a position ratio of the importance barycenter calculated by the importance barycenter calculation unit 53 .
- the background image position ratio calculation unit 72 calculates a barycenter position ratio of a white background image 130 .
- the conversion unit 45 Based on the position ratio of the importance barycenter calculated by the importance barycenter position ratio calculation unit 71 , and based on the position ratio of the barycenter of the white background image 130 calculated by the background image position ratio calculation unit 72 , the conversion unit 45 synthesizes the data of the white background image 130 and the data of the ink-wash-painting-like image, thereby setting the margin region 140 to be added to the ink-wash-painting-like image 120 .
- the importance information calculation unit 43 of the image processing apparatus of the present embodiment further includes the importance region calculation unit 52 .
- the importance region calculation unit 52 calculates the importance region 102 , based on the importance calculated for each pixel 101 by the importance calculation unit 51 .
- the gradation unit 61 calculates an amount of gradation in each pixel.
- the image processing apparatus 1 according to the first embodiment of the present invention has been described above.
- the image processing apparatus 1 according to the second embodiment of the present invention can have a hardware configuration and a functional configuration that are basically similar to those of the image processing apparatus 1 according to the first embodiment.
- FIG. 1 is also a block diagram showing the hardware configuration of the image processing apparatus 1 according to the second embodiment.
- FIGS. 8 , 9 and 10 are also flowcharts illustrating the ink-wash-painting-like image processing, the importance region calculation processing, and the ink-wash-painting-like image generation processing according to the second embodiment.
- FIG. 13 is a functional block diagram showing a functional configuration for executing overall gradation processing according to the second embodiment, regarding the functional configuration of the image processing apparatus 1 of the present invention.
- the functional configuration of the image processing unit 14 of the image processing apparatus 1 according to the second embodiment is basically similar to that of the image processing apparatus 1 according to the first embodiment, except in a case in which there are a plurality of importance regions, the plurality of importance regions are integrated as an integrated importance region; therefore, descriptions thereof are omitted.
- the gradation unit 61 of the image processing apparatus 1 of the first embodiment calculates an amount of gradation in each pixel, based on the importance of each pixel inside the importance region, and a distance between each pixel and the importance barycenter.
- the gradation unit 61 of the image processing apparatus 1 of the second embodiment integrates the plurality of importance regions as an integrated importance region, and calculates an amount of gradation inside such an importance region, based on processing similar to that in the first embodiment. In addition, outside the importance region, the gradation unit 61 of the image processing apparatus 1 of the second embodiment calculates an amount of gradation in each pixel, based on the importance of each pixel inside the integrated importance region, a distance between each pixel and the importance barycenter, and the importance barycenter as well as an edge direction of the integrated importance region.
- the units including the original image acquisition unit 41 to the ink-wash painting conversion unit 44 as well as the margin region setting unit 62 in the conversion unit 45 of the image processing apparatus 1 of the second embodiment are similar to the units including the original image acquisition unit 41 to the ink-wash painting conversion unit 44 as well as the margin region setting unit 62 in the conversion unit 45 of the image processing apparatus 1 of the first embodiment, respectively; therefore, detailed descriptions thereof are omitted, and only different points are described.
- FIG. 13 is a functional block diagram showing a functional configuration for executing overall gradation processing according to the second embodiment, regarding the functional configuration of the image processing apparatus shown in FIG. 1 .
- the overall gradation processing according to the second embodiment is executed similarly to the gradation processing of the first embodiment.
- the gradation unit 61 integrates the plurality of importance regions as an integrated importance region, and calculates an amount of gradation in each pixel, based on the importance of each pixel inside the integrated importance region, a distance between each pixel and the importance barycenter, and the importance barycenter as well as an edge direction of the integrated importance region.
- the gradation unit 61 executes gradation processing on the entire image of the ink-wash-painting-like image, based on the amount of gradation thus calculated.
- the gradation unit 61 further includes an importance region integration unit 81 for executing the overall gradation processing according to the second embodiment.
- the importance region integration unit 81 integrates the plurality of importance regions calculated by the importance region calculation unit 52 as an integrated importance region.
- the gradation unit 61 of the second embodiment calculates an amount of gradation in each pixel, based on the importance of each pixel in the importance region, and a distance between each pixel and the importance barycenter calculated by the importance barycenter calculation unit; and outside the importance region, the gradation unit 61 of the second embodiment calculates an amount of gradation in each pixel, based on the importance of each pixel inside the integrated importance region, a distance between each pixel and the importance barycenter calculated by the importance barycenter calculation unit, and the importance barycenter as well as an edge direction of the integrated importance region.
- FIG. 14 shows an example of data of an original image 100 of the second embodiment.
- gray regions are a plurality of importance regions 102 A and 102 B, among each of a plurality of pixels 101 composing the original image 100 .
- an integrated importance region 202 integrated from the plurality of importance regions 102 A and 102 B is shown in a black frame.
- an importance barycenter 211 calculated based on the integrated importance region 202 is shown in a substantially central portion of the integrated importance region 202 .
- FIG. 15 is a flowchart illustrating the overall gradation processing in Step S 16 , regarding the ink-wash-painting-like image processing shown in FIG. 8 .
- Step S 121 the gradation unit 61 acquires importance barycenter information calculated by the importance barycenter calculation unit 53 .
- Step S 122 the gradation unit 61 acquires information of the importance region calculated by the importance region calculation unit 52 .
- Step S 123 the gradation unit 61 determines whether there are plurality of importance regions calculated by the importance region calculation unit 52 .
- the processing advances to Step S 130 , and processing similar to the overall gradation processing in Steps S 83 to S 85 shown in FIG. 11 is executed.
- processing in Steps S 130 to S 132 of the second embodiment is similar to the processing in Steps S 83 to S 85 of the first embodiment, descriptions thereof are omitted.
- the processing advances to Step S 124 .
- Step S 124 the importance region integration unit 81 integrates the plurality of importance regions 102 A and 102 B shown in FIG. 14 as the integrated importance region 202 .
- Step S 125 the gradation unit 61 sets one pixel 101 in the integrated importance region 202 of the original image 100 shown in FIG. 14 as an attention pixel.
- Step S 126 the gradation unit 61 determines whether the attention pixel that was set in Step S 125 is a pixel in the importance region. For example, in the example shown in FIG. 14 , the gradation unit 61 determines whether the attention pixel is a pixel 101 in the importance region 102 A or 102 B. In a case in which the attention pixel is a pixel in the importance region, the processing advances to Step S 127 , and processing similar to the overall gradation processing in Step S 84 of the first embodiment shown in FIG. 11 is executed. In this case, since processing in Step S 127 of the second embodiment is similar to the processing in Step S 84 of the first embodiment, descriptions thereof are omitted. On the other hand, in a case in which it is determined that the attention pixel is not a pixel in the importance region, the processing advances to Step S 128 .
- Step S 128 the gradation unit 61 calculates an amount of gradation, based on the importance of the attention pixel, a distance from the importance barycenter acquired in Step S 121 , and the importance barycenter as well as an edge direction of the integrated importance region.
- Step S 129 the gradation unit 61 determines whether all pixels in the integrated importance region have been set as attention pixels. In a case in which all pixels in the integrated importance region have not been set as attention pixels in Step S 125 , the determination in Step S 129 is NO, and the processing returns to Step S 125 . In other words, the processing in Steps S 125 to S 129 is repeated until the amount of gradation is calculated for all pixels in the integrated importance region. On the other hand, in a case in which all pixels in the integrated importance region have been set as attention pixels in Step S 125 , the determination in Step S 129 is YES, and the processing advances to Step S 133 .
- Step S 133 the gradation unit 61 executes gradation processing of gradating the entire image of the original image, based on the gradation amount calculated for each pixel in Steps S 128 , S 128 and S 131 .
- the overall gradation processing is terminated, i.e. the processing in Step 16 shown in FIG. 8 is terminated, and the processing advances to Step S 17 .
- the gradation unit 61 of the image processing apparatus of the second embodiment configured as above includes the importance region integration unit 81 .
- the importance region integration unit 81 integrates the plurality of importance regions 102 A and 102 B shown in FIG. 14 calculated by the importance region calculation unit 52 as the integrated importance region 202 .
- the gradation unit 61 calculates an amount of gradation in each pixel 101 , based on the importance of each pixel 101 in the importance regions 102 A and 102 B, and a distance between each pixel 101 and the importance barycenter 211 calculated by the importance barycenter calculation unit 53 ; and outside the importance regions 102 A and 102 B, the gradation unit 61 calculates an amount of gradation in each pixel 101 , based on the importance of each pixel 101 inside the integrated importance region 202 , a distance between each pixel 101 and the importance barycenter 211 calculated by the importance barycenter calculation unit 53 , and the importance barycenter 211 as well as an edge direction of the integrated importance region 202 .
- an important pixel 101 being at a distance from the importance barycenter may be determined to be not very important although it is actually important.
- the gradation unit 61 integrates the plurality of importance regions 102 A and 102 B as the integrated importance region 202 .
- an object drawn as an ink-wash painting can be expressed based on a drawing method peculiar to the ink-wash painting.
- the importance calculation unit 51 calculates importance, based on color information, brightness (brightness information), or information of an edge direction, within a predetermined range around each pixel; however, it is not limited thereto. For example, a technique can also be employed, in which the importance calculation unit 51 uses a spatial frequency in a particular region for the attention pixel to increase importance of a corresponding portion. Moreover, regarding data of a single original image, the importance calculation unit 51 scans a face search frame in a predetermined direction, identifies characteristic portions (face parts) corresponding to eyes, nose, mouth and the like, and determines whether it is a face, based on a positional relationship of each face part.
- the importance calculation unit 51 can also calculate predetermined importance in accordance with classification of a characteristic portion thus determined. Furthermore, the importance calculation unit 51 can also identify a particular object in a scenery photograph or a portrait to calculate predetermined importance in accordance with classification of the object thus identified. Moreover, the importance calculation unit 51 calculates importance of each pixel, but it is not limited thereto, and can calculate importance of each region.
- the margin region setting unit 62 of the conversion unit 45 synthesizes the data of the background image and the data of the ink-wash-painting-like image, thereby setting the margin region to be added to the ink-wash-painting-like image; however, it is not limited thereto.
- the margin region setting unit 62 of the conversion unit 45 may further include an image region division unit (not shown) that divides the data of the ink-wash-painting-like image into a plurality of regions.
- the margin region setting unit 62 may not add a margin region to the periphery of such a region. In this way, as a result of not adding a margin region to the periphery of the region including a pixel with high importance, a natural composition of an ink-wash painting can be faithfully expressed. As a result, it is possible to obtain an image that is more similar to a real ink-wash painting.
- the amount of gradation calculated by the gradation unit 61 can be calculated so as to maximize an amount of gradation in the edge, by gradually changing a white-based contrast to zero, based on a predetermined function.
- the image processing apparatus of the present invention can be applied to electronic devices in general that can execute the aforementioned image processing. More specifically, for example, the present invention can be applied to a personal computer, a smart phone, a printer, a television, a video camera, a portable navigation device, a cell phone device, a portable game device, and the like.
- the processing sequence described above can be executed by hardware, and can also be executed by software.
- the hardware configuration shown in FIG. 2 is merely an illustrative example, and the present invention is not particularly limited thereto. More specifically, the types of functional blocks employed to realize the aforementioned functions are not particularly limited to the example in FIG. 2 , so long as the image processing apparatus 1 can be provided with the functions enabling the aforementioned processing sequence to be executed as its entirety.
- a single functional block may be configured by a single piece of hardware, a single installation of software, or any combination thereof.
- a program configuring the software is installed from a network or a storage medium into a computer or the like.
- the computer may be a computer embedded in dedicated hardware.
- the computer may be a computer capable of executing various functions by installing various programs, e.g., a general-purpose personal computer.
- the storage medium containing such a program can not only be constituted by the removable media 31 shown in FIG. 1 distributed separately from the device main body for supplying the program to a user, but can also be constituted by a storage medium or the like supplied to the user in a state incorporated in the device main body in advance.
- the removable media 31 is composed of a magnetic disk (including a floppy disk), an optical disk, a magnetic optical disk, or the like, for example.
- the optical disk is composed of a CD-ROM (Compact Disk-Read Only Memory), a DVD (Digital Versatile Disk), or the like, for example.
- the magnetic optical disk is composed of an MD (Mini-Disk) or the like.
- the storage medium supplied to the user in a state incorporated in the device main body in advance may include, for example, the ROM 12 shown in FIG. 1 , a hard disk included in the storage unit 19 shown in FIG. 1 or the like, in which the program is recorded.
- the steps describing the program recorded in the storage medium include not only the processing executed in a time series following this order, but also processing executed in parallel or individually, which is not necessarily executed in a time series.
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Image Processing (AREA)
- Processing Or Creating Images (AREA)
- Editing Of Facsimile Originals (AREA)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US15/005,623 US10311323B2 (en) | 2011-09-28 | 2016-01-25 | Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2011-213375 | 2011-09-28 | ||
JP2011213375A JP5807489B2 (ja) | 2011-09-28 | 2011-09-28 | 画像処理装置及び画像処理方法、並びにプログラム |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/005,623 Division US10311323B2 (en) | 2011-09-28 | 2016-01-25 | Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130077869A1 true US20130077869A1 (en) | 2013-03-28 |
Family
ID=47911362
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/615,521 Abandoned US20130077869A1 (en) | 2011-09-28 | 2012-09-13 | Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns |
US15/005,623 Expired - Fee Related US10311323B2 (en) | 2011-09-28 | 2016-01-25 | Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US15/005,623 Expired - Fee Related US10311323B2 (en) | 2011-09-28 | 2016-01-25 | Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns |
Country Status (3)
Country | Link |
---|---|
US (2) | US20130077869A1 (enrdf_load_stackoverflow) |
JP (1) | JP5807489B2 (enrdf_load_stackoverflow) |
CN (1) | CN103198498B (enrdf_load_stackoverflow) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150103257A1 (en) * | 2012-03-21 | 2015-04-16 | Seiko Epson Corporation | Image processing device, projector, and method of controlling projector |
US9762869B2 (en) | 2012-03-21 | 2017-09-12 | Seiko Epson Corporation | Image processing device, projector, and method of controlling projector |
US9767582B2 (en) * | 2015-05-05 | 2017-09-19 | Google Inc. | Painterly picture generation |
US9842416B2 (en) * | 2015-05-05 | 2017-12-12 | Google Llc | Animated painterly picture generation |
CN111753847A (zh) * | 2020-06-28 | 2020-10-09 | 浙江大华技术股份有限公司 | 图像预处理方法及装置、存储介质、电子装置 |
US11062422B2 (en) * | 2019-08-26 | 2021-07-13 | Ricoh Company, Ltd. | Image processing apparatus, image communication system, image processing method, and recording medium |
US11250295B2 (en) | 2019-01-24 | 2022-02-15 | Casio Computer Co., Ltd. | Image searching apparatus, classifier training method, and recording medium |
Families Citing this family (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109325499A (zh) * | 2018-08-02 | 2019-02-12 | 浙江中农在线电子商务有限公司 | 病虫害识别方法及装置 |
TWI796108B (zh) * | 2022-01-21 | 2023-03-11 | 廣達電腦股份有限公司 | 取得字體軌跡的方法及系統 |
Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4949279A (en) * | 1984-03-22 | 1990-08-14 | Sharp Kabushiki Kaisha | Image processing device |
US5038223A (en) * | 1988-02-29 | 1991-08-06 | Canon Kabushiki Kaisha | Image processing method and apparatus for imparting a pictorial or painter-like effect |
US20070127847A1 (en) * | 2005-12-07 | 2007-06-07 | Shinji Aoki | Image processing apparatus |
US20080297859A1 (en) * | 2007-05-31 | 2008-12-04 | Brother Kogyo Kabushiki Kaisha | Image-processing device |
US20080317348A1 (en) * | 1999-05-13 | 2008-12-25 | Canon Kabushiki Kaisha | Image processing apparatus, image reproduction apparatus, system, method and storage medium for image processing and image reproduction |
US7486296B2 (en) * | 2004-10-18 | 2009-02-03 | Reallusion Inc. | Caricature generating system and method |
US20110187732A1 (en) * | 2010-02-04 | 2011-08-04 | Casio Computer Co. Ltd. | Image processing device and non-transitory computer-readable storage medium |
US20110193876A1 (en) * | 2010-02-08 | 2011-08-11 | Casio Computer Co., Ltd. | Display processing apparatus |
US20110235905A1 (en) * | 2008-12-18 | 2011-09-29 | Yokokawa Masatoshi | Image processing apparatus and method, and program |
US20110249863A1 (en) * | 2010-04-09 | 2011-10-13 | Sony Corporation | Information processing device, method, and program |
US20110261066A1 (en) * | 2010-04-23 | 2011-10-27 | Casio Computer Co., Ltd. | Image processing apparatus and image processing method |
US20110292062A1 (en) * | 2010-05-28 | 2011-12-01 | Casio Computer Co., Ltd. | Image processing apparatus, method, and storage medium storing a program |
US20130028471A1 (en) * | 2011-07-26 | 2013-01-31 | Casio Computer Co., Ltd. | Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4112423A (en) * | 1976-09-13 | 1978-09-05 | Kelsey-Hayes Co. | Dual-screen data display terminal for data processing units |
JPS61221878A (ja) * | 1985-03-18 | 1986-10-02 | Sharp Corp | 画像処理装置 |
JPH01219965A (ja) | 1988-02-29 | 1989-09-01 | Canon Inc | 画像処理方法及びその装置 |
US20030165262A1 (en) * | 2002-02-21 | 2003-09-04 | The University Of Chicago | Detection of calcifications within a medical image |
US7379651B2 (en) * | 2003-06-10 | 2008-05-27 | Abu-Ageel Nayef M | Method and apparatus for reducing laser speckle |
JP4324043B2 (ja) * | 2004-07-15 | 2009-09-02 | キヤノン株式会社 | 画像処理装置およびその方法 |
US7602529B2 (en) * | 2004-09-07 | 2009-10-13 | Eastman Kodak Company | Method and system for controlling printer text/line art and halftone independently |
JP5088161B2 (ja) * | 2008-02-15 | 2012-12-05 | ソニー株式会社 | 画像処理装置、カメラ装置、通信システム、画像処理方法、およびプログラム |
US8270039B2 (en) * | 2009-01-05 | 2012-09-18 | Konica Minolta Laboratory U.S.A., Inc. | System and method for reducing toner usage in printing and cutting applications |
JP5434582B2 (ja) | 2009-12-28 | 2014-03-05 | カシオ計算機株式会社 | 画像処理装置、及び、画像加工プログラム |
JP5089713B2 (ja) * | 2010-01-18 | 2012-12-05 | シャープ株式会社 | 画像圧縮装置、圧縮画像出力装置、画像圧縮方法、コンピュータプログラム及び記録媒体 |
JP5556194B2 (ja) * | 2010-01-27 | 2014-07-23 | カシオ計算機株式会社 | 表示制御装置及び表示制御プログラム |
US8908206B2 (en) * | 2011-03-15 | 2014-12-09 | Xerox Corporation | Automated print job modification for reduced total and color page counts |
-
2011
- 2011-09-28 JP JP2011213375A patent/JP5807489B2/ja not_active Expired - Fee Related
-
2012
- 2012-09-13 US US13/615,521 patent/US20130077869A1/en not_active Abandoned
- 2012-09-26 CN CN201210364005.3A patent/CN103198498B/zh not_active Expired - Fee Related
-
2016
- 2016-01-25 US US15/005,623 patent/US10311323B2/en not_active Expired - Fee Related
Patent Citations (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4949279A (en) * | 1984-03-22 | 1990-08-14 | Sharp Kabushiki Kaisha | Image processing device |
US5038223A (en) * | 1988-02-29 | 1991-08-06 | Canon Kabushiki Kaisha | Image processing method and apparatus for imparting a pictorial or painter-like effect |
US20080317348A1 (en) * | 1999-05-13 | 2008-12-25 | Canon Kabushiki Kaisha | Image processing apparatus, image reproduction apparatus, system, method and storage medium for image processing and image reproduction |
US7486296B2 (en) * | 2004-10-18 | 2009-02-03 | Reallusion Inc. | Caricature generating system and method |
US20070127847A1 (en) * | 2005-12-07 | 2007-06-07 | Shinji Aoki | Image processing apparatus |
US20080297859A1 (en) * | 2007-05-31 | 2008-12-04 | Brother Kogyo Kabushiki Kaisha | Image-processing device |
US20110235905A1 (en) * | 2008-12-18 | 2011-09-29 | Yokokawa Masatoshi | Image processing apparatus and method, and program |
US20110187732A1 (en) * | 2010-02-04 | 2011-08-04 | Casio Computer Co. Ltd. | Image processing device and non-transitory computer-readable storage medium |
US20110193876A1 (en) * | 2010-02-08 | 2011-08-11 | Casio Computer Co., Ltd. | Display processing apparatus |
US20110249863A1 (en) * | 2010-04-09 | 2011-10-13 | Sony Corporation | Information processing device, method, and program |
US20110261066A1 (en) * | 2010-04-23 | 2011-10-27 | Casio Computer Co., Ltd. | Image processing apparatus and image processing method |
US20110292062A1 (en) * | 2010-05-28 | 2011-12-01 | Casio Computer Co., Ltd. | Image processing apparatus, method, and storage medium storing a program |
US20130028471A1 (en) * | 2011-07-26 | 2013-01-31 | Casio Computer Co., Ltd. | Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20150103257A1 (en) * | 2012-03-21 | 2015-04-16 | Seiko Epson Corporation | Image processing device, projector, and method of controlling projector |
US9426436B2 (en) * | 2012-03-21 | 2016-08-23 | Seiko Epson Corporation | Image processing device, projector, and method of controlling projector |
US9762869B2 (en) | 2012-03-21 | 2017-09-12 | Seiko Epson Corporation | Image processing device, projector, and method of controlling projector |
US9767582B2 (en) * | 2015-05-05 | 2017-09-19 | Google Inc. | Painterly picture generation |
US9842416B2 (en) * | 2015-05-05 | 2017-12-12 | Google Llc | Animated painterly picture generation |
US11250295B2 (en) | 2019-01-24 | 2022-02-15 | Casio Computer Co., Ltd. | Image searching apparatus, classifier training method, and recording medium |
US12277745B2 (en) | 2019-01-24 | 2025-04-15 | Casio Computer Co., Ltd. | Image searching apparatus, classifier training method, and recording medium |
US11062422B2 (en) * | 2019-08-26 | 2021-07-13 | Ricoh Company, Ltd. | Image processing apparatus, image communication system, image processing method, and recording medium |
CN111753847A (zh) * | 2020-06-28 | 2020-10-09 | 浙江大华技术股份有限公司 | 图像预处理方法及装置、存储介质、电子装置 |
Also Published As
Publication number | Publication date |
---|---|
CN103198498B (zh) | 2016-01-06 |
US20160140411A1 (en) | 2016-05-19 |
US10311323B2 (en) | 2019-06-04 |
JP5807489B2 (ja) | 2015-11-10 |
JP2013074538A (ja) | 2013-04-22 |
CN103198498A (zh) | 2013-07-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10311323B2 (en) | Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns | |
JP5463866B2 (ja) | 画像処理装置および画像処理方法、並びにプログラム | |
US10783691B2 (en) | Generating a stylized image or stylized animation by matching semantic features via an appearance guide, a segmentation guide, and/or a temporal guide | |
KR102304674B1 (ko) | 얼굴 표정 합성 방법과 장치, 전자 기기, 및 저장 매체 | |
CN111066060B (zh) | 虚拟面部化妆去除和模拟、快速面部检测和地标跟踪 | |
US11741639B2 (en) | Locating and augmenting object features in images | |
US8896622B2 (en) | Methods and apparatus for marker-based stylistic rendering | |
CN104798101B (zh) | 化妆辅助装置、化妆辅助方法以及化妆辅助程序 | |
KR20230097157A (ko) | 개인화된 3d 헤드 모델 변형을 위한 방법 및 시스템 | |
US10558849B2 (en) | Depicted skin selection | |
CN116648733A (zh) | 用于从面部图像提取颜色的方法和系统 | |
US8929649B2 (en) | Image processing apparatus for converting image in characteristic region of original image into image of brushstroke patterns | |
WO2019113888A1 (zh) | 一种基于交互式的中国水墨画风格笔画生成方法以及装置 | |
CN105574814A (zh) | 一种肖像剪纸特效的生成方法 | |
JP2014016688A (ja) | 顕著性マップを利用した非写実変換プログラム、装置及び方法 | |
JP5284994B2 (ja) | 画像処理装置 | |
US9959672B2 (en) | Color-based dynamic sub-division to generate 3D mesh | |
US9036938B2 (en) | Image processing apparatus, image processing method, and program | |
CN110516202B (zh) | 文档生成器的获取方法、文档生成方法、装置及电子设备 | |
JP2009253324A (ja) | 画像処理装置、画像処理方法、及びプログラム | |
JP6292010B2 (ja) | 画像処理装置 | |
CN108171766A (zh) | 一种具有笔画轮廓纠正功能的图像生成方法 | |
CN114820894A (zh) | 一种虚拟角色的生成方法及系统 | |
JP7675367B2 (ja) | 画像着色方法及び装置、着色画像生成モデルの生成装置、並びにプログラム | |
JP2023143446A (ja) | 情報処理装置の動作方法、情報処理装置、及びプログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: CASIO COMPUTER CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:TESHIMA, YOSHIHIRO;IWANAGA, MASAKUNI;REEL/FRAME:028959/0772 Effective date: 20120904 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |