JP5006263B2 - Image processing apparatus, program, and image processing method - Google Patents

Image processing apparatus, program, and image processing method Download PDF

Info

Publication number
JP5006263B2
JP5006263B2 JP2008145832A JP2008145832A JP5006263B2 JP 5006263 B2 JP5006263 B2 JP 5006263B2 JP 2008145832 A JP2008145832 A JP 2008145832A JP 2008145832 A JP2008145832 A JP 2008145832A JP 5006263 B2 JP5006263 B2 JP 5006263B2
Authority
JP
Japan
Prior art keywords
image data
image
processing
parameter value
function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
JP2008145832A
Other languages
Japanese (ja)
Other versions
JP2009296140A (en
Inventor
広文 西田
Original Assignee
株式会社リコー
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社リコー filed Critical 株式会社リコー
Priority to JP2008145832A priority Critical patent/JP5006263B2/en
Publication of JP2009296140A publication Critical patent/JP2009296140A/en
Application granted granted Critical
Publication of JP5006263B2 publication Critical patent/JP5006263B2/en
Application status is Active legal-status Critical
Anticipated expiration legal-status Critical

Links

Images

Description

  The present invention relates to an image processing apparatus, a program, and an image processing method.

  Conventionally, with the widespread use of color scanners and digital cameras, image data input devices and output devices may be different, for example, image data captured by a digital camera is output by a printing device. In the output device, for example, image data is output after correction according to the characteristics of the image data, such as correction of the background color. As described above, if the input device and the output device are different, the image data It may be difficult to identify features.

  In order to solve such a problem, an image processing apparatus that performs appropriate image processing on each image data is known (see, for example, “Patent Document 1”). In this apparatus, for example, undo (cancel) processing is performed on color document image data input from an image device such as a scanner in order to perform data management and procedures for generating image data optimal for various applications. ) And Redo (redo) process history and status.

  Also disclosed is an image processing apparatus that allows the user to visually and intuitively grasp the state transition by outputting the state transition of the image processing (see, for example, “Patent Document 2”).

JP 2006-053690 A JP 2006-074331 A

  Incidentally, not only the types of image data to be subjected to image processing, but also user preferences, purposes of using image data, and the like have been diversified. For example, the background treatment includes removal of the background to make it white regardless of the color of the background, and background cleaning to remove dirt and show-through while maintaining the original color. Which one to select depends on the user's preference.

  However, when the configuration is such that the user selects the contents of the image processing and its parameters one by one, there is a problem that the user's operation becomes complicated and the work efficiency also decreases.

  The present invention has been made in view of the above, and recommends an optimum processing parameter predicted for image data to a user, so that a user can perform a desired image with less operations (selection from a menu or parameter setting). An object is to provide an image processing apparatus, a program, and an image processing method.

In order to solve the above-described problems and achieve the object, an image processing apparatus according to a first aspect of the present invention is a combination of image data held in a case holding means and a processing parameter value indicating processing contents for the image data. There a function constructing means for constructing a function using the nearest neighbor method to calculate the optimum process parameter values for the unknown image data from case data will be additionally registered, an image data acquisition means for acquiring image data, acquired the aforementioned feature amount calculating means for calculating a feature quantity of the image data, and inputs the feature quantity of the image data obtained, the optimum the processing parameter value for the image data, the optimum parameter value calculating of calculating using said function And the function construction means has a property that is a state of change of processing parameter values around each case data, and Constructing the function in response to two factors contribution is the degree of influence on the prediction of the feature quantity in the entire example data, characterized in that.
According to a second aspect of the present invention, in the image processing apparatus according to the first aspect, the image data acquisition unit acquires the relationship between the image data stored in the case storage unit and the processing parameter value. When a change is detected in the relationship between the image data and the processing parameter value calculated by the optimum parameter value calculating unit, the image data and the processing parameter value in which the change is detected are additionally registered in the case holding unit. A change detecting means is further provided.

According to a third aspect of the present invention, in the image processing apparatus according to the first or second aspect , the property takes a larger absolute value when the process parameter value changes drastically around the predetermined case data. When the processing parameter value is almost constant, the value is close to 0.

According to a fourth aspect of the present invention, in the image processing apparatus according to the third aspect , when the property is 0, the processing parameter value is constant around the predetermined case data. Features.

According to a fifth aspect of the present invention, in the image processing apparatus according to the first aspect, the contribution takes a larger absolute value when the predetermined feature dimension has a large influence on the prediction, and has an influence on the prediction. When the influence is small, the value is close to 0.

The invention according to claim 6 is the image processing apparatus according to claim 5 , wherein when the degree of contribution is 0, the predetermined feature dimension does not contribute to the prediction at all.

According to a seventh aspect of the present invention, there is provided a program for an unknown computer from case data in which a combination of image data held in case holding means and a processing parameter value indicating processing contents for the image data is additionally registered . Function construction means for constructing a function using a nearest neighbor method for calculating an optimum processing parameter value for image data, image data acquisition means for obtaining image data, and feature quantities for calculating feature quantities of the obtained image data The function construction means is caused to function as a calculation means, and an optimum parameter value calculation means for calculating the optimum processing parameter value for the image data using the function, using the acquired feature amount of the image data as input. Is a property that is a state of change of processing parameter values around each case data, and the entire case data Kicking constructing the function in response to two factors contribution is the degree of influence on the prediction of the characteristic quantity, characterized in that.

An image processing method according to an eighth aspect of the present invention is an image processing method executed by an image processing apparatus, and the image processing apparatus includes a control unit and a storage unit, and is executed by the control unit. function building means, calculates the optimum processing parameter values from case data combination of process parameter values showing the contents of the image data and the image data held in the case holding means it will be additionally registered for the unknown image data The step of constructing a function using the nearest neighbor method, the step of acquiring the image data by the image data acquisition unit, the step of calculating the feature amount of the acquired image data by the feature amount calculation unit, and the optimum parameter value calculating means, and inputs the feature quantity of the image data obtained, the optimum the processing parameter value for the image data, the function The function construction means includes a property that is a state of change in processing parameter values around each case data, and an effect on the prediction of the feature amount in the whole case data. The function is constructed in accordance with two factors such as a degree of contribution.

  According to the present invention, the function used for calculating the optimum processing parameter value is expressed by the property that is the state of change of the processing parameter value around each case data and the degree of influence on the prediction of the feature amount in the entire case data. By constructing according to two factors of a certain degree of contribution, even if the applicable processing parameter takes a continuous value or there are many candidates, it is optimal for image data based on a function. Since the processing parameters can be calculated with high accuracy, the user can recommend the optimum processing parameters predicted for the image data, and the user can select a desired image with few operations (selection from the menu and parameter setting). There is an effect that it can be obtained.

  Exemplary embodiments of an image processing apparatus, a program, and an image processing method according to the present invention will be explained below in detail with reference to the accompanying drawings.

[First Embodiment]
A first embodiment of the present invention will be described with reference to FIGS. FIG. 1 is a block diagram showing electrical connections of the image processing apparatus 1 according to the first embodiment of the present invention. As shown in FIG. 1, an image processing apparatus 1 is a computer such as a PC (Personal Computer), and includes a CPU (Central Processing Unit) 2 that centrally controls each unit of the image processing apparatus 1 and a ROM ( Secondary storage such as primary storage device 5 such as Read Only Memory (RAM) 3 and RAM (Random Access Memory) 4 and HDD (Hard Disk Drive) 6 that is a storage unit for storing data files (for example, color bitmap image data). Information is transmitted by communication with other external computers via a network 7, a removable disk device 8 such as a CD-ROM drive for storing information, distributing information to the outside, and obtaining information from the outside, and a network 9. Network interface 10, CRT (Cathode Ray Tube), LCD (Liquid Crystal Display), etc. for displaying the process progress and results to the operator The display device 11 and a keyboard 12 for an operator to input commands and information to the CPU 2, a pointing device 13 such as a mouse, and the like. The data bus 14 arbitrates data transmitted and received between these components. Works.

  In the present embodiment, a general personal computer is applied as the image processing apparatus 1. However, the present invention is not limited to this, and a portable information terminal called PDA (Personal Digital Assistants). , PalmTopPC, mobile phone, PHS (Personal Handyphone System), etc.

  In such an image processing apparatus 1, when the user turns on the power, the CPU 2 activates a program called a loader in the ROM 3, loads a program for managing the hardware and software of the computer called the operating system from the HDD 6 into the RAM 4, and Start the system. Such an operating system starts a program, reads information, and performs storage according to a user operation. As typical operating systems, Windows (registered trademark), UNIX (registered trademark), and the like are known. An operation program running on these operating systems is called an application program.

  Here, the image processing apparatus 1 stores an image processing program in the HDD 6 as an application program. In this sense, the HDD 6 functions as a storage medium that stores the image processing program.

  In general, the application program installed in the secondary storage device 7 such as the HDD 6 of the image processing apparatus 1 is stored in an optical information recording medium such as a CD-ROM or DVD-ROM, or a magnetic medium such as an FD. The application program recorded on the medium 8 a and recorded on the storage medium 8 a is installed in the secondary storage device 7 such as the HDD 6. Therefore, the portable storage medium 8a such as an optical information recording medium such as a CD-ROM or a magnetic medium such as an FD can also be a storage medium for storing an image processing program. Further, the image processing program is stored on a computer connected to a network such as the Internet, and is installed in the secondary storage device 7 such as the HDD 6 by being downloaded from the outside via the network interface 10, for example. You may do it. The image processing program executed by the image processing apparatus 1 according to the present embodiment may be provided or distributed via a network such as the Internet.

  In the image processing apparatus 1, when an image processing program that operates on an operating system is started, the CPU 2 executes various arithmetic processes according to the image processing program and centrally controls each unit. Among various types of arithmetic processing executed by the CPU 2 of the image processing apparatus 1, image processing that is characteristic processing of the present embodiment will be described below.

  In addition, when real-time property is regarded as important, it is necessary to speed up the processing. For this purpose, it is desirable to separately provide a logic circuit (not shown) and execute various arithmetic processes by the operation of the logic circuit.

  Here, image processing executed by the CPU 2 of the image processing apparatus 1 will be described. FIG. 2 is a functional block diagram illustrating functions related to image processing executed by the CPU 2 of the image processing apparatus 1. The image processing apparatus 1 includes, as functional configurations, an image data acquisition unit 100, a feature amount calculation unit 101, a case database 102 as a case holding unit, an optimum parameter value calculation unit 103, a prediction function construction unit 104, a processing A content output unit 105, a designation receiving unit 106, an image processing unit 107, and a change detection unit 108 are provided.

  The image data acquisition unit 100 functions as an image data acquisition unit and acquires image data. Further, when the input image data is related to a document, the tilt of the document, that is, skew correction is performed.

  The feature amount calculation unit 101 functions as a feature amount calculation unit, and uses the image data acquired by the image data acquisition unit 100 as an input to calculate a feature amount in the entire image data. Here, the feature amount includes statistical information such as the ratio of characters in the entire image, the ratio of pictures in the entire image, the degree of scattering between characters and pictures, and the layout density. In addition, there are a spatial distribution of characters and pictures, a distribution of colors and edges, and a background color. As a feature amount calculation method, a method disclosed in Japanese Patent Application Laid-Open No. 2007-193528 can be used. In general, the input image is exclusively divided into rectangular blocks of the same size, and each block is classified into one of three types: “picture”, “character”, and “other”, and then all blocks are Based on the classification result, the image feature vector of the entire image is calculated. Furthermore, a statistic obtained from the color and luminance distribution, or a statistic obtained from the edge intensity distribution is added to constitute a multidimensional image feature vector. The feature amount calculation unit 101 outputs the feature amount (image feature amount vector) of the entire image to the case database 102 and the optimum parameter value calculation unit 103.

  The designation accepting unit 106 accepts designation of processing contents to be performed on the image data acquired by the image data acquiring unit 100. The processing content is designated by an input from the user via the keyboard 12, the pointing device 13, or the like. The processing content includes the type of processing and its parameters. Examples of the processing type include a background color correction process, a spatial filter process, and a resolution enlargement process.

As the background color correction processing, there are background removal excluding the background color and background cleaning for correcting the background color. Background removal is described in Japanese Patent Application Laid-Open Nos. 2004-320701 and 2005-110184. Here, assuming that a set of algorithms and parameters is A, the background color correction process is defined as follows.
A = {Skin removal, skin cleaning, do nothing}

Spatial filter processing includes smoothing processing, edge enhancement processing, and adaptive filtering over the entire processing target image. Here, adaptive filtering performs different processing for each pixel. Details are described in JP-A No. 2003-281526. Spatial filtering is defined as follows.
A = {smoothing process, edge enhancement process, adaptive filtering, nothing}

Examples of the resolution enlarging process include a process for enlarging the resolution of characters as described in Japanese Patent Application Laid-Open No. 2005-063055, and normal image interpolation. The resolution enlargement process is defined as follows.
A = {Character resolution enlargement, image interpolation, do nothing}

The case database 102 is specified by the user for the same image data via the specification receiving unit 106 and the feature amount (image feature amount vector) obtained by the feature amount calculation unit 101 for predetermined image data. Are stored for each user (with a user ID) in association with the processed contents. The case database 102 stores combinations of feature amounts (image feature amount vectors) and processing contents in a time series in the order in which the processing contents are designated. In other words, the case database 102 stores the history information H as expressed by Equation 1.

Here, x i is a feature amount (image feature amount vector) extracted from the i-th image data. f i is an algorithm or processing parameter suitable for the image data, that is, processing content. M is the number of cases.

Further, the feature amount (image feature amount vector) x i is expressed by Expression 2.

  Here, D is the number of feature dimensions.

  FIG. 3 is a diagram illustrating an example of history information registered in the case database 102. In the example shown in FIG. 3, the image data, the feature amount obtained from the image data, and the processing content designated for the image data are associated with each other. Further, as shown in FIG. 4, the processing content designated for the image data includes background color correction processing (background removal, background cleaning), spatial filter processing (smoothing processing, edge enhancement processing, adaptive filtering), The contents include resolution enlargement processing (character resolution enlargement, image interpolation) and the like. Further, as shown in FIG. 4, the processing content designated for the image data includes not only the processing itself but also parameters. The “background removal 3”, “edge enhancement 1”, and the like shown in FIG. 4 are parameters.

  The optimum parameter value calculation unit 103 functions as an optimum parameter value calculation unit, and receives the feature amount (image feature amount vector) output from the feature amount calculation unit 101, and details a prediction function construction described later. Using the prediction function constructed by the unit 104, an optimum processing parameter value (processing content) for the image data acquired by the image data acquisition unit 100 is calculated.

  The prediction function construction unit 104 functions as a function construction unit, and uses the data stored in the case database 102 to construct a prediction function used for calculating the optimum processing parameter value for unknown data.

  The processing content output unit 105 displays and presents the processing result to which the optimum processing parameter value (processing content), which is the recommended value calculated by the optimal parameter value calculation unit 103, is applied on the display device 11. The user inputs, via the keyboard 12 and the pointing device 13, whether he likes or dislikes the processing result obtained by applying the recommended value presented on the display device 11. If the user does not like the processing result to which the recommended value is applied, the user can input the processing parameter value (processing content) via the keyboard 12, the pointing device 13, etc. The specification of the processing content to be performed on the image data acquired by the data acquisition unit 100 is accepted. The user response result is output to the case database 102.

  The image processing unit 107 acquires the image data acquisition unit 100 according to the processing content selected by the designation receiving unit 106 or the optimal processing parameter value (processing content) that is the recommended value calculated by the optimal parameter value calculation unit 103. Image processing is performed on the processed image data.

  The change detection unit 108 registers in the case database 102 in a predetermined period, that is, the relationship between the image data registered in the case database 102 in the first period and the processing content, and a second period that is a period after the first period. The presence / absence of a change between the processed image data and the relationship between the processing contents is detected. And the change detection part 108 updates the content of the case database 102, when a change is detected. Note that the image data includes a feature amount (image feature amount vector) of the image data.

  FIG. 5 is a flowchart showing a history registration process in which the image data feature quantity (image feature quantity vector) and the processing content are registered in the case database 102 in the image processing apparatus 1.

  First, the image data acquisition unit 100 acquires image data (step S100). Next, the feature amount calculation unit 101 calculates the feature amount (image feature amount vector) of the image data acquired by the image data acquisition unit 100 (step S102).

  FIG. 6 is a flowchart showing detailed processing in the feature amount calculation processing (step S102). The feature amount calculation unit 101 exclusively divides the image data acquired from the image data acquisition unit 100 into rectangular blocks having the same size (step S110). Specifically, the image data is divided into blocks of the same size, for example, a 1 cm × 1 cm rectangle (80 pixels × 80 pixels if the resolution is 200 dpi, 120 pixels × height 120 pixels if the resolution is 300 dpi). .

Next, each block is classified into one of three types of “picture”, “character”, and “other” (step S112). Specifically, as shown in FIG. 7, first, an image I obtained by reducing a block image to be processed to a low resolution of about 100 dpi is generated (step S120), and the number L of resolution levels is set (step S120). In step S122, the resolution reduction level k is initialized (k ← 0) (step S124). The reason why the processes in steps S120 to S124 are performed is to extract features from an image with a further reduced resolution as well as an image I as shown in FIG. Although details will be described later, for example, when the resolution level number L 2, the image I, the images I 1 resolution 1/2, the resolution is the image I 2 1/4 image meter 3 Extract features from two images.

When the resolution reduction level k has not reached the resolution level number L (Yes in step S126), the image I k (k = 0,...) Obtained by reducing the resolution to 1/2 k from the image I generated in step S120. ., L) is generated (step S128). Next, the image I k is binarized (step S130). However, in a binary image, a black pixel has a value 1 and a white pixel has a value 0.

Then, from the image I k of binarized resolution 1/2 k, after calculating the feature vectors f k M-dimensional (step S132), the resolution reduction level k by "1" is incremented (k ← k + 1) (Step S134).

Here, a method for extracting features from an image obtained by binarizing the image I k (k = 0,..., L) will be described. The “higher order autocorrelation function (Nth order autocorrelation function)”, which is an extension of the autocorrelation function to the higher order (Nth order), indicates that the displacement direction (S 1 , S 2 ,..., S N ), defined by (Equation 3).

Here, the sum Σ is addition for the pixel r of the entire image. Therefore, an infinite number of high-order autocorrelation functions can be considered depending on the order and the direction of displacement (S 1 , S 2 ,..., S N ). Here, for simplicity, the order N of the higher-order autocorrelation function is set to “2”. Further, the displacement direction is limited to a local 3 × 3 pixel region around the reference pixel r. If equivalent features are removed by translation, the number of features is 25 in total for the binary image as shown in FIG. For the calculation of each feature, the product of the corresponding pixel values of the local pattern may be added to the entire image. For example, the feature corresponding to the local pattern of “No. 3” is calculated by taking the sum of products for the entire image of the gray value at the reference pixel r and the gray value at the point immediately adjacent to the reference pixel r. In this manner, an M = 25-dimensional feature vector f k is calculated from (Equation 4) from an image having a resolution of 1/2 k .

  The processes in steps S128 to S134 as described above are repeated until the resolution reduction level k incremented in step S18 exceeds the number L of resolution levels (No in step S126).

If incremented resolution reduction level k has exceeded the number of resolution levels L at step S134 (No in step S126), feature vectors f 0, · · ·, based on f L, the block, " It is classified into one of three types of picture, “character” and “other” (step 136).

Here, the block classification method will be described in detail. First, from the aforementioned M = 25-dimensional feature vector f k = (g (k, 1),..., G (k, 25)) (k = 0,..., L) to (25 × L ) Dimension feature vector x = (g (0,1),..., G (0,25),..., G (L, 1),. Generate. In order to perform classification using such a block feature quantity vector x, it is necessary to perform learning in advance. Therefore, in the present embodiment, the feature amount vector x is calculated by dividing the learning data into two types, one containing only characters and one not containing characters. Thereafter, the feature quantity vector p 0 of the character pixel and the feature quantity vector p 1 of the non-character pixel are calculated in advance by taking the respective averages. Then, if the feature vector x obtained from the block image to be classified is decomposed into a linear combination of the known feature vectors p 0 and p 1 , the coupling coefficients a 0 and a 1 become character pixels and non-characters. It represents the ratio of pixels or the “characteristic” and “non-characteristic” of the block. Such decomposition is possible because the feature based on the higher-order local autocorrelation is invariant to the position of the object in the screen, and is additive with respect to the number of objects. The feature vector x is decomposed as in (Equation 5).
Here, e is an error vector. F is defined by (Formula 6).
By the least square method, the optimum coupling coefficient vector a is given by (Equation 7).

Each block is classified into “picture”, “not a picture”, and “undecided” by performing threshold processing on the parameter a 1 representing “non-characteristic”. Each block is classified as “undecided” or “not a picture”, and is classified as “character” if the parameter a 0 representing the character character is greater than or equal to a threshold value, and “other” otherwise. FIG. 10 shows an example of block classification. In the example of FIG. 10, the black portion represents “character”, the gray portion represents “picture”, and the white portion represents “other”.

The description returns to FIG. 6 again. After classification into any of the three types of “picture”, “character” and “other”, about 20 image feature quantities are calculated based on the classification results of all blocks (step S114). Image features include, for example, the ratio of characters and pictures, the density ratio: how the layout is crowded (the degree to which it is packed in a narrow space), and the degree of scattering of letters and pictures: the characters and pictures are scattered throughout the paper. There is a degree. In this case, specifically, the following five values are calculated as image feature amounts. Note that the feature quantity calculation unit 101 calculates various image feature quantities in addition to this. The feature quantity calculation unit 101 extracts about 20 types, that is, about 20 dimensions. It is preferable to use as many feature quantities as possible from the viewpoint of creating a prediction function that selects appropriate process contents according to the history of process contents designation by various users.
1. Character ratio Rtε [0,1]: Ratio of blocks classified as “character” in all blocks 2. Non-character ratio Rpε [0, 1]: Ratio of blocks classified as “pictures” in all blocks. Layout density Dε [0, 1]: the sum of the area of the number of blocks of “character” and “picture” divided by the area of the drawing area 4. Character scattering degree St (> 0): the spatial distribution of the character blocks in the x and y directions, normalized by the area of the image, the determinant of the variance / covariance matrix Non-character scattering degree Sp (> 0): the spatial distribution of picture blocks in the x and y directions, normalized by the area of the image and the determinant of the variance / covariance matrix

  Next, the optimum parameter value calculation unit 103 receives the feature amount (image feature amount vector) output from the feature amount calculation unit 101 as an input, and uses the prediction function constructed by the prediction function construction unit 104 to generate an image data acquisition unit. The optimum processing parameter value (processing content) for the image data acquired in 100 is calculated (step S103).

However, the following four points are considered when constructing the prediction function.
1. The history information H stored in the case database 102 depends on individual users, tasks, and preferences of the image processing apparatus 1. Therefore, on-site learning is required.
2. In the image processing apparatus 1, it is assumed that the history information that can be used for learning is relatively small (several tens to hundreds). This is a condition resulting from the need to read user preferences and tasks from as little data as possible and adapt immediately.
3. The feature space (character ratio, etc.) is multidimensional (tens to hundreds). A feature selection mechanism for selecting only features suitable for prediction and removing disturbance factors, or weighting to each feature dimension is necessary. In addition, it is necessary to consider that the feature subsets suitable for prediction differ depending on individual selection targets and users.
4). Possible values of processing parameters are continuous or have a large number of candidates.

  On the other hand, the nearest neighbor method is suitable for on-site learning and is an identification method that does not assume the form of a probability distribution function. The nearest neighbor method is a prediction method using a past case that is most similar to the one currently being processed, and the prediction accuracy improves as the number of similar data increases.

  Furthermore, for problems with less learning data and multidimensional features, the distance measure in the nearest neighbor method is weighted according to the contribution to the prediction of each feature dimension, and learning data (combination of feature quantity and processing content) The dilemma between the number of data and the number of dimensions can be eliminated by weighting according to the importance of each. Therefore, in this embodiment, such a nearest neighbor method, that is, a nearest neighbor method incorporating weighted distance learning is used.

Here, the calculation of the optimum processing parameter value (processing content) in the present embodiment will be described in detail. As described above, the optimum processing parameter value (processing content) is calculated using the prediction function constructed by the prediction function construction unit 104. Specifically, the feature amount (image feature amount vector) calculated from the image data acquired by the image data acquisition unit 100 is set to y, and the processing parameter value (processing content) that is the i-th case with respect to Expression 8 shown below. ) The square Δ (y, x i ) of the distance to xi is calculated as shown in Equation 9 below.
However, w d (d = 1, 2,..., D) is a coefficient parameter determined for each dimension of the feature vector, and how to determine it will be described later. In the example, K pieces (K is a predetermined constant and a natural number) are selected in ascending order of Δ (y, x i ). A set that collects the indexes of the selected cases is represented as N (y), which is called “neighborhood of y”. For each element kεN (y) in the vicinity of y, using the feature quantity (image feature quantity vector) and processing parameter values (x k , f k ) stored in the case database 102 as case data, the feature quantity A prediction function g used to calculate a processing parameter value for unknown data represented by (image feature vector) y is calculated as shown in Equation 10 below, and output.
Here, v k (k = 1, 2,..., M) is a coefficient determined for each case data, and the determination method will be described later. #N (y) is the number in the vicinity of y.

As shown in Expression 10, the coefficient parameters used when calculating the processing parameter value for the unknown data y, that is, the coefficient parameter v i for the i-th case and the coefficient parameter w j for the j-th feature dimension are Learn automatically from data. The criterion of learning is the error evaluated by Leave-One-Out, that is, for each case data (x i , f i ), the value estimated from its neighborhood N (x i ) by Equation 4 and the true value f i (2) is added to all the case data i = 1, 2,..., M, and the sum J shown in the following equation is minimized.
The steepest descent method is used to calculate the evaluation value minimization.
here,
An initial value is set (for example, 1) for each of v j (j = 1, 2,..., M) and w d (d = 1, 2,..., D), and all v j and w to d converges, by equation 12 and equation 13, v j (j = 1,2 , ..., M) and w d (d = 1,2, ... , D) computation for updating the Repeat. That is, a parameter that defines the prediction function g for predicting the optimum processing parameter value is obtained by simple iterative calculation.

By this iterative calculation, v i and w j are obtained as follows.
(1) The absolute value of v j increases when the process parameter value changes drastically around the j-th case data, and takes a value close to 0 when the process parameter value is substantially constant. v j = 0 means that the processing parameter value is constant (f j ) around the case data (x j , f j ).
(2) if more larger effect of d-th feature dimension on prediction, the absolute value of w d takes a value greater than, when adversely affected is small, it takes a value close to 0. w d = 0 means that the feature dimension does not contribute to the prediction at all.

  That is, the prediction function g used for calculating the optimum processing parameter value is a state of change of the parameter values around each case data (the absolute value increases when the change is severe, and a value close to 0 when the change is almost constant). And the effect of each feature amount on the overall case data (the absolute value increases if the effect is large, and the value is close to 0 if the effect is small). According to the degree of contribution (importance), it is possible to predict with high accuracy.

  Through the processing as described above, the optimum processing parameter value (processing content) for the image data acquired by the image data acquisition unit 100 is calculated.

  In the subsequent step S104, the process waits for the output of the processing parameter value to the user and the input from the user. Here, FIG. 11 is a flowchart showing the flow of processing in step S104. As shown in FIG. 11, the image processing unit 107 executes image processing by applying the optimum processing parameter value (processing content) calculated in step S103 as a recommended value (step S201), and the processing content output unit 105. Displays the image processing result stored in the storage device on the display device 11 and presents it to the user (step S202).

  If the user does not like the processing result to which the recommended value is applied and the user re-enters the processing parameter value (processing content) via the keyboard 12, the pointing device 13, or the like, the designation receiving unit 106 The specification of the processing content is accepted (Yes in step S203), and the image processing unit 107 executes image processing according to the input processing parameter value (processing content) (step S204). Then, the processing content output unit 105 saves the result of the image processing with the applied recommended value or the result of the image processing with the input processing parameter value (processing content) in a storage device (for example, the RAM 4 or the HDD 6) ( Step S205).

  When the image processing is determined as described above, as shown in FIG. 5, when the change detection unit 108 detects a change between the image data and the relationship between the processing contents, the image feature of the image data and the user-specified processing are detected. The parameter value (processing content) is additionally registered in the case database 102 (step S105).

Finally, the prediction function construction unit 104 constructs a prediction function that is used to calculate the optimum processing parameter value for unknown data, using the data stored in the case database 102 (step S106). As described above, the prediction function construction unit 104, for each element kεN (y) in the vicinity of y, the feature quantity (image feature quantity vector) stored in the case database 102 as the case data and the processing parameter value ( x k , f k ) is used to construct a prediction function used to calculate a processing parameter value for unknown data represented by a feature quantity (image feature quantity vector) y.

  As described above, according to the present embodiment, the function used for calculating the optimum processing parameter value is used to predict the characteristic amount of the entire case data and the property that is the state of change of the processing parameter value around each case data. By constructing according to two factors of contribution, which is the degree of influence, based on the function even if the applicable processing parameter takes a continuous value or there are many candidates Since it is possible to calculate the optimum processing parameters for the image data with high accuracy, by recommending the optimum processing parameters predicted for the image data to the user, the number of operations (selection from the menu and parameter setting) is reduced. A desired image can be obtained.

  Further, by adding a processing example to a database and updating a prediction function according to a change in user's preference, high-precision prediction can be performed even on-site (on an actual machine in operation).

[Second Embodiment]
Next, a second embodiment of the present invention will be described with reference to FIGS. The same parts as those in the first embodiment described above are denoted by the same reference numerals, and description thereof is also omitted.

  The image processing apparatus 1 according to the second embodiment is different from the first embodiment in that the processing content prediction processing is performed in batch for a plurality of target image data. Yes.

  FIG. 12 is a functional block diagram illustrating functions related to image processing of the image processing apparatus 1 according to the second embodiment of the present invention. As shown in FIG. 12, the function of the image processing apparatus 1 according to the second embodiment includes a processing result holding unit 110 in addition to the function of the image processing apparatus 1 described in the first embodiment. I have. The processing result holding unit 110 holds the processing result of the image processing performed by the image processing unit 107.

  FIG. 13 is a flowchart showing the flow of processing in step S104. As shown in FIG. 13, the image processing unit 107 executes image processing by applying the optimum processing parameter value (processing content) calculated in step S103 as a recommended value (step S201), and the processing result holding unit 110. Stores the result of the image processing using the applied recommended value in a storage device (for example, the RAM 4 or the HDD 6) (step S301).

  In the present embodiment, the processing is repeated until the processing in steps S201 to S301 is completed for all image data (a plurality of input images) to be processed (Yes in step S302).

  When all the image data (a plurality of input images) to be processed is processed (Yes in step S302), the processing content output unit 105 sends the image processing result stored in the storage device to the display device 11 to the user. Display and present (step S202). Since step S203 and subsequent steps are the same as those in the first embodiment, description thereof will be omitted.

  As described above, according to the present embodiment, all image data (a plurality of input images) are batch-processed, the processing results are stored in the recording device, and all the image data (a plurality of input images) are stored. ), The processing results are collectively presented to the user. If the user accepts all results, the process is complete. On the other hand, if there is a process that the user does not like, the user respecifies the process parameter value and restarts the process. For the reprocessed image, the image feature value and the processing parameter value are accumulated in the history / case database, and the predictor is learned as necessary.

[Third Embodiment]
Next, a third embodiment of the present invention will be described with reference to FIG. In addition, the same part as 1st Embodiment mentioned above or 2nd Embodiment is shown with the same code | symbol, and description is also abbreviate | omitted.

  FIG. 14 is a perspective view showing the configuration of the multi-function device 50 according to the third embodiment of the present invention. In the first and second embodiments, a computer such as a PC is applied as the image processing apparatus 1. However, in this embodiment, an information processing apparatus provided in a digital multifunction peripheral or the like is applied as the image processing apparatus 1. Is.

  As illustrated in FIG. 14, the multi-function device 50 includes a scanner unit 51 that is an image reading device and a printer unit 52 that is an image printing device. The multi-function device 50 has the functions of the image processing apparatus 1 described in the other embodiments.

  More specifically, the image data acquisition unit 100 acquires a scan image read by the scanner unit 51 as image data and target image data, and performs various processes related to process content prediction.

  Other configurations and processes of the multi-function device 50 according to the third embodiment are the same as the configurations and processes of the image processing apparatus 1 according to the other embodiments.

[Fourth Embodiment]
Next, a fourth embodiment of the present invention will be described with reference to FIG. In addition, the same part as 1st Embodiment mentioned above or 2nd Embodiment is shown with the same code | symbol, and description is also abbreviate | omitted.

  FIG. 15 is a system configuration diagram showing an image processing system 60 according to the fourth embodiment of the present invention. In the first and second embodiments, a local system (for example, a personal computer alone) is applied as the image processing apparatus 1. However, in the present embodiment, a server that constitutes a server client system as the image processing apparatus 1. A computer is applied.

  As shown in FIG. 15, the image processing system 60 is a server client system, and a server computer S and a plurality of client computers C are connected via a network N. The server computer S performs processing in the image processing apparatus 1. Each client computer C transmits an image to the server computer S. The server computer S has each function of the image processing apparatus 1. Further, a network scanner NS is provided on the network N. The image data acquisition unit 100 of the server computer S acquires image data from each client computer S or network scanner NS.

  The case database 102 may be held in a server (not shown) other than the server computer S.

  Other configurations and processes of the image processing system 60 according to the fourth embodiment are the same as the configurations and processes of the image processing apparatus 1 according to the other embodiments.

1 is a block diagram showing electrical connections of an image processing apparatus according to a first embodiment of the present invention. It is a functional block diagram which shows the function concerning image processing. It is a figure which shows an example of the historical information registered into the case database. It is a figure which shows an example of the processing content registered into the case database. It is a flowchart which shows a history registration process. It is a flowchart which shows the detailed process in a feature-value calculation process. It is a flowchart which shows the detailed process in a classification | category process. It is a schematic diagram which shows multi-resolution processing. It is a figure which shows an example of the mask pattern for high-dimensional autocorrelation function calculation. It is a schematic diagram which shows the example of a block classification | category. It is a flowchart which shows the flow of a process of step S104. It is a functional block diagram which shows the function concerning the image processing of the image processing apparatus concerning the 2nd Embodiment of this invention. It is a flowchart which shows the flow of a process of step S104. It is a perspective view which shows the structure of the multifunctional device concerning the 3rd Embodiment of this invention. It is a figure which shows the whole structure of the image processing system concerning the 4th Embodiment of this invention.

Explanation of symbols

DESCRIPTION OF SYMBOLS 1 Image processing apparatus 100 Image data acquisition means 101 Feature-value calculation means 102 Case holding means 103 Optimal parameter value calculation means 104 Function construction means

Claims (8)

  1. Nearest neighbor method in which a combination of a processing parameter value showing the processing content for the image data and the image data held in the case holding means to calculate an optimum process parameter values for the unknown image data from case data will be additionally registered and functions construction means for constructing a function using,
    Image data acquisition means for acquiring image data;
    Feature amount calculating means for calculating the feature amount of the acquired image data;
    Optimum parameter value calculating means for calculating the optimum processing parameter value for the image data, using the obtained feature value of the image data as an input, and using the function;
    With
    The function construction means responds to two factors: a property that is a change state of processing parameter values around each case data, and a contribution that is a degree of influence on the prediction of the feature amount in the whole case data. To build the function,
    An image processing apparatus.
  2. The relationship between the image data acquired by the image data acquisition unit and the processing parameter value calculated by the optimum parameter value calculation unit with respect to the relationship between the image data and the processing parameter value held in the case holding unit A change detecting means for additionally registering the image data and the processing parameter value in which the change has been detected in the case holding means,
    The image processing apparatus according to claim 1.
  3. The property takes a larger absolute value when the process parameter value changes drastically around the predetermined case data, and takes a value close to 0 when the process parameter value is substantially constant.
    The image processing apparatus according to claim 1 , wherein the image processing apparatus is an image processing apparatus.
  4. When the property is 0, the processing parameter value is constant around the predetermined case data.
    The image processing apparatus according to claim 3 .
  5. The contribution degree takes a larger absolute value when the influence of the predetermined feature dimension on the prediction is large, and takes a value close to 0 when the influence on the prediction is small.
    The image processing apparatus according to claim 1 .
  6. If the contribution is 0, the predetermined feature dimension does not contribute to the prediction at all,
    The image processing apparatus according to claim 5.
  7. Computer
    A nearest neighbor method for calculating an optimum processing parameter value for unknown image data from case data in which a combination of image data held in case holding means and a processing parameter value indicating processing contents for the image data is additionally registered Function construction means for constructing the function used,
    Image data acquisition means for acquiring image data;
    Feature amount calculating means for calculating the feature amount of the acquired image data;
    Optimum parameter value calculating means for calculating the optimum processing parameter value for the image data, using the obtained feature value of the image data as an input, and using the function;
    Function as
    The function construction means responds to two factors: a property that is a change state of processing parameter values around each case data, and a contribution that is a degree of influence on the prediction of the feature amount in the whole case data. To build the function,
    A program characterized by that.
  8. An image processing method executed by an image processing apparatus,
    The image processing apparatus includes a control unit and a storage unit,
    Executed in the control unit,
    The function construction means calculates the optimum processing parameter value for the unknown image data from the case data in which the combination of the image data held in the case holding means and the processing parameter value indicating the processing content for the image data is additionally registered. Constructing a function using the nearest neighbor method,
    An image data obtaining unit obtaining image data;
    A feature amount calculating means calculating a feature amount of the acquired image data;
    An optimal parameter value calculating means, using the acquired feature amount of the image data as an input, and calculating the optimal processing parameter value for the image data using the function;
    Including
    The function construction means responds to two factors: a property that is a change state of processing parameter values around each case data, and a contribution that is a degree of influence on the prediction of the feature amount in the whole case data. To build the function,
    An image processing method.
JP2008145832A 2008-06-03 2008-06-03 Image processing apparatus, program, and image processing method Active JP5006263B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2008145832A JP5006263B2 (en) 2008-06-03 2008-06-03 Image processing apparatus, program, and image processing method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2008145832A JP5006263B2 (en) 2008-06-03 2008-06-03 Image processing apparatus, program, and image processing method

Publications (2)

Publication Number Publication Date
JP2009296140A JP2009296140A (en) 2009-12-17
JP5006263B2 true JP5006263B2 (en) 2012-08-22

Family

ID=41543952

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2008145832A Active JP5006263B2 (en) 2008-06-03 2008-06-03 Image processing apparatus, program, and image processing method

Country Status (1)

Country Link
JP (1) JP5006263B2 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10296933B2 (en) 2013-04-12 2019-05-21 Facebook, Inc. Identifying content in electronic images
US10163028B2 (en) * 2016-01-25 2018-12-25 Koninklijke Philips N.V. Image data pre-processing

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4118749B2 (en) * 2002-09-05 2008-07-16 株式会社リコー Image processing apparatus, image processing program, and storage medium
US20050206912A1 (en) * 2004-03-22 2005-09-22 Kabushiki Kaisha Toshiba Image processing apparatus
JP2006180189A (en) * 2004-12-22 2006-07-06 Fuji Xerox Co Ltd Color processing method, color processing apparatus, and lookup table
JP4768451B2 (en) * 2006-01-18 2011-09-07 株式会社リコー Image processing apparatus, image forming apparatus, program, and image processing method
JP4615462B2 (en) * 2006-03-15 2011-01-19 株式会社リコー Image processing apparatus, image forming apparatus, program, and image processing method

Also Published As

Publication number Publication date
JP2009296140A (en) 2009-12-17

Similar Documents

Publication Publication Date Title
Chetverikov A simple and efficient algorithm for detection of high curvature points in planar curves
KR101114135B1 (en) Low resolution ocr for camera acquired documents
US8013870B2 (en) Image masks generated from local color models
EP1168247A2 (en) Method for varying an image processing path based on image emphasis and appeal
US20040207600A1 (en) System and method for transforming an ordinary computer monitor into a touch screen
US7103218B2 (en) Image processing methods and apparatus for detecting human eyes, human face, and other objects in an image
US5828771A (en) Method and article of manufacture for determining whether a scanned image is an original image or fax image
US8812978B2 (en) System and method for dynamic zoom to view documents on small displays
EP0741487B1 (en) Method and apparatus for performing text/image segmentation
JP2004320701A (en) Image processing device, image processing program and storage medium
US8542923B2 (en) Live coherent image selection
US8112706B2 (en) Information processing apparatus and method
DE60114469T2 (en) Method and device for determining interesting images and for image transmission
US6807304B2 (en) Feature recognition using loose gray scale template matching
EP1703444B1 (en) Detecting an orientation of characters in a document image
US7978918B2 (en) Digital image cropping using a blended map
EP1999688B1 (en) Converting digital images containing text to token-based files for rendering
JP2007115193A (en) Electronic document comparison program, electronic document comparison device and electronic document comparison method
US8515208B2 (en) Method for document to template alignment
US8270771B2 (en) Iterative selection of pixel paths for content aware image resizing
KR20030011632A (en) Image processing method and apparatus using self-adaptive binarization
US20130208983A1 (en) Up-sampling binary images for segmentation
JP2007286864A (en) Image processor, image processing method, program, and recording medium
US7580571B2 (en) Method and apparatus for detecting an orientation of characters in a document image
JP2001109895A (en) Processing method for digital images

Legal Events

Date Code Title Description
A621 Written request for application examination

Free format text: JAPANESE INTERMEDIATE CODE: A621

Effective date: 20110112

A977 Report on retrieval

Free format text: JAPANESE INTERMEDIATE CODE: A971007

Effective date: 20120119

A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20120124

A521 Written amendment

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20120313

TRDD Decision of grant or rejection written
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20120508

A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20120524

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20150601

Year of fee payment: 3

R150 Certificate of patent or registration of utility model

Free format text: JAPANESE INTERMEDIATE CODE: R150