WO2017110640A1 - Image-processing device, image-processing method, and computer program - Google Patents

Image-processing device, image-processing method, and computer program Download PDF

Info

Publication number
WO2017110640A1
WO2017110640A1 PCT/JP2016/087324 JP2016087324W WO2017110640A1 WO 2017110640 A1 WO2017110640 A1 WO 2017110640A1 JP 2016087324 W JP2016087324 W JP 2016087324W WO 2017110640 A1 WO2017110640 A1 WO 2017110640A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
classification
processed
input image
file
Prior art date
Application number
PCT/JP2016/087324
Other languages
French (fr)
Japanese (ja)
Inventor
伊藤 直樹
Original Assignee
キヤノン株式会社
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by キヤノン株式会社 filed Critical キヤノン株式会社
Publication of WO2017110640A1 publication Critical patent/WO2017110640A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/10Image acquisition

Definitions

  • the present invention relates to an image processing apparatus, a control method thereof, and a computer program.
  • image data of the processing target form is acquired, feature data related to the processing target form is acquired, and feature data that matches or resembles the feature data is specified from the stored feature data (matching processing is performed) ) To estimate the form type. If the form type is not estimated, feature data is added (see).
  • the document recognition result is presented to the user every time, and the user is made to determine whether or not the recognition result is correct. Then, the determination result by the user is transmitted to the document recognition system.
  • the present invention solves the above-described problem, and provides an image processing apparatus capable of generating a document classification rule, which is used when performing a document identification process by a natural operation without requiring an explicit instruction from a user. For the purpose of provision.
  • the image processing apparatus determines a classification unit that classifies the input image to be processed using a feature amount and a classification rule acquired from the input image to be processed, and a result classified by the classification unit.
  • a storage unit for storing a file including the input image to be processed in the storage location, an acquisition unit for acquiring an operation history for the file stored by the storage unit, and an operation history acquired by the acquisition unit.
  • Generation means for regenerating the classification rule, wherein the classification means classifies the input image to be processed using the classification rule regenerated by the generation means. .
  • the image processing apparatus of the present invention it is possible to generate a document classification rule that is used when performing a document identification process by performing a natural operation unintentionally by a user.
  • FIG. 1 is an overview diagram showing a system configuration in Embodiment 1.
  • FIG. 1 is a block diagram illustrating a schematic configuration of an MFP according to a first embodiment.
  • 1 is a block diagram illustrating a schematic configuration of an MFP according to a first embodiment.
  • 1 is a block diagram illustrating a schematic configuration of a server in Embodiment 1.
  • FIG. 3 is a flowchart in the first embodiment.
  • FIG. 6 is a diagram illustrating calculation of an image feature amount based on gradient information in the first embodiment.
  • FIG. 6 is a diagram for explaining extraction of a patch image for calculating an image feature amount according to the first exemplary embodiment. It is a figure explaining the increase process of the learning data in Example 1.
  • FIG. It is a figure explaining the increase process of the learning data in Example 1.
  • FIG. It is a figure explaining the increase process of the learning data in Example 1.
  • FIG. It is a figure explaining the increase process of the learning data in Example 1.
  • FIG. 10 is a flow
  • FIG. 1 is a diagram illustrating a system configuration of the first embodiment.
  • the system according to the first embodiment includes an MFP 101 serving as an image input device and a server 102.
  • the MFP 103 is connected to the LAN 103.
  • the LAN 103 is connected to the Internet 104 and is also connected to the server 102 that provides the service via the Internet 104.
  • the MFP 101 and the server 102 are connected to each other via the LAN 103 and transmit / receive image data and various types of information.
  • the MFP 101 is connected to the LAN 103, but this is not a limitation.
  • the MFP 101 only needs to be connectable to the server 102.
  • the server 102 is an information processing apparatus that generates learning data used for learning classification rules for images input from the image input apparatus.
  • the “classification rule” is a combination of a feature amount acquired from an image and a classifier for classifying (identifying) the image.
  • a rule for classifying (identifying) the image.
  • “generate learning data” is to generate a document classification rule used when performing document identification processing.
  • the server 102 generates learning data in which the rotation direction of the image input from the image input device is unified, and constructs an image classification rule based on the generated learning data.
  • FIG. 2A and 2B are diagrams illustrating a configuration example of the MFP 101.
  • FIG. 2A and 2B are diagrams illustrating a configuration example of the MFP 101.
  • the MFP 101 includes a controller 20 to an operation unit 207.
  • the device control unit 200 exchanges data with the outside of the MFP 101 and the outside via the network I / F 206 and accepts an operation from the operation unit 207.
  • the image reading unit 201 reads a document and outputs image data.
  • the image processing unit 202 converts print information including image data input from the image reading unit 201 or the outside into intermediate information (hereinafter referred to as “object”), and stores it in the object buffer of the storage unit 203.
  • Objects have text, graphic, and image attributes.
  • bitmap data is generated based on the buffered object and stored in the buffer of the storage unit 203.
  • pseudo halftone processing such as color conversion processing, density adjustment processing, toner total amount control processing, video count processing, printer gamma correction processing, and dithering is performed.
  • the storage unit 203 includes a ROM, a RAM, a hard disk (HDD), and the like.
  • the ROM stores various control programs and image processing programs executed by the CPU 204.
  • the RAM is used as a reference area or work area in which the CPU 204 stores data and various types of information.
  • the RAM and HDD are used for the object buffer and the like. Image data is stored on the RAM and HDD, the pages are sorted, the originals that are sorted over a plurality of pages are stored, and a plurality of copies are printed out.
  • the image output unit 205 forms and outputs a color image on a recording medium such as recording paper.
  • a network I / F unit 206 connects the MFP 101 to the LAN 103 and transmits / receives various information to / from the Internet 104 and other devices.
  • the operation unit 207 includes a touch panel and operation buttons, receives an operation from the user, and transmits information on the operation to the device control unit 200.
  • FIG. 2B shows the appearance of the MFP 101.
  • the image reading unit 201 has a plurality of CCDs. If the sensitivity of each CCD is different, it is recognized that each pixel has a different density even if the density of each pixel on the document is the same. Therefore, the image reading unit first performs exposure scanning on a white plate (uniformly white plate), converts the amount of reflected light obtained by the exposure scanning into an electrical signal, and outputs it to the controller.
  • the shading correction unit in the image processing 202 recognizes the difference in sensitivity of each CCD based on the electrical signal obtained from each CCD. Then, using the recognized difference in sensitivity, the value of the electric signal obtained by scanning the image on the document is corrected.
  • the shading correction unit receives gain adjustment information from the CPU 204 in the controller, the shading correction unit performs gain adjustment according to the information.
  • the gain adjustment is used to adjust how the value of the electrical signal obtained by exposing and scanning the document is assigned to the luminance signal value of 0 to 255.
  • the value of the electrical signal obtained by exposing and scanning the document can be converted into a high luminance signal value or converted into a low luminance signal value. That is, the dynamic range of the read signal can be adjusted by this gain adjustment.
  • the image reading unit 201 converts the image information into an electrical signal by inputting the reflected light obtained by exposing and scanning the image on the document to the CCD. Further, the electrical signal is converted into a luminance signal composed of R, G, and B colors, and the luminance signal is output as an image to the controller 20.
  • the document is set on the tray 212 of the document feeder 211.
  • a document reading instruction is given from the controller 20 to the image reading unit 201.
  • the image reading unit 201 feeds originals one by one from the tray 212 of the original feeder 211, and performs an original reading operation.
  • the document reading method is not an automatic feeding method using the document feeder 211, but may be a method of scanning the document by placing the document on a glass surface (not shown) and moving the exposure unit.
  • the image output unit 205 is an image forming device that forms an image received from the controller 20 on a sheet.
  • the image forming method is an electrophotographic method using a photosensitive drum or a photosensitive belt, but the present embodiment is not limited to this.
  • the present invention can also be applied to an ink jet system that prints on paper by ejecting ink from a micro nozzle array.
  • the image output unit 205 is provided with a plurality of paper cassettes 213, 214, and 215 that can select different paper sizes or different paper orientations. The printed paper is discharged to the paper discharge tray 216.
  • FIG. 3 is a diagram illustrating a configuration example of the server.
  • the server 102 includes a CPU 301 to a data bus 306.
  • the CPU 301 reads a control program stored in ROM 303 and executes various control processes.
  • the RAM 302 is used as a temporary storage area such as a main memory or work area for the CPU 301.
  • a network I / F unit 304 connects the server 102 to the Internet 104 and transmits / receives various information to / from other devices.
  • the HDD 305 stores image data, feature amount data, and various programs.
  • the image data from the MFP 101 received via the network I / F unit 304 is sent and received by the CPU 301, RAM 302, and ROM 303 via the data bus 306.
  • the CPU 301 executes an image processing program stored in the ROM 303 or the HDD 305, image processing for the image data is realized.
  • FIG. 4 a classification process is performed on an image obtained by scanning a document, and a classification destination (storage destination) is determined according to the result. And it is a flowchart explaining the process which acquires the operation log (operation history) with respect to an image (file containing this image), and re-learns (regenerates) the image classification rule.
  • the processing shown in this flowchart is executed by the MFP 101 and the server 102.
  • the processes executed by the MFP 101 are realized by the CPU 204 loading and executing a processing program stored in the storage unit 203.
  • 4 is realized by the CPU 301 loading the processing program stored in the HDD 305 into the RAM 302 and executing the processing.
  • step S 401 when the MFP 101 receives a user instruction from the operation unit 207, the MFP 101 feeds documents one by one from the tray 212 of the document feeder 211, and scans the document by the image reading unit 201.
  • step S402 the CPU 204 executes a process for acquiring the feature amount of the input image obtained by scanning the document.
  • the feature amount of the image will be described later with reference to FIGS. 6, 7A, 7B, and 7C.
  • step S403 the CPU 204 stores the feature amount acquired in step S402 in the storage unit 203.
  • step S404 the CPU 204 performs a classification process according to the feature amount stored in the storage unit 203 in step S403 and the classification rule created by learning in advance.
  • the classification process for example, an ID number of a previously learned image (learned image) is associated with a feature amount acquired from the image, and this feature is obtained for an image having a certain feature amount. An ID number corresponding to the quantity is output. Therefore, when the input image is classified as the learning image ID1 as a result of the classification process, the corresponding image ID “1” is output. If the input image cannot be classified as a learned image, an image ID indicating an unknown form (unknown image) is output.
  • a known technique such as classification using machine learning can be applied.
  • step S405 the CPU 204 determines whether the input image in step S404 is classified as a learned image or an unlearned unknown image as a result of the classification.
  • the CPU 204 uses the image ID of the classification result described in step S404, it is determined whether the image is classified as a learned image or an unknown image. It is possible to apply a method that can determine whether the image is classified as a learned image or an unknown image by other methods.
  • step S ⁇ b> 406 the CPU 201 determines whether or not the transmission destination for saving the input image is a transmission destination that can acquire information on the operation of the file including the input image from the MFP 101. For example, the CPU 204 determines whether or not the transmission destination is beyond the firewall via the network I / F unit 206. If there is an input image transmission destination beyond the firewall, it is determined that there is no access right to the input image storage destination. Here, whether or not there is an access right indicates whether or not an operation log performed on a file in a save destination folder can be acquired.
  • step S407 the CPU 204 performs an input image storage process when it is determined in step S406 that the transmission destination has the access right.
  • the input image is stored in a folder named “unknown” (hereinafter referred to as an unknown folder).
  • step S408 the MFP 101 acquires an operation log for the file saved in step S407.
  • the operation log is information about operations performed to move folders to files, operations performed to change file names, and operations performed to delete files. It is.
  • step S409 the CPU 204 generates feedback information from the operation log acquired in step S408.
  • an operation log indicating where an image stored in an unknown folder is moved to is used to newly generate an image stored in the unknown folder.
  • assigning an image ID for example, using an operation log indicating that a first image stored in an unknown folder has been moved to a folder called Document 1, a document is applied to the first image. An image ID for entering 1 is generated and assigned. This assigned information is used as feedback information.
  • step S410 the CPU 204 performs a relearning process from the feedback information generated in step S409.
  • a threshold value update process used in performing the classification process is performed using the feature amount held in step S403 and the feedback information generated in step S409.
  • image ID information is acquired from the feedback information, and a threshold value calculation process is performed so that the image feature amount is classified into the image ID.
  • the updated threshold value is reflected in the classification process, and the classifier used for classification is updated (reproducing the document classification rule used when performing the document identification process).
  • step S411 if it is determined in step S406 that there is no access right for the storage destination of the input image, the CPU 204 stores the file in a storage destination having a separate access right. For example, the data is stored in the storage unit 203 of the MFP 101.
  • step S412 the MFP 101 acquires an operation log for the file saved in step S411. Since the acquisition of the operation log here is the same as that described in step S408, description thereof is omitted.
  • step S405 determines whether the image is a known image. If it is determined in step S405 that the image is a known image, the process proceeds to step S413. In step S413, the CPU 204 performs storage processing in accordance with a rule designated in advance.
  • FIG. 5 is a diagram for explaining a method for calculating an image feature amount based on gradient information.
  • the gradient strength and gradient direction calculated for each pixel in the patch image are used. Specifically, the CPU 301 obtains the gradient strength and gradient direction from the edge information in the vertical direction and the horizontal direction for all pixels in the patch image. The CPU 301 uses the gradient information to calculate 9-dimensional (9) feature amounts from one patch, as shown in FIG. First, for each pixel, a pixel having a gradient strength equal to or greater than a certain value is defined as an edge pixel, and a pixel having a gradient intensity smaller than the certain value is defined as a non-edge pixel.
  • the gradient direction is quantized into 8 directions from the edge pixel group, and the gradient intensity integrated value / patch pixel number for each direction is calculated, and combined with the non-edge pixel number / patch pixel number, a 9-dimensional feature from one patch image Calculate the amount.
  • the edge pixels and the non-edge pixels it is possible to express not only ruled line and character information but also a margin part which is a large feature of the document image.
  • the description so far is the description of the feature amount in one patch image, but actually, a large number of feature amounts are used by cutting out and using a plurality of patch images.
  • FIG. 6 is a diagram for explaining extraction of a patch image.
  • the CPU 301 deletes image edges (margins) where noise is likely to appear, and creates images with a plurality of resolutions.
  • the reason for preparing images with a plurality of resolutions is that the structure of the edge changes for each resolution.
  • the CPU 301 calculates a feature amount in consideration of the patch image position by cutting out patch images of a plurality of sizes from the respective resolution images. For example, assume that a feature amount is extracted from an image scanned at 300 dpi.
  • the CPU 301 creates two types of images obtained by reducing the scanned image to 1 ⁇ 4 size and 8 size.
  • the parameters relating to the image resolution, patch size, and patch cut-out position are not limited to the numbers described above. Further, in order to use document color information as an image feature amount to be acquired, a color histogram, color dispersion, or the like may be used as the image feature amount.
  • a learning data increasing process for increasing learning data when generating a direction discriminator using machine learning will be described.
  • a deformed image is obtained by performing a deformation process on the image by simulation, and this is increased as learning data.
  • FIG. 7A, 7B, and 7C are diagrams illustrating shift processing, rotation processing, and enlargement / reduction processing that are deformation processing. These geometric deformation processes are realized using a projective transformation matrix.
  • FIG. 7A shows the shift process. In the shift process, eight patterns of deformed images are obtained by moving the images in parallel by a fixed amount in the vertical and horizontal directions or in the upper left, upper right, lower left and lower right.
  • FIG. 7B shows the rotation process. In the rotation process, two patterns of deformed images are obtained by rotating clockwise and counterclockwise by a certain amount.
  • FIG. 7C shows the enlargement / reduction processing. In the enlargement / reduction process, two patterns of deformed images are obtained by enlarging or reducing the image by a predetermined magnification.
  • the input image and the output image have the same size.
  • the outside image area that protrudes outside the image area of the output image after projective transformation is discarded.
  • a missing region in which no projection source exists in the output image is complemented by sequentially copying pixel values of non-missing pixels.
  • the handling of this missing area is not limited to complementing by the method described above. For example, another complementing method that replaces the background pixel estimated from the input image may be used, or a method of adding flag information that the missing pixel is a missing pixel without performing the complementing may be used.
  • the learning data increase process by combining the patterns that are not deformed into the shift process, the rotation process, and the enlargement / reduction process, it is possible to obtain deformed images corresponding to the number of combinations from one image data.
  • 9 ⁇ 3 81 patterns of deformed images are generated to increase learning data. Note that the number of patterns of each deformation process is not limited to the above-described numbers.
  • Real AdaBoost is a method capable of selecting a feature amount suitable for classification of a given learning data set from a large amount of feature amounts and combining the feature amounts to configure a classifier. If a large amount of feature amount is used at the time of image classification, performance deteriorates due to the calculation load of the feature amount. Thus, it is a great advantage of Real AdaBoost that a classifier can be configured by selecting feature quantities suitable for classification and using only some of the feature quantities.
  • Real AdaBoost is a two-class classifier and classifies data with two types of labels. That is, as it is, it cannot be used for classification of three or more types of images. Therefore, a known method called OVA (One-Versus-All) that expands the two-class classifier to a multi-class classifier is used. OVA creates classifiers for classifying one class (target class) and other classes by the number of classes, and uses the output of each classifier as the reliability of the target class. At the time of classification, data to be classified is input to all classifiers, and the class having the highest reliability is set as the classification destination.
  • OVA One-Versus-All
  • FIG. 8 is a diagram for explaining an example of machine learning using learning data.
  • image feature amounts corresponding to each of three classes of images are prepared as learning data.
  • OVA prepares three types of classifiers.
  • the three types of classifiers are an image A discriminator for discriminating between image A and other images, an image B discriminator for discriminating between image B and other images, and an image B discriminating device for discriminating between image C and other images.
  • An image C discriminator is an image A discriminator for discriminating between image A and other images.
  • the image A discriminator outputs a large output value (confidence) when the image A is input, and outputs a small output value (confidence) when other images are input.
  • input document images are input to three types of classifiers, and output values are compared to determine which image. For example, when the output of the image B discriminator is maximum, it is determined that the input image is the image B.
  • the learning of the multi-class classifier using Real AdaBoost and OVA described with reference to FIG. 8 and the document image classification using the multi-class classifier are executed by the CPU 301.
  • the machine learning technique that can be used in the present embodiment is not limited to the technique described above. You may utilize well-known methods, such as Suppprot Vector Machine and Random Forest. If the feature selection framework is not included in the machine learning method and you want to improve the classification speed during classification, select a known feature value such as feature value selection using principal component analysis or discriminant analysis. Do.
  • a known method such as All-Versus-All (AVA) or Error- Correcting Output-Coding (ECOC) other than OVA may be used.
  • AVA All-Versus-All
  • ECOC Error- Correcting Output-Coding
  • an operation log for a user's file is acquired for an input image whose classification destination (storage destination) is unknown, and document identification processing is performed by using the acquired operation log. It is possible to generate (relearn) a document classification rule to be used when performing. That is, the classifier used for document classification is updated.
  • Example 2 In the first embodiment, the operation for the image stored in the unknown folder explained the folder moving process.
  • Example 2 it is assumed that a deletion operation is performed on an image stored in an unknown folder. Below, only the part which has a difference with Example 1 is demonstrated.
  • FIG. 9 is a flowchart for describing processing for detecting whether or not a file including an input image has been deleted from an operation log for an unclassified input image and generating feedback information in response to the deletion.
  • the process executed by the MFP 101 is realized by the CPU 204 loading and executing a processing program stored in the storage unit 203. Steps S401 to S413 are omitted since they have been described in the first embodiment.
  • step S901 the CPU 204 determines whether the file has been deleted from the log acquired in step S408 or step S412. Detection of whether or not a file has been deleted can be obtained by a known method.
  • step S902 when the CPU 204 determines that the operation for deleting the file has been performed in step S901, the CPU 204 generates a folder for deletion and stores the image therein.
  • the operation unit 207 is displayed for a user instruction.
  • what is displayed on the operation unit 207 is to cause the user to select whether to create a folder for deletion and move the file there, or delete the file without generating the folder for deletion.
  • step S903 the CPU 204 determines whether the user has input to the operation unit 207 when generating a deletion folder or not generating a deletion folder.
  • step S904 the CPU 204 generates a deletion folder when instructed to generate a deletion folder in step S903, and stores a file including an image to be deleted in the folder. Then, an image ID is assigned to the file so that the file including the image is classified into the deletion folder.
  • step S409 an image ID for classifying the processing target image into a folder for deletion is input.
  • step S903 if it is determined that the deletion folder is not generated, the CPU 204 deletes the file, and ends the process without generating feedback information.
  • an operation log in which an image in an unknown folder is deleted is acquired. For example, when an image stored in an unknown folder is deleted, it is considered that the image is an unnecessary image from the operation log. Therefore, an image ID that is unnecessary or classified as a trash can is assigned to the input image. As a result, it is possible to avoid unnecessary images from being classified into unknown folders many times, so that it is possible to reduce troublesome operations such as the user deleting images classified into unknown folders every time.
  • the present invention is also realized by executing the following processing. That is, software (computer program) that realizes the functions of the above-described embodiments is supplied to a system or apparatus via a network or various storage media, and the computer of the system or apparatus (or CPU, MPU, etc.) reads the program. To be executed.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Character Input (AREA)
  • Image Analysis (AREA)
  • Facsimiles In General (AREA)

Abstract

This image-processing device is characterized by having a classification means for classifying an input image to be processed using a feature value and classification rule acquired from the input image to be processed, a storage means for storing a file that includes the input image to be processed in a storage location determined on the basis of the result of classification by the classification means, an acquisition means for acquiring an operation history for the file stored by the storage means, and a generation means for regenerating the classification rule using the operation history acquired by the acquisition means, the classification means classifying the input image to be processed using the classification rule regenerated by the generation means.

Description

画像処理装置、画像処理方法、コンピュータプログラムImage processing apparatus, image processing method, and computer program
 本発明は、画像処理装置、その制御方法およびコンピュータプログラムに関する。 The present invention relates to an image processing apparatus, a control method thereof, and a computer program.
 スキャナ等を用いて、紙媒体として発生した帳票の画像データを生成して帳票の種別を識別し、OCR(Optical Character Recognition)処理を行う技術がある。このような技術では、帳票の種別を識別し、OCR処理を行うために、全ての帳票の種別について、帳票種別毎に予め書式定義(帳票定義)を作成しておく必要があった。しかし、このような書式定義を予め作成するには、帳票種別の識別サービスの利用者または提供者に手間をかけるものであった。そこで、予め書式定義を作成することなく帳票種別を推定するために、情報処理装置において帳票の外観上の特徴に係る特徴データを帳票種別毎に蓄積する。そして、処理対象帳票の画像データを取得して、その処理対象帳票に係る特徴データを取得し、蓄積されている特徴データから、特徴データに一致または類似する特徴データを特定する(マッチング処理を行う)ことで、帳票種別を推定する。また、帳票種別が推定されなかった場合は、特徴データを追加する(参照)。 There is a technology that uses a scanner or the like to generate image data of a form generated as a paper medium, identify the type of the form, and perform an OCR (Optical Character Recognition) process. In such a technique, in order to identify a form type and perform OCR processing, it is necessary to create a format definition (form definition) for each form type in advance for all form types. However, creating such a format definition in advance requires time and effort for the user or provider of the form type identification service. Therefore, in order to estimate the form type without creating a format definition in advance, the information processing apparatus accumulates feature data relating to the appearance characteristics of the form for each form type. Then, image data of the processing target form is acquired, feature data related to the processing target form is acquired, and feature data that matches or resembles the feature data is specified from the stored feature data (matching processing is performed) ) To estimate the form type. If the form type is not estimated, feature data is added (see).
特許第5670787号Patent No. 5670787
 従来技術のようにマッチング処理により文書の種別を認識する方法を利用する場合、文書認識結果を毎回ユーザーに提示し、ユーザーにこの認識結果に対して正しいか否かの判定を行わせる。そして、このユーザーによる判定結果を文書認識システムへ送信する。 When using a method for recognizing a document type by matching processing as in the conventional technology, the document recognition result is presented to the user every time, and the user is made to determine whether or not the recognition result is correct. Then, the determination result by the user is transmitted to the document recognition system.
 このように、文書認識処理を行うには、分類された文書に対して毎回分類結果が合っているか否かをユーザーに提示し、ユーザーからの判定結果の指示を受信することを必要とする。よって、ユーザーが判定結果の指示を行うまで文書認識が完了しない。 Thus, in order to perform document recognition processing, it is necessary to present to the user whether or not the classification result matches the classified document and to receive an instruction of the determination result from the user. Therefore, the document recognition is not completed until the user instructs the determination result.
 本発明は、上記課題を解決するものであり、ユーザーの明示的な指示を必要とせず、自然な操作で文書識別処理を行う際に用いられる、文書分類ルールの生成が可能な画像処理装置の提供を目的とする。 The present invention solves the above-described problem, and provides an image processing apparatus capable of generating a document classification rule, which is used when performing a document identification process by a natural operation without requiring an explicit instruction from a user. For the purpose of provision.
 本画像処理装置は、処理対象の入力画像から取得した特徴量と分類ルールを用いて前記処理対象の入力画像に対して分類を行う分類手段と、前記分類手段により分類された結果に基づいて決められた格納先に前記処理対象の入力画像を含むファイルを格納する格納手段と、前記格納手段により格納されたファイルに対する操作履歴を取得する取得手段と、前記取得手段により取得した操作履歴を用いて前記分類ルールを再生成する生成手段と、を有し、前記分類手段は、前記生成手段により再生成された分類ルールを用いて、処理対象の入力画像に対して分類を行うことを特徴とする。 The image processing apparatus determines a classification unit that classifies the input image to be processed using a feature amount and a classification rule acquired from the input image to be processed, and a result classified by the classification unit. A storage unit for storing a file including the input image to be processed in the storage location, an acquisition unit for acquiring an operation history for the file stored by the storage unit, and an operation history acquired by the acquisition unit. Generation means for regenerating the classification rule, wherein the classification means classifies the input image to be processed using the classification rule regenerated by the generation means. .
 本発明の画像処理装置によれば、ユーザーが意図せずに自然な操作をすることで文書識別処理を行う際に用いられる、文書分類ルールの生成が可能になる。 According to the image processing apparatus of the present invention, it is possible to generate a document classification rule that is used when performing a document identification process by performing a natural operation unintentionally by a user.
実施例1におけるシステム構成を示す概観図である。1 is an overview diagram showing a system configuration in Embodiment 1. FIG. 実施例1におけるMFPの概略構成を示すブロック図である。1 is a block diagram illustrating a schematic configuration of an MFP according to a first embodiment. 実施例1におけるMFPの概略構成を示すブロック図である。1 is a block diagram illustrating a schematic configuration of an MFP according to a first embodiment. 実施例1におけるサーバの概略構成を示すブロック図である。1 is a block diagram illustrating a schematic configuration of a server in Embodiment 1. FIG. 実施例1におけるフローチャートである。3 is a flowchart in the first embodiment. 実施例1における勾配情報に基づく画像特徴量の算出を説明する図である。FIG. 6 is a diagram illustrating calculation of an image feature amount based on gradient information in the first embodiment. 実施例1における画像特徴量の算出のためのパッチ画像の切り出しを説明する図である。FIG. 6 is a diagram for explaining extraction of a patch image for calculating an image feature amount according to the first exemplary embodiment. 実施例1における学習データの増加処理を説明する図である。It is a figure explaining the increase process of the learning data in Example 1. FIG. 実施例1における学習データの増加処理を説明する図である。It is a figure explaining the increase process of the learning data in Example 1. FIG. 実施例1における学習データの増加処理を説明する図である。It is a figure explaining the increase process of the learning data in Example 1. FIG. 実施例1における学習データ学習し分類処理を説明する図である。It is a figure explaining learning data in Example 1, and explaining classification processing. 実施例2におけるフローチャートである。10 is a flowchart in the second embodiment.
 (実施例1)
 図1は、実施例1のシステム構成を示す図である。
Example 1
FIG. 1 is a diagram illustrating a system configuration of the first embodiment.
 実施例1のシステムは、画像入力装置であるMFP101とサーバ102を備える。 The system according to the first embodiment includes an MFP 101 serving as an image input device and a server 102.
 LAN103には、MFP101が接続されている。また、LAN103はインターネット104に接続されており、サービスを提供しているサーバ102ともインターネット104経由で接続されている。MFP101、サーバ102は、LAN103を介して、互いに接続されており、画像データや各種情報の送受信を行う。この例では、MFP101はLAN103に接続されているが、その限りではない。MFP101はサーバ102に接続可能であればよい。なお、サーバ102は、画像入力装置から入力される画像の分類ルールの学習に用いる学習データを生成する情報処理装置である。 The MFP 103 is connected to the LAN 103. The LAN 103 is connected to the Internet 104 and is also connected to the server 102 that provides the service via the Internet 104. The MFP 101 and the server 102 are connected to each other via the LAN 103 and transmit / receive image data and various types of information. In this example, the MFP 101 is connected to the LAN 103, but this is not a limitation. The MFP 101 only needs to be connectable to the server 102. The server 102 is an information processing apparatus that generates learning data used for learning classification rules for images input from the image input apparatus.
 ここで、「分類ルール」とは、画像から取得された特徴量と画像を分類(識別)するための分類器との組合せのことである。ある特徴を有する画像は分類1へ、別の特徴を有する画像は分類2へと分類するため予め決まり(ルール)を設ける。 Here, the “classification rule” is a combination of a feature amount acquired from an image and a classifier for classifying (identifying) the image. In order to classify an image having a certain feature into category 1 and an image having another feature into category 2, a rule (rule) is provided in advance.
 また、「学習データを生成する」とは、文書識別処理を行う際に用いられる文書分類ルールを生成することである。 In addition, “generate learning data” is to generate a document classification rule used when performing document identification processing.
 具体的には、サーバ102は、画像入力装置から入力された画像の回転方向を統一した学習データを生成し、生成した学習データに基づいて、画像の分類ルールを構築する。 Specifically, the server 102 generates learning data in which the rotation direction of the image input from the image input device is unified, and constructs an image classification rule based on the generated learning data.
 図2A、2Bは、MFP101の構成例を示す図である。 2A and 2B are diagrams illustrating a configuration example of the MFP 101. FIG.
 図2Aに示すように、MFP101は、コントローラ20~操作部207を備える。 As shown in FIG. 2A, the MFP 101 includes a controller 20 to an operation unit 207.
 装置制御部200は、MFP101内およびネットワークI/F206を経由した外部とのデータの受け渡しや、操作部207からの操作の受け付けを行う。 The device control unit 200 exchanges data with the outside of the MFP 101 and the outside via the network I / F 206 and accepts an operation from the operation unit 207.
 画像読取部201は、原稿を読み取り画像データを出力する。 The image reading unit 201 reads a document and outputs image data.
 画像処理部202は、画像読取部201や外部から入力される画像データを含む印刷情報を中間情報(以下「オブジェクト」と呼ぶ)に変換し、記憶部203のオブジェクトバッファに格納する。オブジェクトは、テキスト、グラフィック、イメージの属性を持つ。さらに、バッファしたオブジェクトに基づきビットマップデータを生成し、記憶部203のバッファに格納する。その際、色変換処理、濃度調整処理、トナー総量制御処理、ビデオカウント処理、プリンタガンマ補正処理、ディザなどの疑似中間調処理を行う。 The image processing unit 202 converts print information including image data input from the image reading unit 201 or the outside into intermediate information (hereinafter referred to as “object”), and stores it in the object buffer of the storage unit 203. Objects have text, graphic, and image attributes. Further, bitmap data is generated based on the buffered object and stored in the buffer of the storage unit 203. At that time, pseudo halftone processing such as color conversion processing, density adjustment processing, toner total amount control processing, video count processing, printer gamma correction processing, and dithering is performed.
 記憶部203は、ROM、RAM、ハードディスク(HDD)などから構成される。ROMは、CPU204が実行する各種の制御プログラムや画像処理プログラムを格納する。RAMは、CPU204がデータや各種情報を格納する参照領域や作業領域として用いられる。また、RAMとHDDは、上記のオブジェクトバッファなどに用いられる。RAMとHDD上で画像データを蓄積し、ページのソートや、ソートされた複数ページにわたる原稿を蓄積し、複数部プリント出力を行う。 The storage unit 203 includes a ROM, a RAM, a hard disk (HDD), and the like. The ROM stores various control programs and image processing programs executed by the CPU 204. The RAM is used as a reference area or work area in which the CPU 204 stores data and various types of information. The RAM and HDD are used for the object buffer and the like. Image data is stored on the RAM and HDD, the pages are sorted, the originals that are sorted over a plurality of pages are stored, and a plurality of copies are printed out.
 画像出力部205は、記録紙などの記録媒体にカラー画像を形成して出力する。ネットワークI/F部206は、MFP101をLAN103に接続し、インターネット104や他の装置との間で各種情報を送受信する。 The image output unit 205 forms and outputs a color image on a recording medium such as recording paper. A network I / F unit 206 connects the MFP 101 to the LAN 103 and transmits / receives various information to / from the Internet 104 and other devices.
 操作部207は、タッチパネルや操作ボタンを備え、ユーザーからの操作を受け付けて装置制御部200へ該操作の情報を送信する。 The operation unit 207 includes a touch panel and operation buttons, receives an operation from the user, and transmits information on the operation to the device control unit 200.
 図2Bは、MFP101の外観を示す。画像読取部201は、複数のCCDを有している。この各CCDの感度が夫々異なっていると、たとえ原稿上の各画素の濃度が同じであったとしても、各画素が夫々違う濃度であると認識されてしまう。そのため、画像読取部では、最初に白板(一様に白い板)を露光走査し、露光走査して得られた反射光の量を電気信号に変換してコントローラに出力している。なお、画像処理202内のシェーディング補正部は、各CCDから得られた電気信号を元に、各CCDの感度の違いを認識している。そして、この認識された感度の違いを利用して、原稿上の画像をスキャンして得られた電気信号の値を補正している。さらに、シェーディング補正部は、コントローラ内のCPU204からゲイン調整の情報を受取ると、当該情報に応じたゲイン調整を行う。ゲイン調整は、原稿を露光走査して得られた電気信号の値を、どのように0~255の輝度信号値に割り付けるかを調整するために用いられる。このゲイン調整により、原稿を露光走査して得られた電気信号の値を高い輝度信号値に変換したり、低い輝度信号値に変換したりすることができるようになっている。つまり、このゲイン調整により、読み取り信号のダイナミックレンジの調整が可能である。続いて、この原稿上の画像をスキャンする構成について説明する。 FIG. 2B shows the appearance of the MFP 101. The image reading unit 201 has a plurality of CCDs. If the sensitivity of each CCD is different, it is recognized that each pixel has a different density even if the density of each pixel on the document is the same. Therefore, the image reading unit first performs exposure scanning on a white plate (uniformly white plate), converts the amount of reflected light obtained by the exposure scanning into an electrical signal, and outputs it to the controller. The shading correction unit in the image processing 202 recognizes the difference in sensitivity of each CCD based on the electrical signal obtained from each CCD. Then, using the recognized difference in sensitivity, the value of the electric signal obtained by scanning the image on the document is corrected. Further, when the shading correction unit receives gain adjustment information from the CPU 204 in the controller, the shading correction unit performs gain adjustment according to the information. The gain adjustment is used to adjust how the value of the electrical signal obtained by exposing and scanning the document is assigned to the luminance signal value of 0 to 255. By this gain adjustment, the value of the electrical signal obtained by exposing and scanning the document can be converted into a high luminance signal value or converted into a low luminance signal value. That is, the dynamic range of the read signal can be adjusted by this gain adjustment. Next, a configuration for scanning the image on the document will be described.
 画像読取部201は、原稿上の画像を露光走査して得られた反射光をCCDに入力することで画像の情報を電気信号に変換する。さらに電気信号をR,G,B各色からなる輝度信号に変換し、当該輝度信号を画像としてコントローラ20に対して出力する。 The image reading unit 201 converts the image information into an electrical signal by inputting the reflected light obtained by exposing and scanning the image on the document to the CCD. Further, the electrical signal is converted into a luminance signal composed of R, G, and B colors, and the luminance signal is output as an image to the controller 20.
 なお、原稿は原稿フィーダ211のトレイ212にセットされる。ユーザーが操作部207から読み取り開始を指示すると、コントローラ20から画像読取部201に原稿読み取り指示が与えられる。画像読取部201は、この指示を受けると原稿フィーダ211のトレイ212から原稿を1枚ずつフィードして、原稿の読み取り動作を行う。なお、原稿の読み取り方法は原稿フィーダ211による自動送り方式ではなく、原稿を不図示のガラス面上に載置し露光部を移動させることで原稿の走査を行う方法であってもよい。 Note that the document is set on the tray 212 of the document feeder 211. When the user gives an instruction to start reading from the operation unit 207, a document reading instruction is given from the controller 20 to the image reading unit 201. Upon receiving this instruction, the image reading unit 201 feeds originals one by one from the tray 212 of the original feeder 211, and performs an original reading operation. The document reading method is not an automatic feeding method using the document feeder 211, but may be a method of scanning the document by placing the document on a glass surface (not shown) and moving the exposure unit.
 画像出力部205は、コントローラ20から受取った画像を用紙上に形成する画像形成デバイスである。なお、本実施例において画像形成方式は感光体ドラムや感光体ベルトを用いた電子写真方式となっているが、本実施例はこれに限られることはない。例えば、微少ノズルアレイからインクを吐出して用紙上に印字するインクジェット方式などでも適用可能である。また、画像出力部205には、異なる用紙サイズ又は異なる用紙向きを選択可能とする複数の用紙カセット213、214、215が設けられている。排紙トレイ216には印字後の用紙が排出される。 The image output unit 205 is an image forming device that forms an image received from the controller 20 on a sheet. In this embodiment, the image forming method is an electrophotographic method using a photosensitive drum or a photosensitive belt, but the present embodiment is not limited to this. For example, the present invention can also be applied to an ink jet system that prints on paper by ejecting ink from a micro nozzle array. The image output unit 205 is provided with a plurality of paper cassettes 213, 214, and 215 that can select different paper sizes or different paper orientations. The printed paper is discharged to the paper discharge tray 216.
 図3は、サーバの構成例を示す図である。 FIG. 3 is a diagram illustrating a configuration example of the server.
 サーバ102は、CPU301~データバス306を備える。 The server 102 includes a CPU 301 to a data bus 306.
 CPU301は、ROM303に記憶された制御プログラムを読み出して各種制御処理を実行する。RAM302は、CPU301の主メモリ、ワークエリア等の一時記憶領域として用いられる。ネットワークI/F部304は、サーバ102をインターネット104に接続し、他の装置との間で各種情報を送受信する。HDD305は、画像データや特徴量データ、各種プログラムを記憶する。 CPU 301 reads a control program stored in ROM 303 and executes various control processes. The RAM 302 is used as a temporary storage area such as a main memory or work area for the CPU 301. A network I / F unit 304 connects the server 102 to the Internet 104 and transmits / receives various information to / from other devices. The HDD 305 stores image data, feature amount data, and various programs.
 図3において、ネットワークI/F部304を介して受信したMFP101からの画像データを、データバス306を介してCPU301、RAM302、ROM303が送受する。CPU301がROM303やHDD305に格納された画像処理プログラムを実行することによって、画像データに対する画像処理が実現される。 3, the image data from the MFP 101 received via the network I / F unit 304 is sent and received by the CPU 301, RAM 302, and ROM 303 via the data bus 306. When the CPU 301 executes an image processing program stored in the ROM 303 or the HDD 305, image processing for the image data is realized.
 <フローチャートを用いた本実施例の詳細説明>
 図4は、原稿をスキャンして得られた画像に対して分類処理を行い、その結果に応じて分類先(格納先)を決定する。そして、画像(この画像が含まれたファイル)に対する操作ログ(操作履歴)を取得し、画像分類ルールを再学習(再生成)する処理を説明するフローチャートである。
<Detailed description of this embodiment using flowchart>
In FIG. 4, a classification process is performed on an image obtained by scanning a document, and a classification destination (storage destination) is determined according to the result. And it is a flowchart explaining the process which acquires the operation log (operation history) with respect to an image (file containing this image), and re-learns (regenerates) the image classification rule.
 このフローチャートに示される処理は、MFP101およびサーバ102にて実行される。図4のフローチャートに示される処理のうちMFP101が実行する処理は、CPU204が記憶部203に格納されている処理プログラムをロードして実行することで実現される。また、図4のフローチャートに示される処理のうちサーバ102が実行する処理は、CPU301が、HDD305に格納されている処理プログラムをRAM302にロードして実行することで実現される。 The processing shown in this flowchart is executed by the MFP 101 and the server 102. Of the processes shown in the flowchart of FIG. 4, the processes executed by the MFP 101 are realized by the CPU 204 loading and executing a processing program stored in the storage unit 203. 4 is realized by the CPU 301 loading the processing program stored in the HDD 305 into the RAM 302 and executing the processing.
 ステップS401において、MFP101は、操作部207からユーザーの指示を受け付けると、原稿フィーダ211のトレイ212から原稿を1枚ずつフィードして、画像読取部201で原稿をスキャンする。 In step S 401, when the MFP 101 receives a user instruction from the operation unit 207, the MFP 101 feeds documents one by one from the tray 212 of the document feeder 211, and scans the document by the image reading unit 201.
 ステップS402において、CPU204は、原稿をスキャンすることで得られた入力画像の特徴量の取得処理を実行する。画像の特徴量に関しては、図6、図7A、7B,7Cを用いて後述する。 In step S402, the CPU 204 executes a process for acquiring the feature amount of the input image obtained by scanning the document. The feature amount of the image will be described later with reference to FIGS. 6, 7A, 7B, and 7C.
 ステップS403において、CPU204はステップS402で取得した特徴量を記憶部203へ格納する。 In step S403, the CPU 204 stores the feature amount acquired in step S402 in the storage unit 203.
 ステップS404において、CPU204は、ステップS403で記憶部203に格納している特徴量と予め学習して作成された分類ルールに沿って分類処理を行う。分類処理の結果として、例えば、予め学習されている画像(学習済画像)のID番号と画像から取得される特徴量とが紐づけられており、ある特徴量を有する画像に対して、この特徴量に対応するID番号が出力される。よって、分類処理した結果、入力画像が学習画像のID1の画像である分類された場合、該当する画像IDである“1”を出力する。また、入力画像が学習済画像に分類できない場合には、未知帳票(未知画像)であることを示す画像IDを出力する。ここで分類処理に関しては、機械学習を利用した分類など、公知の技術を適用することが可能である。 In step S404, the CPU 204 performs a classification process according to the feature amount stored in the storage unit 203 in step S403 and the classification rule created by learning in advance. As a result of the classification process, for example, an ID number of a previously learned image (learned image) is associated with a feature amount acquired from the image, and this feature is obtained for an image having a certain feature amount. An ID number corresponding to the quantity is output. Therefore, when the input image is classified as the learning image ID1 as a result of the classification process, the corresponding image ID “1” is output. If the input image cannot be classified as a learned image, an image ID indicating an unknown form (unknown image) is output. Here, with regard to the classification process, a known technique such as classification using machine learning can be applied.
 ステップS405において、CPU204は、ステップS404の入力画像に対して分類をした結果、学習済画像に分類されるのか、学習していない未知画像に分類されるのかの判断を行う。ここではステップS404で説明した分類結果の画像IDを用いて、学習済画像に分類されるのか、未知画像に分類されるかの判断を行う。これ以外の方法により、学習済画像に分類されるか、未知画像に分類されるかを判断できる方法の適用が可能である。 In step S405, the CPU 204 determines whether the input image in step S404 is classified as a learned image or an unlearned unknown image as a result of the classification. Here, using the image ID of the classification result described in step S404, it is determined whether the image is classified as a learned image or an unknown image. It is possible to apply a method that can determine whether the image is classified as a learned image or an unknown image by other methods.
 ステップS406においてCPU201は、入力画像を保存するべく送信する送信先がMFP101から入力画像を含んだファイルを操作した情報が取得可能な送信先であるか否かの判断を行う。例えば、CPU204は、ネットワークI/F部206を介して送信先がファイヤーウォールを越えた先にあるか否かの判断を行う。ファイヤーウォールを越えた先に入力画像の送信先がある場合は、入力画像の保存先にアクセス権がないと判断をする。ここでアクセス権があるか否かについては、保存先のフォルダでファイルに対して行われた操作ログが取得できるか出来ないか否かを指している。 In step S <b> 406, the CPU 201 determines whether or not the transmission destination for saving the input image is a transmission destination that can acquire information on the operation of the file including the input image from the MFP 101. For example, the CPU 204 determines whether or not the transmission destination is beyond the firewall via the network I / F unit 206. If there is an input image transmission destination beyond the firewall, it is determined that there is no access right to the input image storage destination. Here, whether or not there is an access right indicates whether or not an operation log performed on a file in a save destination folder can be acquired.
 ここでは、保存先のフォルダに対してファイル操作ログが取得できるか否かを検出する公知の方法を適用することが可能である。 Here, it is possible to apply a publicly known method for detecting whether or not a file operation log can be acquired for a storage destination folder.
 ステップS407においてCPU204は、ステップS406で送信先にアクセス権があると判定された場合に入力画像の保存処理を行う。ここで説明する一例として、入力画像の保存先にアクセス権があると判定され、このアクセス権があると判定された場所に、分類処理にて分類ができなかった(画像が未知であった)画像を保存するためのフォルダへ保存される。具体例としては、“未知”という名前のフォルダ(以下、未知フォルダとする)に入力画像が保存される。 In step S407, the CPU 204 performs an input image storage process when it is determined in step S406 that the transmission destination has the access right. As an example to be described here, it is determined that the input image storage destination has access right, and the classification processing could not be performed at the location determined to have this access right (the image was unknown). Saved to a folder for saving images. As a specific example, the input image is stored in a folder named “unknown” (hereinafter referred to as an unknown folder).
 ステップS408においてMFP101は、ステップS407で保存したファイルに対する操作ログの取得を行う。ここで、操作ログとは、ファイルに対してフォルダを移動するために行われた操作や、ファイル名を変更するために行われた操作や、ファイルを削除するために行われた操作についての情報である。具体例としては、未知フォルダに保存されている画像が、どこのフォルダへ移動するよう行われた操作の操作ログを取得する。 In step S408, the MFP 101 acquires an operation log for the file saved in step S407. Here, the operation log is information about operations performed to move folders to files, operations performed to change file names, and operations performed to delete files. It is. As a specific example, an operation log of an operation in which an image stored in an unknown folder is moved to which folder is acquired.
 ステップS409においてCPU204は、ステップS408で取得した操作ログからフィードバック情報を生成する。フィードバック情報を生成するとは、未知フォルダに保存された画像がどこのフォルダへ移動するよう操作が行われたのかを示す操作ログを用いて、未知フォルダに保存されていた画像に対して、新たに画像IDを付与することである具体例として例えば、未知フォルダに保存されている第1の画像をドキュメント1というフォルダへ移動したことを示す操作ログを用いて、第1の画像に対して、ドキュメント1に入るための画像IDを生成して付与する。この付与された情報をフィードバック情報とする。 In step S409, the CPU 204 generates feedback information from the operation log acquired in step S408. When generating feedback information, an operation log indicating where an image stored in an unknown folder is moved to is used to newly generate an image stored in the unknown folder. As a specific example of assigning an image ID, for example, using an operation log indicating that a first image stored in an unknown folder has been moved to a folder called Document 1, a document is applied to the first image. An image ID for entering 1 is generated and assigned. This assigned information is used as feedback information.
 ステップS410においてCPU204は、ステップS409で生成されたフィードバック情報から再学習処理を行う。再学習処理では、ステップS403で保持している特徴量とステップS409で生成されたフィードバック情報とを用いて、分類処理を行う際に用いられる閾値の更新処理を行う。例えば、フィードバック情報から画像ID情報を取得し、画像特徴量がその画像IDに分類されるようになるための閾値の算出処理が行われる。ここで、更新された閾値は分類処理へ反映され、分類に用いられる分類器の更新(文書識別処理を行う際に用いられる文書分類ルールの再生成)が行われる。 In step S410, the CPU 204 performs a relearning process from the feedback information generated in step S409. In the relearning process, a threshold value update process used in performing the classification process is performed using the feature amount held in step S403 and the feedback information generated in step S409. For example, image ID information is acquired from the feedback information, and a threshold value calculation process is performed so that the image feature amount is classified into the image ID. Here, the updated threshold value is reflected in the classification process, and the classifier used for classification is updated (reproducing the document classification rule used when performing the document identification process).
 ステップS411においてCPU204は、ステップS406において入力画像の保存先に対してアクセス権がないと判断された場合、別途アクセス権がある保存先へファイルへの格納を行う。例えばMFP101の記憶部203へ格納を行う。 In step S411, if it is determined in step S406 that there is no access right for the storage destination of the input image, the CPU 204 stores the file in a storage destination having a separate access right. For example, the data is stored in the storage unit 203 of the MFP 101.
 ステップS412においてMFP101は、ステップS411で保存したファイルに対する操作ログの取得を行う。ここでの操作ログの取得は、ステップS408で説明したことと同じであるため説明は省略する。 In step S412, the MFP 101 acquires an operation log for the file saved in step S411. Since the acquisition of the operation log here is the same as that described in step S408, description thereof is omitted.
 一方、ステップS405で画像が既知の画像であると判断されると、ステップS413に進む。ステップS413においてCPU204は、あらかじめ指定されているルールに沿って保存処理が行われる。 On the other hand, if it is determined in step S405 that the image is a known image, the process proceeds to step S413. In step S413, the CPU 204 performs storage processing in accordance with a rule designated in advance.
 <画像分類に利用する画像特徴量の詳細>
 画像分類処理で利用される画像特徴量について、図5と図6を用いて詳細を説明する。図5は、勾配情報に基づく画像特徴量の算出方法を説明する図である。
<Details of image features used for image classification>
Details of the image feature amount used in the image classification process will be described with reference to FIGS. 5 and 6. FIG. 5 is a diagram for explaining a method for calculating an image feature amount based on gradient information.
 図5に示すようにパッチ画像内の画素ごとに算出した勾配強度と勾配方向を利用する。具体的には、CPU301が、パッチ画像内の全ての画素について、縦方向と横方向のエッジ情報から勾配強度と勾配方向を求める。CPU301は、勾配情報を利用して、図5で示すように、1パッチから9次元(9個)の特徴量を算出する。まず、各画素について、勾配強度が一定値以上の画素をエッジ画素、一定値より小さい画素を非エッジ画素とする。エッジ画素群から勾配方向を8方向に量子化して、方向ごとの勾配強度積算値/パッチ画素数を計算し、非エッジ画素数/パッチ画素数と合わせて、1つのパッチ画像から9次元の特徴量を算出する。このように、エッジ画素と非エッジ画素を利用することで、罫線や文字の情報だけでなく、文書画像の大きな特徴である余白部分を表現することが可能になる。これまでの説明は、1つのパッチ画像における特徴量の説明であるが、実際には、複数のパッチ画像を切り出して利用することにより、多数の特徴量を利用する。 As shown in FIG. 5, the gradient strength and gradient direction calculated for each pixel in the patch image are used. Specifically, the CPU 301 obtains the gradient strength and gradient direction from the edge information in the vertical direction and the horizontal direction for all pixels in the patch image. The CPU 301 uses the gradient information to calculate 9-dimensional (9) feature amounts from one patch, as shown in FIG. First, for each pixel, a pixel having a gradient strength equal to or greater than a certain value is defined as an edge pixel, and a pixel having a gradient intensity smaller than the certain value is defined as a non-edge pixel. The gradient direction is quantized into 8 directions from the edge pixel group, and the gradient intensity integrated value / patch pixel number for each direction is calculated, and combined with the non-edge pixel number / patch pixel number, a 9-dimensional feature from one patch image Calculate the amount. As described above, by using the edge pixels and the non-edge pixels, it is possible to express not only ruled line and character information but also a margin part which is a large feature of the document image. The description so far is the description of the feature amount in one patch image, but actually, a large number of feature amounts are used by cutting out and using a plurality of patch images.
 図6は、パッチ画像の切り出しを説明する図である。 FIG. 6 is a diagram for explaining extraction of a patch image.
 まず、CPU301が、ノイズが表れやすい画像端(余白)を削除して、複数の解像度の画像を作成する。複数の解像度の画像を用意するのは、解像度ごとにエッジの構造が変わるためである。そして、CPU301が、それぞれの解像度の画像から複数のサイズのパッチ画像を走査しながら切り出すことで、パッチ画像位置を考慮した特徴量を算出する。例えば、300dpiでスキャンした画像から特徴量を抽出する場合を想定する。まず、CPU301が、スキャン画像を1/4サイズと1/8サイズに縮小した2種類の画像を作成する。CPU401は、上記縮小した各解像度の画像から1/4サイズのパッチ画像を1/5ずつずらして5×5=25枚、1/8サイズのパッチ画像を1/10ずつずらして、10×10=100枚切り出すことで、計250個のパッチを切り出す。このような設定においては、各パッチから9次元のエッジ特徴量を算出するので、画像1枚から9×250=2250次元の特徴量を算出することが可能となる。 First, the CPU 301 deletes image edges (margins) where noise is likely to appear, and creates images with a plurality of resolutions. The reason for preparing images with a plurality of resolutions is that the structure of the edge changes for each resolution. Then, the CPU 301 calculates a feature amount in consideration of the patch image position by cutting out patch images of a plurality of sizes from the respective resolution images. For example, assume that a feature amount is extracted from an image scanned at 300 dpi. First, the CPU 301 creates two types of images obtained by reducing the scanned image to ¼ size and 8 size. The CPU 401 shifts the 1/4 size patch image by 1/5 from each reduced resolution image by 5 × 5 = 25 sheets, and shifts the 1/8 size patch image by 1/10 by 10 × 10. By cutting out 100 sheets, a total of 250 patches are cut out. In such a setting, since a 9-dimensional edge feature amount is calculated from each patch, it is possible to calculate a 9 × 250 = 2250-dimensional feature amount from one image.
 なお、画像解像度、パッチサイズ、パッチ切り出し位置に関するパラメータは、上述した数字に限定されるものではない。また、取得する画像特徴量として、原稿の色の情報を利用するために、色ヒストグラムや色分散等を画像特徴量としてもよい。 Note that the parameters relating to the image resolution, patch size, and patch cut-out position are not limited to the numbers described above. Further, in order to use document color information as an image feature amount to be acquired, a color histogram, color dispersion, or the like may be used as the image feature amount.
 <学習データ増加処理の詳細>
 画像方向統一処理において、機械学習を利用して方向判別器の生成をする際に学習データを増やす学習データ増加処理について説明をする。本実施例では、画像をシミュレーションによって変形処理を施すことで変形画像を得て、それを学習データとして増やす。
<Details of learning data increase processing>
In the image direction unification process, a learning data increasing process for increasing learning data when generating a direction discriminator using machine learning will be described. In the present embodiment, a deformed image is obtained by performing a deformation process on the image by simulation, and this is increased as learning data.
 図7A、7B、7Cは、変形処理であるシフト処理、回転処理、拡大縮小処理を説明する図である。これらの幾何学的変形処理は、射影変換行列を利用して実現する。図7Aは、シフト処理を表している。シフト処理では、上下左右あるいは左上、右上、左下、右下に一定量だけ画像を並行移動させることで8パターンの変形画像を得る。図7Bは、回転処理を表している。回転処理では、一定量だけ時計回りと反時計回りの回転させることで、2パターンの変形画像を得る。図7Cは、拡大縮小処理を表している。拡大縮小処理では、画像を一定倍率だけ拡大または縮小することで2パターンの変形画像を得る。なお、それぞれの変形処理において、入力画像と出力画像とは同サイズである。射影変換後に出力画像の画像領域外にはみ出てしまう画像外領域については破棄する。また、出力画像の内部で射影元の存在しない欠損領域については、非欠損画素の画素値を順次コピーしていくことで補完する。この欠損領域の扱いについては、上述した方法による補完に限られるものではない。例えば、入力画像から推定した背景画素に置き換えるような他の補完方法でもよいし、補完は行なわずに欠損画素について欠損画素であるというフラグ情報を付加してマスク処理に利用する方法でもよい。 7A, 7B, and 7C are diagrams illustrating shift processing, rotation processing, and enlargement / reduction processing that are deformation processing. These geometric deformation processes are realized using a projective transformation matrix. FIG. 7A shows the shift process. In the shift process, eight patterns of deformed images are obtained by moving the images in parallel by a fixed amount in the vertical and horizontal directions or in the upper left, upper right, lower left and lower right. FIG. 7B shows the rotation process. In the rotation process, two patterns of deformed images are obtained by rotating clockwise and counterclockwise by a certain amount. FIG. 7C shows the enlargement / reduction processing. In the enlargement / reduction process, two patterns of deformed images are obtained by enlarging or reducing the image by a predetermined magnification. In each transformation process, the input image and the output image have the same size. The outside image area that protrudes outside the image area of the output image after projective transformation is discarded. In addition, a missing region in which no projection source exists in the output image is complemented by sequentially copying pixel values of non-missing pixels. The handling of this missing area is not limited to complementing by the method described above. For example, another complementing method that replaces the background pixel estimated from the input image may be used, or a method of adding flag information that the missing pixel is a missing pixel without performing the complementing may be used.
 学習データ増加処理では、これらシフト処理、回転処理、拡大縮小処理のパターンのそれぞれに変形しないというパターンを組み合わせることで、1枚の画像データからその組み合わせの数だけ変形画像を得ることができる。具体的には、上述した各変形処理のパターンに補正無しの場合を加えて、シフト処理が9パターン、回転処理が3パターン、拡大縮小処理が3パターンであるので、1枚の画像から3×9×3=81パターンの変形画像を生成して、学習データを増加させる。なお、それぞれの変形処理のパターン数は上述の数字に限定されるものではない。 In the learning data increase process, by combining the patterns that are not deformed into the shift process, the rotation process, and the enlargement / reduction process, it is possible to obtain deformed images corresponding to the number of combinations from one image data. Specifically, in addition to the case where there is no correction to the above-described deformation process patterns, there are 9 shift processes, 3 rotation processes, and 3 enlargement / reduction processes. 9 × 3 = 81 patterns of deformed images are generated to increase learning data. Note that the number of patterns of each deformation process is not limited to the above-described numbers.
 <利用する機械学習の詳細>
 次に、本実施例において画像を分類する分類器の生成に利用する機械学習の手法について説明をする。本実施例では、機械学習の手法としてReal AdaBoostと呼ばれる公知の手法を利用する。Real AdaBoostは、大量の特徴量から、与えられた学習データセットの分類に適した特徴量を選択して、その特徴量を組み合わせて分類器を構成することが可能な方法である。画像の分類時に大量の特徴量を利用すると、特徴量の計算負荷のためにパフォーマンスが低下してしまう。このように、分類に適した特徴量を選択して、一部の特徴量だけを利用し、分類器を構成できることは、Real AdaBoostの大きな利点である。ただし、Real AdaBoostは、2クラス分類器であり、2種類のラベルがついたデータを分類するものである。つまり、このままでは、3種類以上の画像の分類には利用することができない。そこで、2クラス分類器を多クラス分類器に拡張するOVA(One-Versus-All)と呼ばれる公知の方法を利用する。OVAは、1つのクラス(対象クラス)とそれ以外のクラスを分類する分類器をクラスの数だけ作成し、それぞれの分類器の出力を、対象クラスの信頼度とする。分類の際には、分類したいデータをすべての分類器に入力し、信頼度が最大であったクラスを分類先とする。
<Details of machine learning to be used>
Next, a machine learning method used for generating a classifier that classifies images in this embodiment will be described. In the present embodiment, a known technique called Real AdaBoost is used as a machine learning technique. Real AdaBoost is a method capable of selecting a feature amount suitable for classification of a given learning data set from a large amount of feature amounts and combining the feature amounts to configure a classifier. If a large amount of feature amount is used at the time of image classification, performance deteriorates due to the calculation load of the feature amount. Thus, it is a great advantage of Real AdaBoost that a classifier can be configured by selecting feature quantities suitable for classification and using only some of the feature quantities. However, Real AdaBoost is a two-class classifier and classifies data with two types of labels. That is, as it is, it cannot be used for classification of three or more types of images. Therefore, a known method called OVA (One-Versus-All) that expands the two-class classifier to a multi-class classifier is used. OVA creates classifiers for classifying one class (target class) and other classes by the number of classes, and uses the output of each classifier as the reliability of the target class. At the time of classification, data to be classified is input to all classifiers, and the class having the highest reliability is set as the classification destination.
 図8は、学習データを用いた機械学習の例を説明する図である。 FIG. 8 is a diagram for explaining an example of machine learning using learning data.
 この例では、学習データとして、3つのクラスの画像(画像A、画像B、画像C)のそれぞれに対応する画像特徴量が用意されているものとする。この3つのクラスを分類するために、OVAでは3種類の分類器を用意する。3種類の分類器は、画像Aとその他の画像を判別するための画像A判別器、画像Bとその他の画像を判別するための画像B判別器、画像Cとその他の画像を判別するための画像C判別器である。 In this example, it is assumed that image feature amounts corresponding to each of three classes of images (image A, image B, and image C) are prepared as learning data. In order to classify these three classes, OVA prepares three types of classifiers. The three types of classifiers are an image A discriminator for discriminating between image A and other images, an image B discriminator for discriminating between image B and other images, and an image B discriminating device for discriminating between image C and other images. An image C discriminator.
 画像A判別器は、画像Aが入力されたときに、大きい出力値(確信度)が出力され、それ以外の画像が入力されたときは、小さい出力値(確信度)が出力される。画像B判別器、画像C判別器についても同様である。実際の分類を行う際には、入力文書画像を3種類の分類器に入力し、その出力値の比較を行って、どの画像かを決定する。例えば画像B判別器の出力が最大であった場合は、その入力画像が画像Bであると判別する。 The image A discriminator outputs a large output value (confidence) when the image A is input, and outputs a small output value (confidence) when other images are input. The same applies to the image B classifier and the image C classifier. When actual classification is performed, input document images are input to three types of classifiers, and output values are compared to determine which image. For example, when the output of the image B discriminator is maximum, it is determined that the input image is the image B.
 図8を参照して説明したReal AdaBoostとOVAを利用した多クラス分類器の学習と、多クラス分類器を利用した文書画像分類は、CPU301で実行される。なお、本実施例で利用可能な機械学習の手法は、上述した手法に限定されるものではない。Supprot Vector MachineやRandom Forest等の公知の手法を利用してもよい。特徴選択の枠組みが機械学習の手法に含まれていない場合に、分類時の分類速度を向上させたい場合には、主成分分析や判別分析を利用した特徴量選択等の公知の特徴量選択を行う。機器学習手法が2クラス分類器である場合は、OVA以外の、All-Versus-All(AVA)やError-Correcting Output-Coding(ECOC)等の公知の手法を用いてもよい。 The learning of the multi-class classifier using Real AdaBoost and OVA described with reference to FIG. 8 and the document image classification using the multi-class classifier are executed by the CPU 301. Note that the machine learning technique that can be used in the present embodiment is not limited to the technique described above. You may utilize well-known methods, such as Suppprot Vector Machine and Random Forest. If the feature selection framework is not included in the machine learning method and you want to improve the classification speed during classification, select a known feature value such as feature value selection using principal component analysis or discriminant analysis. Do. When the device learning method is a two-class classifier, a known method such as All-Versus-All (AVA) or Error- Correcting Output-Coding (ECOC) other than OVA may be used.
 以上、実施例1によれば、分類先(格納先)が未知であった入力画像に対して、ユーザーのファイルに対する操作ログを取得し、取得された操作ログを利用することにより文書識別処理を行う際に用いられる文書分類ルールの生成(再学習)が可能となる。つまり、文書の分類に用いられる分類器が更新される。 As described above, according to the first embodiment, an operation log for a user's file is acquired for an input image whose classification destination (storage destination) is unknown, and document identification processing is performed by using the acquired operation log. It is possible to generate (relearn) a document classification rule to be used when performing. That is, the classifier used for document classification is updated.
 これにより、ユーザーが明示的に再学習を指示することなく、分類ルールを有さない(学習をしていない)画像に対しても画像分類が行うことが可能になる。 This makes it possible to classify images even for images that do not have classification rules (no learning) without the user explicitly instructing re-learning.
 (実施例2)
 実施例1では、未知フォルダに格納された画像に対する操作は、フォルダの移動処理を説明していた。
(Example 2)
In the first embodiment, the operation for the image stored in the unknown folder explained the folder moving process.
 実施例2では、未知フォルダに格納した画像に対して削除の操作が行われることを想定する。以下では、実施例1と差分がある部分についてのみ説明する。 In Example 2, it is assumed that a deletion operation is performed on an image stored in an unknown folder. Below, only the part which has a difference with Example 1 is demonstrated.
 <フローチャートを用いた本実施例の詳細説明>
 図9は、未分類となった入力画像に対する操作ログから入力画像を含むファイルを削除したか否かを検知し、削除したことに応じてフィードバック情報を生成する処理について説明するフローチャートである。図9に示す処理のうちMFP101が実行する処理は、CPU204が記憶部203に格納されている処理プログラムをロードして実行することで実現される。また、ステップS401~S413については実施例1にて説明したので割愛する。
<Detailed description of this embodiment using flowchart>
FIG. 9 is a flowchart for describing processing for detecting whether or not a file including an input image has been deleted from an operation log for an unclassified input image and generating feedback information in response to the deletion. Of the processes shown in FIG. 9, the process executed by the MFP 101 is realized by the CPU 204 loading and executing a processing program stored in the storage unit 203. Steps S401 to S413 are omitted since they have been described in the first embodiment.
 ステップS901においてCPU204は、ステップS408あるいはステップS412で取得したログからファイルが削除されたか否かの判断を行う。ファイルが削除されたか否かの検知は公知の方法で取得が可能である。 In step S901, the CPU 204 determines whether the file has been deleted from the log acquired in step S408 or step S412. Detection of whether or not a file has been deleted can be obtained by a known method.
 ステップS902においてCPU204は、ステップS901でファイルが削除された操作がなされたと判断した場合に、削除用のフォルダを生成しそこへ画像を格納する。または、ユーザーの指示を仰ぐ表示を操作部207に対して行う。ここで操作部207に表示されるものとしては、削除用のフォルダを生成しそこへファイルを移動するか、削除用のフォルダ生成は行わずファイルを削除するかをユーザーに選択させるものである。 In step S902, when the CPU 204 determines that the operation for deleting the file has been performed in step S901, the CPU 204 generates a folder for deletion and stores the image therein. Alternatively, the operation unit 207 is displayed for a user instruction. Here, what is displayed on the operation unit 207 is to cause the user to select whether to create a folder for deletion and move the file there, or delete the file without generating the folder for deletion.
 ステップS903において、CPU204はユーザーが操作部207に対して、削除用フォルダを生成すると入力されたか、削除用フォルダを生成しないと入力されたか判断を行う。 In step S903, the CPU 204 determines whether the user has input to the operation unit 207 when generating a deletion folder or not generating a deletion folder.
 ステップS904においてCPU204は、ステップS903で削除用のフォルダを生成すると指示があった場合に削除用のフォルダを生成し、そのフォルダへ削除対象の画像を含むファイルを格納する。そして、削除用のフォルダへ画像を含むファイルが分類されるように、画像IDをファイルへ付与する。 In step S904, the CPU 204 generates a deletion folder when instructed to generate a deletion folder in step S903, and stores a file including an image to be deleted in the folder. Then, an image ID is assigned to the file so that the file including the image is classified into the deletion folder.
 そして、ステップS409にて、処理対象の画像を削除用のフォルダに分類するための画像IDが入力される。 In step S409, an image ID for classifying the processing target image into a folder for deletion is input.
 ステップS903においてCPU204は、削除用のフォルダを生成しないとされた場合にはファイルを削除し、フィードバック情報は生成せずに処理を終える。 In step S903, if it is determined that the deletion folder is not generated, the CPU 204 deletes the file, and ends the process without generating feedback information.
 実施例2によれば、未知フォルダに入っている画像が削除された操作ログを取得する。例えば、未知フォルダに保存されている画像が削除された場合には、その操作ログから、その画像は不要な画像であると考えられる。よって、入力画像に対して不要、あるいはゴミ箱に分類されるような画像IDを付与する。これにより、不要な画像が何度も未知フォルダへ分類されることを避けることが可能になるため、ユーザーが毎回未知フォルダへ分類された画像を削除するようなわずらわしい操作を減らすことができる。 According to the second embodiment, an operation log in which an image in an unknown folder is deleted is acquired. For example, when an image stored in an unknown folder is deleted, it is considered that the image is an unnecessary image from the operation log. Therefore, an image ID that is unnecessary or classified as a trash can is assigned to the input image. As a result, it is possible to avoid unnecessary images from being classified into unknown folders many times, so that it is possible to reduce troublesome operations such as the user deleting images classified into unknown folders every time.
 (その他の実施例)
 本発明は、以下の処理を実行することによっても実現される。即ち、上述した実施例の機能を実現するソフトウェア(コンピュータプログラム)を、ネットワーク又は各種記憶媒体を介してシステム或いは装置に供給し、そのシステム或いは装置のコンピュータ(またはCPUやMPU等)がプログラムを読み出して実行する処理である。
(Other examples)
The present invention is also realized by executing the following processing. That is, software (computer program) that realizes the functions of the above-described embodiments is supplied to a system or apparatus via a network or various storage media, and the computer of the system or apparatus (or CPU, MPU, etc.) reads the program. To be executed.
 本発明は上記実施の形態に制限されるものではなく、本発明の精神及び範囲から離脱することなく、様々な変更及び変形が可能である。従って、本発明の範囲を公にするために以下の請求項を添付する。 The present invention is not limited to the above embodiment, and various changes and modifications can be made without departing from the spirit and scope of the present invention. Therefore, in order to make the scope of the present invention public, the following claims are attached.
 本願は、2015年12月25日提出の日本国特許出願特願2015-254377を基礎として優先権を主張するものであり、その記載内容の全てをここに援用する。 This application claims priority based on Japanese Patent Application No. 2015-254377 filed on December 25, 2015, the entire contents of which are incorporated herein by reference.

Claims (8)

  1.  処理対象の入力画像から取得した特徴量と分類ルールを用いて前記処理対象の入力画像に対して分類を行う分類手段と、
     前記分類手段により分類された結果に基づいて決められた格納先に前記処理対象の入力画像を含むファイルを格納する格納手段と、
     前記格納手段により格納されたファイルに対する操作履歴を取得する取得手段と、
    前記取得手段により取得した操作履歴を用いて前記分類ルールを再生成する生成手段と、を有し、
     前記分類手段は、前記生成手段により再生成された分類ルールを用いて、処理対象の入力画像に対して分類を行うことを特徴とする画像処理装置。
    Classification means for classifying the input image to be processed using a feature amount and a classification rule acquired from the input image to be processed;
    Storage means for storing a file containing the input image to be processed in a storage location determined based on the result of classification by the classification means;
    Obtaining means for obtaining an operation history for the file stored by the storage means;
    Generating means for regenerating the classification rule using the operation history acquired by the acquisition means,
    The image processing apparatus according to claim 1, wherein the classification unit classifies the input image to be processed using the classification rule regenerated by the generation unit.
  2.  前記分類手段は、前記操作履歴を反映して再生成された分類ルールを用いて、処理対象の入力画像を分類することを特徴とする請求項1に記載の画像処理装置。 2. The image processing apparatus according to claim 1, wherein the classification unit classifies the input image to be processed using a classification rule regenerated by reflecting the operation history.
  3.  前記分類ルールとは、画像から取得された特徴量と処理対象の入力画像を分類するための分類器との組合せであることを特徴とする請求項1に記載の画像処理装置。 The image processing apparatus according to claim 1, wherein the classification rule is a combination of a feature amount acquired from an image and a classifier for classifying an input image to be processed.
  4.  前記操作履歴とは、処理対象の入力画像を含むファイルに対するフォルダ移動または前記ファイル名の変更または前記ファイルの削除に関する情報であることを特徴とする請求項1に記載の画像処理装置。 2. The image processing apparatus according to claim 1, wherein the operation history is information relating to folder movement, file name change, or file deletion for a file including an input image to be processed.
  5.  前記格納手段により処理対象の入力画像を含むファイルを格納する格納先に対してアクセス権がない場合、前記格納先とは別の格納先でありアクセス権がある格納先へ前記ファイルを格納することを特徴とする請求項1に記載の画像処理装置。 When the storage means does not have access right to the storage destination storing the file including the input image to be processed, the file is stored in a storage destination different from the storage destination and having access right. The image processing apparatus according to claim 1.
  6.  前記操作履歴が特定の操作である場合、画面に指示を受けるための表示を行う表示制御手段を有することを特徴とする請求項1に記載の画像処理装置。 The image processing apparatus according to claim 1, further comprising display control means for performing display for receiving an instruction on a screen when the operation history is a specific operation.
  7.  処理対象の入力画像から取得した特徴量と分類ルールを用いて前記処理対象の入力画像に対して分類を行う分類ステップと、
     前記分類ステップにより分類された結果に基づいて決められた格納先に前記処理対象の入力画像を含むファイルを格納する格納ステップと、
     前記格納ステップにより格納されたファイルに対する操作履歴を取得する取得ステップと、
     前記取得ステップにより取得した操作履歴を用いて前記分類ルールを再生成する生成ステップ
    とを有し、
     前記再生成された分類ルールを用いて、処理対象の入力画像に対して分類を行うことを特徴とする画像処理方法。
    A classification step for classifying the input image to be processed using the feature amount and the classification rule acquired from the input image to be processed;
    A storage step of storing a file including the input image to be processed in a storage destination determined based on the result of the classification in the classification step;
    An acquisition step of acquiring an operation history for the file stored in the storage step;
    A generation step of regenerating the classification rule using the operation history acquired in the acquisition step;
    An image processing method, wherein classification is performed on an input image to be processed using the regenerated classification rule.
  8.  コンピュータを、
     処理対象の入力画像から取得した特徴量と分類ルールを用いて前記処理対象の入力画像に対して分類を行う分類ステップと、
     前記分類ステップにより分類された結果に基づいて決められた格納先に前記処理対象の入力画像を含むファイルを格納する格納ステップと、
     前記格納ステップにより格納されたファイルに対する操作履歴を取得する取得ステップと、
     前記取得ステップにより取得した操作履歴を用いて前記分類ルールを再生成する生成ステップと、を有し、
     前記再生成された分類ルールを用いて、処理対象の入力画像に対して分類を行うことを特徴とする画像処理方法として実行させるためのプログラム。
    Computer
    A classification step for classifying the input image to be processed using the feature amount and the classification rule acquired from the input image to be processed;
    A storage step of storing a file including the input image to be processed in a storage destination determined based on the result of the classification in the classification step;
    An acquisition step of acquiring an operation history for the file stored in the storage step;
    And regenerating the classification rule using the operation history acquired in the acquisition step,
    A program for causing an input image to be processed to be classified using the regenerated classification rule as an image processing method.
PCT/JP2016/087324 2015-12-25 2016-12-15 Image-processing device, image-processing method, and computer program WO2017110640A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2015254377A JP2017117335A (en) 2015-12-25 2015-12-25 Image processing apparatus, image processing method, and computer program
JP2015-254377 2015-12-25

Publications (1)

Publication Number Publication Date
WO2017110640A1 true WO2017110640A1 (en) 2017-06-29

Family

ID=59090164

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2016/087324 WO2017110640A1 (en) 2015-12-25 2016-12-15 Image-processing device, image-processing method, and computer program

Country Status (2)

Country Link
JP (1) JP2017117335A (en)
WO (1) WO2017110640A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111837157A (en) * 2018-03-08 2020-10-27 株式会社岛津制作所 Cell image analysis method, cell image analysis device, and learning model creation method

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP7070656B2 (en) * 2018-03-06 2022-05-18 株式会社島津製作所 Cell image analysis method, cell image analysis device, and learning model creation method
JP7339491B2 (en) * 2018-12-28 2023-09-06 キヤノンマーケティングジャパン株式会社 Information processing device, information processing method, and program
JP7492858B2 (en) * 2020-05-15 2024-05-30 シャープ株式会社 Image forming device

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000057349A (en) * 1998-08-10 2000-02-25 Hitachi Ltd Method for sorting defect, device therefor and method for generating data for instruction
WO2012176317A1 (en) * 2011-06-23 2012-12-27 サイバーアイ・エンタテインメント株式会社 Image recognition system-equipped interest graph collection system using relationship search

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2000057349A (en) * 1998-08-10 2000-02-25 Hitachi Ltd Method for sorting defect, device therefor and method for generating data for instruction
WO2012176317A1 (en) * 2011-06-23 2012-12-27 サイバーアイ・エンタテインメント株式会社 Image recognition system-equipped interest graph collection system using relationship search

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111837157A (en) * 2018-03-08 2020-10-27 株式会社岛津制作所 Cell image analysis method, cell image analysis device, and learning model creation method
US11978211B2 (en) 2018-03-08 2024-05-07 Shimadzu Corporation Cellular image analysis method, cellular image analysis device, and learning model creation method

Also Published As

Publication number Publication date
JP2017117335A (en) 2017-06-29

Similar Documents

Publication Publication Date Title
JP6433187B2 (en) Information processing apparatus, control method thereof, and computer program
JP4574235B2 (en) Image processing apparatus, control method therefor, and program
JP2017107455A (en) Information processing apparatus, control method, and program
US8384964B2 (en) Image processing apparatus and image processing method
WO2017110640A1 (en) Image-processing device, image-processing method, and computer program
JP2017146745A (en) Information processing apparatus, control method, information processing system, and program
US20170124390A1 (en) Image processing apparatus, image processing method, and non-transitory computer readable medium
JP2022147074A (en) Image processing apparatus, image processing method, and program
US9338310B2 (en) Image processing apparatus and computer-readable medium for determining pixel value of a target area and converting the pixel value to a specified value of a target image data
JP4213112B2 (en) Image search apparatus, image forming apparatus, image search apparatus control method, image search program, and computer-readable recording medium
US11223743B2 (en) Image processing system, image processing apparatus, method of controlling the image processing apparatus, and storage medium
US8531717B2 (en) Image processing enabling reduction of processing time for data conversion
JP2006203584A (en) Image processor and program
JP6494435B2 (en) Information processing apparatus, control method thereof, and computer program
US8059296B2 (en) Image forming apparatus that synthesizes fiber information extracted from pages of a paper medium having a plurality of pages, and an image forming apparatus control method, a program, and a storage medium relating thereto
US9355473B2 (en) Image forming apparatus having color conversion capability
US8675244B2 (en) Rendering apparatus having a display list clip processing control method, rendering method, and storage medium
JP4424718B2 (en) Image output apparatus, control method therefor, computer program, and image output system
JP4906488B2 (en) Image forming apparatus, image forming method, and program
JP2019153061A (en) Image processing device, control method of the same and program thereof
JP2015166899A (en) Image forming apparatus, control method thereof, and program
US11212419B1 (en) Image reading system, image forming system, and image reading method that perform image processing for each area
JP5389096B2 (en) Apparatus and control method thereof
JP2022036658A (en) Printing time estimation image creation device, printing time estimation image creation method, and program
JP2023158554A (en) Character image quality adjustment system, character image quality adjustment device, and control method and program therefor

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 16878531

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 16878531

Country of ref document: EP

Kind code of ref document: A1