US6850645B2 - Pattern recognizing apparatus - Google Patents
Pattern recognizing apparatus Download PDFInfo
- Publication number
- US6850645B2 US6850645B2 US10/000,525 US52501A US6850645B2 US 6850645 B2 US6850645 B2 US 6850645B2 US 52501 A US52501 A US 52501A US 6850645 B2 US6850645 B2 US 6850645B2
- Authority
- US
- United States
- Prior art keywords
- character
- box
- pattern
- recognizing
- column
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/10—Image acquisition
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/40—Document-oriented image-based pattern recognition
- G06V30/41—Analysis of document content
- G06V30/414—Extracting the geometrical structure, e.g. layout tree; Block segmentation, e.g. bounding boxes for graphics or text
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/10—Character recognition
- G06V30/24—Character recognition characterised by the processing or recognition method
- G06V30/242—Division of the character sequences into groups prior to recognition; Selection of dictionaries
Definitions
- the present invention relates to an apparatus and method for recognizing a pattern, and realizes to recognize characters, graphics, and symbols correctly depending on various states of input images when used with a printed character recognizing apparatus and a graphics recognizing apparatus as well as a handwritten character recognizing apparatus.
- OCR optical character reader
- FIG. 1 is a block diagram showing the configuration of the conventional handwritten character recognizing apparatus.
- a form/document 311 is read using a scanner to obtain a multiple-value image of the form/document 311 .
- a preprocessing unit 312 binarizes a multiple-value image; removes noises, and amends the position of the form/document 311 .
- a character detecting unit 313 detects each character according to information about preliminarily defined ruled lines and positional information about a character.
- a character recognizing unit 314 recognizes each character and outputs a character code.
- the character is recognized by collating each feature of an unknown character pattern detected by the character detecting unit 313 with the feature of each character category preliminarily entered in a recognizing dictionary 315 .
- a distance between feature vectors in a feature space is computed by converting a 2-dimensional character pattern into a feature vector in a feature space representing the feature of the character, as a similarity between the unknown character pattern and the character category preliminarily entered in the recognizing dictionary 315 .
- the character category is recognized corresponding to the unknown character pattern.
- a threshold is set for a distance between two feature vectors to avoid mistakenly recognizing a non-character such as a deletion line, a noise, a symbol, etc. for a character and outputting a character code for a non-character. If the distance between the two feature vectors is larger than the threshold, a reject code is output by determining that the unknown character pattern has no corresponding character category preliminarily entered in the recognizing dictionary 315 , or that the unknown character pattern refers to a non-character.
- the recognizing dictionary 315 also contains the features of the character categories of high-quality characters, obscure characters, and deformed characters.
- a high-quality character recognizing dictionary 315 is referred to for high quality characters.
- An obscure character recognizing dictionary 315 is referred to for obscure characters.
- a deformed-character recognizing dictionary 315 is referred to for deformed characters.
- FIG. 2 shows the configuration of the character recognizing apparatus for recognizing a character with a deletion line.
- the character recognizing apparatus shown in FIG. 2 comprises an image input unit 491 for inputting an original image, containing a character and detecting or preprocessing a character from the input image, and an identifying unit 492 for identifying a character by extracting the feature of the character and comparing the extracted feature with the feature of the standard pattern stored in the recognizing dictionary.
- the handwritten character recognizing apparatus shown in FIG. 1 equally processes a detected character among obscure characters, deformed characters, high-quality characters using the same recognizing dictionary 315 .
- a character may touch its character box.
- a single recognizing dictionary 315 When a single recognizing dictionary 315 is referred to in various environments, they affect each other, thereby generating a problem that the recognizing process cannot be performed with enhanced precision.
- the ‘ 0 ’ to be deleted is not rejected but recognized as ‘ 8 ’.
- the ‘ 1 ’ to be deleted is not rejected but recognized as ‘ 8 ’.
- the ‘ 7 ’ to be deleted is not rejected but recognized as ‘ 4 ’.
- the ‘ 6 ’ to be deleted is not rejected but recognized as ‘ 6 ’.
- the present invention aims at providing a pattern recognizing apparatus and method capable of appropriately recognizing a character with high precision depending on the environment of the character.
- an input pattern is recognized by extracting the first predetermined feature from the input pattern and extracting the second predetermined feature from the input pattern from which the first feature has been extracted.
- a recognizing process can be performed depending on each environment of a character.
- a pattern is recognized by extracting the state of a process object from an input image and selecting a recognizing process suitable for the state for each process object.
- a pattern recognizing process can be performed appropriately for each state on the input image having various states, thereby realizing the recognizing process with high precision.
- a state of a process object is extracted from an input image, and a pattern recognizing process exclusively for the first state is performed on the process object in the first state, and a pattern recognizing process exclusively for the second state is performed on the process object in the second state.
- the recognizing process on the process object in the first state interacts with the recognizing process on the process object in the second state, thereby successfully performing the recognizing processes with high precision.
- recognizing dictionaries are appropriately selected for an input image in various states.
- the recognizing process can be performed with high precision by using an obscure character recognizing dictionary for obscure characters, a deformed-character recognizing dictionary for deformed characters, and high-quality character recognizing dictionary for high-quality characters.
- identification functions are appropriately selected for an input image in various states.
- the recognizing process can be performed with high precision by, for example, recognizing a character using a city block distance on a character written in a one-character box, and recognizing a character using a discriminant function on a character written in a free-pitch box in consideration of the character detection reliability.
- the recognizing process can be performed with high precision by, for example, setting a correspondence between an unknown character and a character category by dividing a character into character segments when an unknown character is considerably deformed and has no correspondence with a character category stored in the recognizing dictionary, computing the detection reliability using a discriminant function generated based on a learning pattern when a character is detected from a character string, and evaluating the recognition reliability on a box-touching character using the reliability obtained through a learning pattern when the box-touching character is recognized.
- the recognizing process is performed according to priority until the reliability of the recognizing process reaches a predetermined value when a plurality of recognizing processes are called for a specified process object.
- the reliability of the recognizing process can be enhanced and the precision of the process can be successfully improved.
- a non-character is extracted from an input image and a non-character recognizing process and a character recognizing process are performed separately on the extracted non-character.
- the recognizing process can be performed with a high precision with less characters mistaken for non-characters and with less non-characters mistaken for characters.
- the first predetermined feature is extracted from an input pattern, and the input pattern is recognized by extracting the second predetermined feature from the input pattern from which the first predetermined feature has not been extracted.
- a character with a deletion line can be distinguished from a character without a deletion line, and only the character without a deletion line can be recognized. Therefore, it is possible to prevent a character with a deletion line from being mistakenly recognized for any other character.
- the first predetermined feature is extracted from an input pattern
- a portion contributing to the first predetermined feature can be removed from the input pattern from which the first predetermined feature has been extracted, and the input pattern is recognized based on a pattern from which the portion contributing to the first predetermined feature has been removed.
- FIG. 1 is a block diagram showing the configuration of the conventional character recognizing apparatus
- FIG. 2 is a block diagram showing the configuration of the conventional character recognizing apparatus
- FIG. 3 shows an example of recognizing a character with a deletion line
- FIG. 4 is a block diagram showing the configuration of the pattern recognizing apparatus according to the first embodiment of the present invention.
- FIG. 5 is a block diagram showing the functions of the pattern recognizing apparatus according to the second embodiment of the present invention.
- FIG. 6 is a block diagram showing an embodiment of the practical configuration of the environment recognizing unit shown in FIG. 5 ;
- FIG. 7 is a block diagram showing an embodiment of the practical configuration of the pattern recognizing apparatus shown in FIG. 5 ;
- FIG. 8 is a flowchart showing an embodiment of the entire operations of the environment recognizing system shown in FIG. 1 ;
- FIG. 9 is a flowchart showing an embodiment of the operations of the preprocessing unit shown in FIG. 8 ;
- FIG. 10 is a flowchart showing an embodiment of the operations of the layout analyzing unit shown in FIG. 8 ;
- FIG. 11 is a flowchart showing an embodiment of the operations of the quality analyzing unit shown in FIG. 8 ;
- FIG. 12 is a flowchart showing an embodiment of the operations of the correction analyzing unit shown in FIG. 8 ;
- FIG. 13 is a flowchart showing an embodiment of the operations of the control unit for controlling character/non-character recognizing processes shown in FIG. 8 ;
- FIG. 14 is a block diagram showing the configuration of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 15 is a block diagram showing a practical configuration of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 16 is an example of the labelling process of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIGS. 17A through 17D show the representations of compressing the labelling process of the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 18 shows an example of the text extracting process of the pattern recognizing apparatus according to an embodiment of the present invention
- FIGS. 19A through 19D show examples of the partial area in the text extracting process of the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 20 shows the contiguous projecting method in the ruled line extracting process of the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 21 shows the pattern projection result in the ruled line extracting process of the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 22 is a flowchart showing the ruled line extracting process of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 23 shows the ruled line extracting process of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 24 shows the method of completing an obscure ruled line in the ruled line extracting process of the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 25 is a flowchart showing the method of completing an obscure ruled line of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 26 shows the searching direction when an obscure ruled line is completed by the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 27 is a flowchart showing the one-character box extracting process of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 28 is a flowchart showing the block character box extracting process of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIGS. 29A through 29E show the types of boxes and tables used in the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 30 is a flowchart showing the image reducing process of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIGS. 31A through 31E show the box-touching state determining process of the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 32 is a flowchart showing the box-touching state determining process of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIGS. 33A through 33E show the types of deletion lines used in the pattern recognizing apparatus according to an embodiment of the present invention.
- FIGS. 34A through 34C show the method of computing the feature of a corrected character used in the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 35 is a block diagram showing an example of the configuration of the basic character recognizing unit shown in FIG. 7 ;
- FIG. 36 shows an example of the method of computing a feature vector in the basic character recognizing unit shown in FIG. 7 ;
- FIG. 37 shows an example of the method of computing the distance between the feature vectors in the basic character recognizing unit shown in FIG. 7 ;
- FIGS. 38A through 38C show the method of extracting a character segment by the detailed identifying method for use with the basic character recognizing unit shown in FIG. 7 ;
- FIG. 39 shows the method of detecting an end point in the detailed identifying method for use with the basic character recognizing unit shown in FIG. 7 ;
- FIG. 40 shows the method of detecting a change in angle in the detailed identifying method for use with the basic character recognizing unit shown in FIG. 7 ;
- FIGS. 41A and 41B show the correspondence among character segments in the detailed identifying method for use with the basic character recognizing unit shown in FIG. 7 ;
- FIG. 42 is a flowchart showing the process of the detailed identifying method for use with the basic character recognizing unit shown in FIG. 7 ;
- FIGS. 43A through 43C show the method of completing a character by the box-touching character recognizing unit shown in FIG. 7 ;
- FIGS. 44A through 44D show the re-completing method by the box-touching character recognizing unit shown in FIG. 7 ;
- FIGS. 45A through 45C show examples of a completed misread character by the box-touching character recognizing unit FIG. 7 ;
- FIG. 46 is a block diagram showing an example of the method of learning a character by the box-touching character recognizing unit FIG. 7 ;
- FIG. 47 shows the method of generating a box-touching character by the box-touching character recognizing unit shown in FIG. 7 ;
- FIG. 48 shows an example of generating a box-touching character by the box-touching character recognizing unit shown in FIG. 7 ;
- FIG. 49 shows an example of a knowledge table for use in the box-touching character recognizing unit shown in FIG. 7 ;
- FIG. 50 shows an example of the type and amount of a change entered on the knowledge table for use in the box-touching character recognizing unit shown in FIG. 7 ;
- FIGS. 51A and 51B show examples of the re-recognized area emphasized by the box-touching character recognizing unit shown in FIG. 7 ;
- FIGS. 52A through 52D show the re-recognizing method using an emphasized area by the box-touching character recognizing unit shown in FIG. 7 ;
- FIG. 53 is a flowchart showing the re-recognizing process using an emphasized area by the box-touching character recognizing unit shown in FIG. 7 ;
- FIG. 54 is a block diagram showing an example of the character re-recognizing method for use with the box-touching character recognizing unit shown in FIG. 7 ;
- FIG. 55 is a block diagram showing the character re-recognizing process for use with the box-touching character recognizing unit shown in FIG. 7 ;
- FIG. 56 shows the graphic meaning of a parameter in the statistic process performed by the character st ring recognizing unit shown in FIG. 7 ;
- FIG. 57 is a flowchart showing the statistic process performed by the character string recognizing unit shown in FIG. 7 ;
- FIG. 58 shows the graphic meaning of a parameter in the delimiting character process performed by the character string recognizing unit shown in FIG. 7 ;
- FIG. 59 is a flowchart showing the delimiting character process per formed by the character string recognizing unit shown in FIG. 7 ;
- FIG. 60 shows the graphic meaning of a parameter in the superscript-stroke process performed by the character string recognizing unit shown in FIG. 7 ;
- FIG. 61 is a flowchart showing the superscript-stroke process performed by the character string recognizing unit shown in FIG. 7 ;
- FIG. 62 is a flowchart showing the process of computing the character-detection-possibility data of the character string recognizing unit shown in FIG. 7 ;
- FIG. 63 shows the method of quantizing the character detection reliability of the character string recognizing unit shown in FIG. 7 ;
- FIG. 64 shows the method of generating the frequency distribution of the character string recognizing unit shown in FIG. 7 ;
- FIG. 65 is a flowchart showing the method of computing the character detection reliability of the character string recognizing unit shown in FIG. 7 ;
- FIG. 66 shows an example of a histogram distribution about the success and failure in character detection by the character string recognizing unit shown in FIG. 7 ;
- FIG. 67 shows the method of computing the overlapping area of the success and failure in character detection by the character string recognizing unit shown in FIG. 7 ;
- FIG. 68 shows the flow of the process of detecting a character by the character string recognizing unit shown in FIG. 7 ;
- FIG. 69 shows the flow of the process of detecting a character in a non-statistic process performed by the character string recognizing unit shown in FIG. 7 ;
- FIG. 70 is a block diagram showing an example of the configuration of the obscure character recognizing unit shown in FIG. 7 ;
- FIG. 71 shows an example of the process performed by the deletion line recognizing unit shown in FIG. 7 ;
- FIG. 72 shows the flow of the clustering process performed by the unique character analyzing unit shown in FIG. 7 ;
- FIG. 73 is a flowchart showing the clustering process performed by the unique character analyzing unit shown in FIG. 7 ;
- FIG. 74 shows the flow of the character category determination result correcting process performed by the unique character analyzing unit shown in FIG. 7 ;
- FIG. 75 is a flowchart showing the character category determination result correcting process performed by the unique character analyzing unit shown in FIG. 7 ;
- FIG. 76 shows an example of a list to be processed by the pattern recognizing apparatus according to the present invention.
- FIG. 77 shows an example of the intermediate process result table for use in the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 78 shows an example of the process order table for use in the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 79 shows an example of the intermediate process result table for use in the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 80 shows an example of the intermediate process result table for use in the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 81 shows an example of the intermediate process result table for use in the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 82 shows an example of the intermediate process result table for use in the pattern recognizing apparatus according to an embodiment of the present invention
- FIG. 83 is a block diagram showing the function of the pattern recognizing apparatus according to the third embodiment of the present invention.
- FIG. 84 is a block diagram showing the function of the pattern recognizing apparatus according to the fourth embodiment of the present invention.
- FIG. 85 is a block diagram showing the function of the pattern recognizing apparatus according to the fifth embodiment of the present invention.
- FIG. 86 is a block diagram showing the function of the pattern recognizing apparatus according to the sixth embodiment of the present invention.
- FIG. 87 is a block diagram showing the function of the pattern recognizing apparatus according to the seventh embodiment of the present invention.
- FIG. 88 is a flowchart showing the operations performed by the pattern recognizing apparatus according to the eighth embodiment of the present invention.
- FIG. 89 is a flowchart showing the operations performed by the pattern recognizing apparatus according to the ninth embodiment of the present invention.
- FIG. 90 is a flowchart showing the operations performed by the pattern recognizing apparatus according to the tenth embodiment of the present invention.
- FIG. 91 is a flowchart showing the operations performed by the pattern recognizing apparatus according to the eleventh embodiment of the present invention.
- FIG. 92 is a block diagram showing the function of the pattern recognizing apparatus according to the twelfth embodiment of the present invention.
- FIG. 93 shows the amount of the complexity of the pattern recognizing apparatus according to an embodiment of the present invention.
- FIG. 94 is a flowchart showing the operations of the pattern recognizing apparatus shown in FIG. 92 ;
- FIG. 95 is a flowchart showing the operations of the pattern recognizing apparatus shown in FIG. 92 .
- FIG. 4 is a block diagram showing the functions of the pattern recognizing apparatus according to the first embodiment of the present invention.
- a feature extracting unit 1000 extracts the first predetermined feature from an input pattern.
- a pattern recognizing unit 1001 recognizes an input pattern by extracting the second predetermined feature from the input pattern from which the feature extracting unit 1000 has extracted the first feature.
- the pattern recognizing unit 1001 can recognize only an input pattern having the first predetermined feature, and can perform a recognizing process on each environment of the input pattern, thereby improving the precision of the recognizing process.
- the pattern recognizing unit 1001 can also recognize an input pattern by extracting the second predetermined feature from the input pattern from which the feature extracting unit 1000 has not extracted the first feature.
- the pattern recognizing unit 1001 can remove the input pattern having the first predetermined feature from a process object, and only an appropriate process object can be selected and recognized, thereby improving the precision of the recognizing process.
- FIG. 5 is a block diagram showing the functions of the pattern recognizing apparatus according to the second embodiment of the present invention.
- a environment recognizing unit 1 extracts the first through N-th states from an input image.
- a state extracted from the input image refers to, for example, the format such as a one-character box, a free-pitch box, table, etc. in which characters are written; the relationship between a character and its box; the obscurity of a character; the deformation of a character; the deletion of a character using a deletion line, etc.
- a first pattern recognizing unit 2 exclusively recognizes a pattern of a process object in the first state.
- a second pattern recognizing unit 4 exclusively recognizes a pattern of a process object in the second state.
- An N-th pattern recognizing unit 6 exclusively recognizes a pattern of a process object in the N-th state.
- the first through N-th pattern recognizing units 2 , 4 , and 6 respectively comprise reliability computing units 3 , 5 , and 7 , and compute the reliability of the recognition results obtained by the first through N-th pattern recognizing units 2 , 4 , and 6 .
- the environment recognizing unit 1 calls any of the first through N-th pattern recognizing units 2 , 4 , and 6 corresponding to the first through N-th state to perform a recognizing process.
- the pattern recognizing process to be performed by the first pattern recognizing unit 2 is called for the process object in the first state.
- the pattern recognizing process to be performed by the second pattern recognizing unit 4 is called for the process object in the second state.
- the pattern recognizing process to be performed by the N-th pattern recognizing unit 6 is called for the process object in the N-th state.
- the environment recognizing unit 1 extracts, for example, the first and second states for a single process object, the pattern recognizing processes to be respectively performed by the first pattern recognizing unit 2 and second pattern recognizing unit 4 are called for the process object.
- the first state refers to the state in which characters are written in one-character boxes
- the second state refers to the state in which characters are written in free-pitch character boxes
- the third state refers to the state in which a character touches its character box
- the fourth state refers to the state in which a character is obscure
- the fifth state refers to the state in which a character is deformed
- the sixth state refers to the state in which a character is corrected using deletion lines.
- the first pattern recognizing unit 2 recognizes a character written in a one-character box
- the second pattern recognizing unit 4 recognizes a character string written in a free-pitch box
- the third pattern recognizing unit recognizes a box-touching character
- the fourth pattern recognizing unit recognizes an obscure character
- the fifth pattern recognizing unit recognizes a deformed character
- the sixth pattern recognizing unit recognizes a corrected character.
- the first pattern recognizing unit 2 performs the recognizing process on the character written in the one-character box.
- the second pattern recognizing unit 4 performs the recognizing process on the character written in the free-pitch box.
- the third pattern recognizing unit performs the recognizing process on the box-touching character.
- the fourth pattern recognizing unit performs the recognizing process on the obscure character.
- the fifth pattern recognizing unit performs the recognizing process on the deformed character.
- the sixth pattern recognizing unit performs the recognizing process on the candidate for the corrected character.
- the pattern recognizing unit 2 and the pattern recognizing unit 3 perform the recognizing processes on the box-touching character touching the free-pitch box.
- the second pattern recognizing unit 4 , the third pattern recognizing unit, and the sixth pattern recognizing unit perform the recognizing processes on the box-touching character with deletion lines which touches the free-pitch box.
- the order of the recognizing processes to be performed by the plurality of pattern recognizing units 2 , 4 , and 6 is determined according to the process order table storing the order of calling the plurality of pattern recognizing units 2 , 4 , and 6 .
- the recognizing processes to be performed by the plurality of pattern recognizing units 2 , 4 , and 6 are sequentially performed in the calling order until the reliability larger than a predetermined threshold can be obtained by the reliability computing units 3 , 5 , and 7 in the recognizing processes performed by the pattern recognizing units 2 , 4 , and 6 .
- the third pattern recognizing unit first performs the recognizing process and then the second pattern recognizing unit performs the recognizing process on the box-touching character which touches the free-pitch box.
- the third pattern recognizing unit first performs the recognizing process
- the sixth pattern recognizing unit performs the recognizing process
- the second pattern recognizing unit 4 performs the recognizing process on the box-touching character with deletion lines which touches the free-pitch box.
- FIG. 6 is a block diagram showing the configuration of an embodiment of the environment recognizing unit 1 shown in FIG. 1 .
- a state extracting unit 1 a extracts the first through the N-th states from an input image.
- a recognizing process controlling unit 1 b calls one or a plurality of the first through N-th pattern recognizing units 2 , 4 , and 6 shown in FIG. 5 corresponding to the first through N-th states extracted by the state extracting unit 1 a for use in a recognizing process.
- a process order table 1 f stores the process order indicating the process order of the first through N-th pattern recognizing units 2 , 4 , and 6 when the plurality of recognizing units are called from among the first through N-th pattern recognizing units 2 , 4 , and 6 .
- a process order control rule storage unit 1 d stores the calling procedure indicating the recognizing unit to be called from among the first through N-th pattern recognizing units 2 , 4 , and 6 based on the first through N-th states extracted by the state extracting unit 1 a.
- An intermediate process result table generating unit 1 c generates an intermediate process result table indicating the process order of the first through N-th pattern recognizing units 2 , 4 , and 6 according to the calling procedure stored in the process order control rule storage unit 1 d and the process order stored in the process order table 1 f.
- a process performing rule storage unit 1 e stores the procedure indicating the next process to be performed based on the result of the recognizing process entered in the intermediate process result table.
- the pattern recognizing apparatus shown in FIG. 5 extracts the state of a process object from an input image, selects an appropriate recognizing process for the state for each process object so that an appropriate pattern recognizing process can be performed for each state on an input image having various states, thereby performing the recognizing process with high precision. Since the process object is evaluated when the state is extracted and also when a recognizing process is performed on the process object, the precision of the recognizing process can be furthermore improved.
- extracting the state of a process object from an input image, performing the pattern recognizing process exclusively for the first state on the process object having the first state, and performing the pattern recognizing process exclusively for the second state on the process object having the second state suppress the mutual influence between the recognizing process on the process object having the first state and the recognizing process on the process object having the second state, thereby realizing a recognizing process with high precision.
- FIG. 7 is a block diagram showing a practical configuration of the pattern recognizing apparatus according to an embodiment of the present invention.
- an environment recognizing system 11 extracts the state of an input image, and calls one or a plurality of a basic character recognizing unit 17 , a character string recognizing unit 15 , a box-touching character recognizing unit 13 , an obscure character recognizing unit 19 , a deformed character recognizing unit 21 of a character recognizing unit 12 and a deletion line recognizing unit 26 and noise recognizing unit 28 of a non-character recognizing unit 25 based on the extracted state.
- the character recognizing unit 12 performs a character recognizing process for each state of an input image and comprises the basic character recognizing unit 17 for recognizing a character, the character string recognizing unit 15 for performing a character recognizing process B and a character detecting process B on a character string, the box-touching character recognizing unit 13 for performing a character recognizing process A and a character detecting process A on a box-touching character, the obscure character recognizing unit 19 for performing a character recognizing process C and a character detecting process C on an obscure character, the deformed character recognizing unit 21 for performing a character recognizing process D and a character detecting process D on a deformed character, and the unique character recognizing unit 23 for performing a unique character recognizing process E and a unique character detecting process E on a unique character.
- the basic character recognizing unit 17 , character string recognizing unit 15 , box-touching character recognizing unit 13 , obscure character recognizing unit 19 , deformed character recognizing unit 21 , and unique character recognizing unit 23 respectively comprise knowledge tables 14 , 16 , 18 , 20 , 22 and 24 storing the knowledge of the character recognizing method.
- the knowledge table 14 stores the knowledge of a box-touching state and reliability of a recognizing process and the knowledge of the overlapping portion pattern method.
- the knowledge table 16 stores, for example, the knowledge of the reliability of a detecting process and the knowledge of the method of combining a detecting process and a recognizing process.
- the knowledge table 18 stores, for example, the knowledge about a detail recognizing method.
- a recognizing process can be performed on an input image having various states by selecting and referring to appropriate knowledge for each state, thereby improving the precision of the recognizing process.
- a recognizing process can be performed on an input image having various states by selecting and referring to appropriate recognizing dictionary for each state, thereby improving the precision of the recognizing process.
- a recognizing process can be performed on an input image having various states by selecting appropriate identification function for each state, thereby improving the precision of the recognizing process.
- the non-character recognizing unit 25 performs a non-character recognizing process for each state of an input image, and comprises the deletion line recognizing unit 26 for performing a non-character recognizing process F and a non-character detecting process F on a character with a deletion line, and the noise recognizing unit 28 for performing a non-character recognizing process G and a non-character detecting process G on a noise.
- the deletion line recognizing unit 26 and noise recognizing unit 28 respectively comprise knowledge tables 27 and 29 storing the knowledge of non-character recognizing methods.
- the recognizing process can be performed with high precision with less characters mistaken for non-characters and with less non-characters mistaken for characters by performing the recognizing process on characters and non-characters separately.
- FIG. 8 is a flowchart showing an example of the entire process of the environment recognizing system 11 .
- an input image is pre-processed in step S 1 .
- the pre-process of the input image is performed by labelling the input image binarized by a facsimile, a scanner, etc., and the input image and the labelled image are stored.
- the input image and the labelled image are stored in a way that they can be accessed at any time in the subsequent processes.
- FIG. 9 is a flowchart showing the pre-process of the input image shown in FIG. 8 .
- a binarized input image is labelled in step S 11 so that a link pattern can be extracted and labelled, and the extracted labelled image and the input image are stored.
- the memory capacity can be saved by compressing the labelled link pattern by adding and subtracting the circumscribing rectangle.
- a layout is analyzed in step S 2 shown in FIG. 8 .
- the analysis of the layout is performed by recognizing a text, extracting a ruled line, extracting a character box, determining the type of box and table, determining the existence of a box-touching character, and recognizing a drawing based on the size, the arrangement, etc. of a labelled link pattern.
- FIG. 10 is a flowchart showing the layout analysis shown in FIG. 8 .
- the text is recognized in step S 21 .
- the size of the labelled link pattern is analyzed, and a link pattern of a relatively small size is extracted and defined as a candidate for a character. Then, a text is extracted by integrating the candidate with the adjacent character candidate.
- a ruled line is extracted in step S 22 .
- the ruled line can be extracted by searching a pattern indicating a larger histogram value in the vertical or horizontal direction for the link patterns which are not recognized as text in step S 21 .
- a character box is extracted in step S 23 .
- the character box is extracted by detecting ruled lines corresponding to the four sides of the box from the ruled lines extracted in step S 22 .
- step S 24 the type of box and table is discriminated.
- the type of the character box extracted in step S 23 is discriminated and the attribute of the type of the character box is assigned.
- the attribute of the type of the character box can be a one-character box, a block character box, a free-pitch character box, a table, etc.
- step S 25 it is determined whether or not a box-touching character exists.
- the determination of the box-touching character is performed by detecting a crossing pattern when the character box is searched along the character box line. If a crossing pattern exists, it is determined that the character touches the character box. Another character in the adjacent character box may be partly out of its box and enters the present box. Therefore, in such cases, the character partly out of its own box and entering the present box is not defined as a box-touching character to the present character box.
- a drawing is recognized in step S 26 .
- a link pattern of a relatively large size which has not been assigned the attribute such as a text, character box, table, etc. is assigned an attribute of the drawing.
- a quality is analyzed in step S 3 shown in FIG. 8 .
- the analysis can be a global quality analysis and a local quality analysis.
- an obscurity is detected in a predetermined area if the value obtained by dividing (the total length of the completed portion when the obscure ruled line is completed) by (the total length of each ruled line) is larger than a predetermined value according to the information into which an obscure ruled line in extracting the ruled line is partly integrated.
- FIG. 11 is a flowchart of the quality analysis shown in FIG. 8 .
- a global quality analysis is performed instep S 31 .
- the global quality analysis is performed on the entire document/form, and determines whether or not the threshold used in binarizing the input image is appropriate, whether or not the quality of the document/form with noises given to them after transmitted through facsimile is acceptable, and whether or not obscurity or deformation has been generated.
- a local quality analysis is performed in step S 32 .
- the local quality analysis is performed by checking whether or not obscurity or deformation has been generated or whether or not noises have been generated in each of the areas assigned attributes of a one-character box, a text, a free-pitch character box, a table, etc. in the layout analysis.
- a correction analysis is performed in step S 4 shown in FIG. 8 .
- a deletion line is extracted from an input image, and a character recognizing process can be omitted for the character corrected with the deletion line.
- FIG. 12 is a flowchart showing the correction analysis shown in FIG. 8 .
- a correction feature is extracted in step S 41 .
- a feature of a corrected character is extracted.
- the corrected character can be one of the four types of character, that is, a deformed character, a character removed with double lines, a character removed with a diagonal line, and a character removed with a symbol ‘x’.
- the feature of each of the corrected characters can be extracted by computing the black picture element density, line density, Euler number, histogram value, etc.
- a correction character candidate extracting process is performed in step S 42 .
- a candidate for a corrected character is extracted based on the difference between the corrected character and the non-corrected normal character in a feature space representing the features of the corrected character.
- a character/non-character recognizing control is performed in step S 5 shown in FIG. 8 .
- the character/non-character recognizing control it is determined which should be called from among the basic character recognizing unit 17 , character string recognizing unit 15 , box-touching character recognizing unit 13 , obscure character recognizing unit 19 , deformed character recognizing unit 21 of the character recognizing unit 12 and the deletion line recognizing unit 26 and noise recognizing unit 28 of the non-character recognizing unit 25 , based on the state of an input image extracted in step S 2 through S 4 shown in FIG. 8 .
- the intermediate process result table is read, process order control rule is followed and terminated, and the process is performed under process execution rules.
- the process order control rule shows the procedure of calling which should be called from among the basic character recognizing unit 17 , character string recognizing unit 15 , box-touching character recognizing unit 13 , obscure character recognizing unit 19 , and deformed character recognizing unit 21 of the character recognizing unit 12 and the deletion line recognizing unit 26 and noise recognizing unit 28 of the non-character recognizing unit 25 , based on the state extracted by the environment recognizing system 11 .
- the process execution rule indicates the procedure of the process to be performed next, based on the result of the recognizing process called according to the process order control rule.
- the intermediate process result table includes the state of the input image extracted in step S 2 through S 4 shown in FIG. 8 for each of the areas assigned the attributes of a one-character box, text, free-pitch character box, table, etc. through the layout analysis.
- the processes called according to the input process order control rule are entered in the process order stored in the process order table.
- the basic character recognizing unit 17 is called to perform a recognizing process on the character.
- the character string recognizing unit 15 is called to perform a recognizing process on the text.
- the box-touching character recognizing unit 13 is called to perform a recognizing process on the box-touching character.
- step S 32 When the environment recognizing system 11 determines in step S 32 that the value obtained by dividing (the number of link areas having the sizes of the area, length, and width smaller than the respectively predetermined threshold values) by (the number of all link areas in the predetermined area) is larger than a predetermined value, the obscure character recognizing unit 19 is called to perform a recognizing process on the character in this area.
- the environment recognizing system 11 determines in step S 32 that the value obtained by dividing (the number of link areas indicating the black picture element density larger than a predetermined threshold) by (the total number of link areas in the predetermined area) is larger than a predetermined value
- the deformed character recognizing unit 21 is called to perform a recognizing process on the character in this area.
- the deletion line recognizing unit 26 is called to perform a recognizing process on the candidate for a corrected character.
- the noise recognizing unit 28 is called to perform a recognizing process on the noise.
- FIG. 13 is a flowchart showing the control of character recognizing process/non-character recognizing process shown in FIG. 8 .
- step S 51 the intermediate process result table is read and the process order control rules are executed in step S 51 .
- step S 52 it is determined in step S 52 whether or not the process has been completed. It is determined that the process has been completed when all processes on the intermediate process result table have been completed based on the process order control rules, and all process instruction columns on the intermediate process result table contain completion entries. If it is determined that the process has not been completed yet, then control is passed to step S 53 , returned to step S 51 after performing the process according to the process execution rules, and the above described processes are repeated until it is determined in step S 52 that the process has been completed.
- FIG. 14 is a block diagram showing the configuration of the pattern recognizing apparatus according to an embodiment of the present invention.
- an image storage unit 41 stores a form image.
- a process condition storage unit 42 stores definitions such as the layout structure of the form and read character information, for example, the position, type, and size of a character box, type of characters, number of characters, etc.
- a labelled image storage unit 43 stores compressed and labelled images.
- An environment recognizing system 30 comprises a layout analyzing unit 31 and correction analyzing unit 32 .
- An environment recognizing system 38 comprises a unique character analyzing unit 39 and a completion determining unit 40 .
- a character recognizing system/non-character recognizing system 33 comprises a basic character recognizing unit 34 , a black-character-box touching character recognizing unit 35 , a free-pitch character string recognizing unit 36 , and a deletion line recognizing unit 37 .
- the layout analyzing unit 31 refers to the definitions stored in the process condition storage unit 42 for a labelled image stored in the labelled image storage unit 43 , and extracts ruled lines, character boxes, and black-character-box touching character.
- the method of preliminarily storing format information about the position and size of a character box and information about the pose of a character box as list data, and extracting ruled lines and character boxes according to the list data is disclosed by, for example, Tokukaisho 62-21288 and Tokukaihei 3-126186.
- ruled lines and character boxes can be extracted without entering format information such as the position and size of a character box.
- the correction analyzing unit 32 extracts a candidate for a deletion line.
- the unique character analyzing unit 39 analyzes a unique character having a personal handwritten feature.
- the completion determining unit 40 determines the completion of a character recognizing process, and outputs a character recognition result when it is determined that the process has been completed.
- the basic character recognizing unit 34 recognizes characters detected one by one.
- the black-character-ox touching character recognizing unit 35 removes a character box from a black-character-box touching character, completes an obscure character by removing the character box, and then recognizes the character.
- the free-pitch character string recognizing unit 36 recognizes a character in a character string in consideration of the detection reliability when the character is detected from the character string.
- the deletion line recognizing unit 37 recognizes a deletion line based on the black picture element density of a corrected character, line density, Euler number, histogram, etc.
- An intermediate process result table 44 stores the process order indicating which process is to be performed, the character recognizing system or the non-character recognizing system 33 , and the result of the process, based on the state extracted by the environment recognizing systems 30 and 38 .
- FIG. 15 is a block diagram showing a practical configuration of the character recognizing system to which the pattern recognizing apparatus shown in FIGS. 5 through 7 is applied.
- a central processing unit (CPU) 51 performs various processes.
- a program memory 50 stores a program to be executed by the CPU 51 .
- a image memory 52 stores image data in a bit map format.
- a work memory 53 is used in processing an image.
- a scanner 59 optically reads an image.
- a memory 58 temporarily stores information read by the scanner 59 .
- a dictionary file 60 stores the feature of each character image.
- a display 56 displays a recognition result.
- a printer 57 prints a recognition result.
- An interface circuit 55 functions for the display 56 and printer 57 .
- a bus 54 connects the CPU 51 , program memory 50 , image memory 52 , work memory 53 , memory 58 , dictionary file 60 , and interface circuit 55 .
- the character recognizing system temporarily stores image data read by the scanner 59 in the memory 58 , and develops the image data in the bit map format into the image memory 52 . Then, a pattern extracting process is performed on the binary image data copied from the image memory 52 to the work memory 53 . Based on the result, a character image is detected from the image data read by the scanner 59 . The feature of the detected character image is compared with the feature data stored in the dictionary file 60 to recognize a character. Then, the recognition result is output to the display 56 or printer 57 .
- the pattern recognizing apparatus shown in FIGS. 5 through 7 can be realized as having the function of the CPU 51 for performing the process according to the program stored by the program memory 50 .
- FIG. 16 shows the labelling process performed in step S 11 shown in FIG. 9 .
- the labelling unit 70 extracts a link pattern comprising link picture elements from the input binary image, generates a labelled image assigned a label for each link pattern, and stores the image in a labelled image storage unit 71 .
- a labelled image 73 is generated with labels ‘ 1 ’, ‘ 2 ’, and ‘ 3 ’ assigned to each link pattern.
- the storage capacity of the labelled image storage unit 71 equals 8 times of the number of picture elements of the entire image, thereby requiring a large storage capacity to store the labelled image.
- FIG. 17 shows the method of reducing the storage capacity required for the labelled image storage unit 71 by compressing the labelled image 73 shown in FIG. 16 .
- labels ‘ 1 ’ and ‘ 2 ’ are respectively assigned to link patterns A 1 and A 2 shown in FIGS. 17A and 17B .
- Rectangle B 1 circumscribes link pattern A 1 and rectangle B 2 circumscribes link pattern A 2 as shown in FIG. 17 C.
- Circumscribing rectangles B 1 and B 2 can be specified by the coordinate (x 1 , y 1 ) of the left-top vertex and the coordinate (x 2 , y 2 ) of the right-bottom vertex of circumscribing rectangles B 1 and B 2 as shown in FIG. 17 D.
- rectangle B 1 circumscribing link pattern A 1 overlaps rectangle B 2 circumscribing link pattern A 2 . If rectangle B 1 circumscribing link pattern A 1 does not overlap rectangle B 2 circumscribing link pattern A 2 then the coordinates (x 1 , y 1 ) of the left-top vertexes and the coordinates (x 2 , y 2 ) of the right-bottom vertex of circumscribing rectangles B 1 and B 2 respectively are stored.
- link pattern A 1 can be represented by the difference between rectangular area ( 1 - 1 ) and rectangular areas ( 1 - 2 ) as shown by the following equation where ( 1 - 1 ) indicates the maximum rectangular area belonging to link pattern A 1 , and ( 1 - 2 ) indicates rectangular areas contained in rectangular area ( 1 - 1 ).
- a 1 ( 1 - 1 ) ⁇ ( 1 - 2 )
- link pattern A 2 can be represented by the difference between rectangular area ( 2 - 1 ) and rectangular areas ( 2 - 2 ) plus rectangular area ( 2 - 3 ) as shown by the following equation where ( 2 - 1 ) indicates the maximum rectangular area belonging to link pattern A 2 , ( 2 - 2 ) indicates rectangular area contained in rectangular area ( 2 - 1 ), and ( 2 - 3 ) 25 indicates rectangular areas contained in rectangular areas ( 2 - 2 ).
- a 2 ( 2 - 1 ) ⁇ ( 2 - 2 )+( 2 - 3 )
- the storage capacity required to store labelled images can be smaller by reducing the volume of information representing link patterns by representing the link pattern by a rectangle circumscribing a series of picture elements.
- FIG. 18 is a flowchart showing an embodiment of the text recognizing process in step S 21 shown in FIG. 10 .
- a document is read by a scanner and the image data of the read document is stored in memory in step S 61 .
- step S 62 only a specified strip portion in the horizontal direction is observed among the image data read in step S 61 , the labelling process is performed on the observed portion, and a circumscribing rectangle of black link picture elements is obtained.
- the area of a character string 81 of document A as shown in FIG. 19A is within section A as shown in FIG. 19D
- the area in character string 82 of document B shown in FIG. 19B is within section A as shown in FIG. 19D
- the area in character string 83 of document C shown in FIG. 19C is within section B as shown in FIG. 19D
- only the portions in sections A and B are observed and the labelling processes are performed in the strip portion only to obtain a rectangle circumscribing of linked black picture elements.
- step S 63 Extracted next in step S 63 is only a circumscribing rectangle indicating the difference, smaller than the threshold thy, between the height obtained in step S 62 and a predetermined height ylen, and indicating the difference, smaller than the threshold thx, between the width obtained in step S 62 and the width xlen preliminarily obtained. Then, the coordinate of the circumscribing rectangle in the y direction (vertical direction) is obtained and stored in the memory.
- step S 64 a wide area having the coordinate as a center point along the y direction obtained in step S 63 is observed with the width of the rectangle extracted in step S 62 equaling the width of an image.
- step S 65 a circumscribing rectangle of linked black picture elements is obtained by labelling the wide area obtained in step S 64 .
- step S 66 Extracted next in step S 66 is only a circumscribing rectangle indicating the difference, smaller than the threshold thy, between the height obtained in step S 65 and a predetermined height ylen, and indicating the difference, smaller than the threshold thx, between the width obtained in step S 65 and the width xlen preliminarily obtained.
- the extracted circumscribing rectangle is stored in the memory.
- step S 67 rectangles extracted in step S 66 are sorted based on the x coordinate.
- the pitch is computed from the intervals of the center lines of the extracted rectangle.
- This text extracting method is described by, for example, Tokukaihei 8-171609.
- Described below is an embodiment of the ruled line extracting process in step S 22 shown in FIG. 10 .
- the linked pattern obtained in the labelling process is divided into plural sections in the horizontal and vertical directions.
- the contiguous projection value of the linked pattern is computed within each section of the pattern divided horizontally and vertically.
- ruled lines are extracted by detecting a portion of a predetermined length of line through the approximation of a rectangle.
- a contiguous projection in this example refers to a sum of a projection value of an object row or column and a projection value of a surrounding row or column.
- the projection values of the object row or column are obtained by computing the total number of black picture elements in the row or column.
- FIG. 20 shows the contiguous projection process
- the projection value in row i is p(i), and the contiguous projection value P(i) can be computed by the following equation (1).
- P ( i ) p ( i ⁇ j )+. . . + p ( i )+. . . + p ( i+j ) (1)
- FIG. 21 shows an example of the projection value of a partial pattern.
- a candidate for a straight line forming part of ruled lines can be extracted by detecting a partial pattern indicating the ratio, equal to or smaller than a predetermined threshold, of a contiguous projection value to each of the lengths of vertical and horizontal divisions.
- FIG. 22 is a flowchart showing the ruled line extracting process.
- step S 601 it is determined in step S 601 whether or not the ratio of a contiguous projection value to each of the lengths of vertical and horizontal divisions is equal to or larger than a predetermined threshold. If it is not equal to or larger than the predetermined threshold, then control is passed to step S 602 , and it is assumed that no lines forming part of ruled lines exist.
- step S 601 If it is determined in step S 601 that the ratio of a contiguous projection value to each of the lengths of vertical and horizontal divisions is equal to or larger than a predetermined threshold, then control is passed to step S 603 , and it is assumed that lines forming part of the ruled lines exist.
- step S 604 It is determined in step S 604 whether or not the pattern regarded as a line in step S 603 touches a line above or below the pattern. If it is determined that the pattern does not touch a line above or below the pattern, control is passed to step S 605 , and the pattern is defined as a line forming part of a rectangle.
- step S 604 If it is determined in step S 604 that the pattern regarded in step S 603 as a line touches the lines above and below the pattern, then control is passed to step S 606 and the pattern is integrated into the lines above and below the pattern.
- step S 607 the lines integrated in step S 606 are detected as rectangular lines. For example, the three rectangular lines 85 as shown by (A) in FIG. 23 are integrated, and a rectangular line 86 indicated by (B) in FIG. 23 25 is obtained. Then, ruled lines are extracted by searching for the rectangular lines obtained in step S 605 or S 607 .
- FIG. 24 shows the search method performed while completing obscure ruled lines in the ruled line extracting process in step S 22 shown in FIG. 10 .
- the method of completing the obscure ruled lines is followed to search for a pattern forming a straight line. Even if an area without a pattern exists in the searching direction, it is assumed that a pattern exists in a blank area containing the number of picture element smaller than a predetermined value.
- a blank area 93 containing the number of picture elements smaller than a predetermined value is searched with the picture elements 92 assumed to exist.
- FIG. 25 is a flowchart showing the method of completing an obscure ruled line in the ruled line extracting process.
- the X coordinate of the thinnest portion of the pattern in a predetermined rectangular range is computed in step S 71 .
- step S 72 a center point of a pattern at the X coordinate computed in step S 71 is computed.
- step S 73 the center point of the pattern computed in step S 72 is set as a search start point.
- the search start point is set at the thinnest portion of the pattern because there is a small possibility that the thinnest portion is a character, and the straight line forming part of the character box can be more probably detected.
- step S 74 the search direction for a straight line is set to ‘right’.
- the initial value of the variable K is set to 0 to count the length of the blank area in step S 75 .
- step S 76 the start point obtained in step S 73 is set as the current position of a pattern search.
- step S 77 it is determined whether the current search position set in step S 76 is in the range of the rectangle recognized in step S 71 . If the current search position is not in the range of the rectangle observed in step S 71 , control is passed to stp S 86 .
- step S 77 If it is determined in step S 77 that the current search position is in the range of the rectangle observed in step S 71 , control is passed to step S 78 , and it is determined whether or not a pattern is positioned next to the current search position in the search direction.
- a pattern positioned next to the current search position in the search direction refers to a pattern 102 next to the right of a pattern 101 as shown in FIG. 26 . If it is determined that the pattern 102 is positioned next to the current search position in the search direction, then control is passed to step S 81 , and the pattern 102 next to the current position in the search direction is set as the current position.
- step S 78 If it is determined in step S 78 that a pattern is not positioned next to the current position in the search direction, control is passed to step S 79 , and it is determined whether or not a pattern is positioned diagonally above or below the current position in the search direction.
- a pattern positioned diagonally above or below the current position in the search direction refers to a pattern 104 a or a pattern 104 b diagonally above or below a pattern 103 as shown in FIG. 26 . If it is determined that the patterns 104 a and 104 b are positioned diagonally above or below the current position in the search direction, then control is passed to step S 83 , and the patterns 104 a and 104 b diagonally above or below the current position are defined as the current position. If there are two patterns 104 a and 104 b positioned diagonally above or below the current position in the search direction, then one of the patterns 104 a and 104 b is set as the current search position.
- step S 79 If it is determined in step S 79 that the patterns 104 a and 104 b are not positioned diagonally above or below the current position in the search direction, then control is passed to step S 80 , and it is determined whether or not the variable K for use in counting the length of a blank area is equal to or smaller than a predetermined threshold. If the variable K for use in counting the length of a blank area is equal to or smaller than a predetermined threshold, then control is passed to step S 84 , and the picture element adjacent in the search direction and not forming part of the pattern is defined as the current position. For example, It is assumed that there is a pattern for a blank area 93 having a predetermined number of or less picture elements as shown in FIG. 24 , and a searching process is performed.
- step S 85 the variable K for use in counting the length of a blank area is increased by 1 dot, and control is returned to step S 77 .
- step S 80 If it is determined in step S 80 that the variable K for use in counting the length of a blank area is not equal to or smaller than a predetermined threshold, then control is passed to step S 86 , and it is determined whether or not the search direction is set to ‘right’. If it is not set to ‘right’, then the process terminates.
- step S 86 When the search direction is set to ‘right’ in step S 86 , then control is passed to step S 87 and the search direction is set to ‘left’. Then, the processes in step S 75 through S 85 are similarly repeated as performed when the search direction is set to ‘right’.
- a pattern positioned next to the current search position in the search direction refers to a pattern 106 next to the left of a pattern 105 as shown in FIG. 26.
- a pattern positioned diagonally above or below the current position in the search direction refers to a pattern 108 a or a pattern 108 b diagonally above or below a pattern 107 as shown in FIG. 26 .
- step S 23 the character box extracting process in step S 23 is described below.
- FIG. 27 is a flowchart showing an embodiment of the one-character box extracting process.
- a searching process is performed on a pattern detected in the process shown in FIG. 22 as a line of a rectangle in step S 91 .
- a searching process is performed on a blank area of a predetermined length assuming that a pattern exists as shown in the flowchart in FIG. 25 , and an obscurity problem can be solved.
- step S 92 it is determined in step S 92 after a search in step S 91 whether or not a pattern is disconnected at a predetermined length. If it is not disconnected at the predetermined length, then control is passed to the block character box extracting process shown in FIG. 28 . If the pattern is disconnected at a predetermined length, then control is passed to step S 93 , and a searched lines are combined into a straight line.
- step S 94 straight lines forming a rectangle are extracted from the straight lines detected in step S 93 .
- step S 95 It is determined in step S 95 whether or not the size of a portion encompassed by four straight lines is within a predetermined range for a one-character box in an image. If it is determined that the size of the portion encompassed by four straight lines is within the predetermined range for the one-character box in the image, then control is passed to step S 96 and the portion encompassed by the four straight lines is regarded as a one-character box. If it is determined that the size of the portion encompassed by four straight lines is not within the predetermined range for the one-character box in the image, then control is passed to step S 97 and the portion encompassed by the four straight lines is not regarded as a one-character box.
- FIG. 28 is a flowchart showing an embodiment of the block character box extracting process.
- step S 101 it is determined in step S 101 whether or not a horizontal line detected in a searching process is longer than a predetermined value. If the horizontal line detected in a searching process is shorter than a predetermined value, then control is passed to step S 102 , and the horizontal line is not regarded as a horizontal line forming part of a character box. If it is determined that the horizontal line detected in a searching process is equal to or larger than a predetermined value, then control is passed to step S 102 , and the horizontal line detected in the searching process is regarded as a horizontal line forming part of the character box.
- Two adjacent horizontal lines forming part of a character box are extracted from the horizontal lines extracted in step S 103 as shown in step S 104 .
- a range encompassed by the two horizontal lines forming part of the character box extracted in step S 104 is regarded as a block character box for one row in step S 105 .
- step S 106 vertical lines are detected by extracting the lines forming a rectangle detected in the process shown in FIG. 22 .
- step S 107 the vertical lines detected in step S 106 is searched. It is determined in step S 108 whether or not the vertical lines have reached the horizontal lines which form part of the character box and are detected in step S 104 . If the vertical lines have not reached the horizontal lines forming part of the character box, then control is passed to step S 109 , the vertical lines are removed from the candidates for vertical lines. When the vertical lines touch the upper and lower sides of the character box, control is passed to S 110 and the vertical lines are regarded as candidates for the vertical lines forming part of the character box.
- step S 111 it is determined whether a process object is a regular table-form block character box or an irregular table-form block character box if the process object is a regular table-form block character box, then control is passed to step S 112 , and the interval between the vertical lines regarded as candidates for vertical lines forming part of a character lines is computed in step S 10 , and a histogram indicating the relationship between the interval of the computed vertical lines and the frequency of the vertical lines is computed.
- step S 113 the vertical lines making intervals different from other intervals are removed from candidates for the vertical lines forming part of the character box within a range encompassed by two adjacent horizontal lines forming part of the character box.
- the remaining vertical lines are regarded as vertical lines forming part of the character box, thereby terminating the process.
- step S 11 If it is determined in step S 11 that the process object is an irregular table-form block character box, then all candidates determined in step S 110 for the vertical lines are regarded as vertical lines forming part of the character box, thereby terminating the process.
- step S 24 Described below is the character-box type/table discriminating process in step S 24 shown in FIG. 10 .
- FIG. 29 shows an example of the character box and table extracted in the character box extracting process in step S 23 shown in FIG. 10 .
- FIG. 29A shows a one-character box.
- FIG. 29B shows a free-pitch character box.
- FIG. 29C shows a block character box.
- FIG. 29D shows a regular table.
- FIG. 29E shows an irregular table.
- the one-character box is assigned an attribute of a one-character box.
- the free-pitch character box is assigned an attribute of a free-pitch character box.
- the block character box is assigned an attribute of a block character box.
- the table is assigned an attribute of a table.
- an original input image is reduced using the reduction ratio of 1/n in the OR process, and the process of determining the existence of a box-touching character is then performed.
- the coordinate is set corresponding to each picture element of the image.
- the X coordinate is set along the horizontal direction of the image.
- the Y coordinate is set along the vertical direction of the image.
- the X coordinate increases in the right direction, and the Y coordinate increases downward.
- FIG. 30 is a flowchart showing an embodiment of reducing an input image.
- an original image is input in step S 121 .
- a range of n horizontal picture elements x n vertical picture elements from the left top point of the original image (left top coordinate ( 1 , 1 ), right bottom coordinate (X, Y)) is set in step S 122 .
- step S 123 It is determined in step S 123 whether or not black picture elements exist in the determined range of the original image. If there are black picture elements in the determined range of the original image, then control is passed to step S 124 and the picture elements at the coordinate (X/n, Y/n) of a reduced image are defined as black picture elements. If there are no black picture elements in the determined range of the original image, then control is passed to step S 125 and the picture elements at the coordinate (X/n, Y/n) of a reduced image are defined as white picture elements.
- step S 126 it is determined in step S 126 whether or not the process has been performed up to the right bottom of the original image if the process has not been performed up to the right bottom of the original image, then control is passed to step S 127 , and it is determined whether or not the process has been performed up to the rightmost portion.
- a range of n horizontal picture elements x n vertical picture elements (left top coordinate (x, y), right bottom coordinate (X, Y)) is set to the right of the processed range. If the process has been performed up to the rightmost portion, then a range of n horizontal picture elements x n vertical picture elements (left top coordinate (x, y), right bottom coordinate (X, Y)) is set below the processed range and to the right of the original image. Then, control is returned to step S 123 , and the above described processes are repeated until the reducing process has been completed on the entire range of the original image.
- a rectangular area is enlarged at the side touching the character outside by a predetermined distance.
- the coordinate of the enlarged rectangular area is converted into the coordinate of the original image data.
- a searching process is performed straight along the inside of the character box. If the search line crosses any pattern, it is assumed that a character exists near the character box, the character possibly touches the character box, and that the character 112 of ‘ 4 ’ in the rectangular area encompassed by the character box touches the character box. In this example, the character 112 of ‘ 4 ’ is assumed to touch the lower side 111 of the character box.
- the searching process is performed along the inside of the side 111 of the character box.
- the rectangular area enclosed by the sides of the character box is enlarged outward from the side 111 of the character box touching the character 112 as shown in FIG. 31 C.
- An enlarged rectangular area 113 is defined as a character area containing the character 112 . If it is assumed that the character does not touch the side of a character box, the portion enclosed by the character box is assumed to be a character area.
- the coordinates of the rectangular area 113 shown in FIG. 31C is converted into the coordinate in the original image data to obtain the character area in the original image data from the character area in the compressed image data.
- a rectangular area 116 can be obtained in the original image data as shown in FIG. 31 D.
- a projecting process is performed on a side 114 of the character box in the rectangular area 116 , and the coordinate of the side 114 is computed from the original image data.
- the side 114 of the character box is represented by a rectangle in the form of a predetermined length of strip.
- the pattern in the rectangular area 116 is transmitted to be processed in a character completing process, and a completing process is performed on a character 115 touching the side 114 of the character box based on the coordinate of the side 114 of the character box computed according to the original image data.
- FIG. 32 is a flowchart showing an embodiment of the process of determining the existence of a character touching a character box.
- a rectangle is represented in compressed image data in step S 131 in, for example, r the process shown in FIG. 30 .
- a rectangular portion encompassed by four vertical and horizontal lines is extracted in step S 132 .
- step S 133 the coordinates indicating the left top corner and right bottom corner inside the rectangle are computed.
- step S 134 the compressed image is searched for along the four sides (upper horizontal side, lower horizontal side, right vertical side, and left vertical side) of the rectangle inside the character box.
- step S 135 If one of the four sides crosses the image pattern during the searching process in step S 135 , then it is assumed that a character touches the side currently being searched.
- the coordinates of the rectangle inside the character box are converted into the coordinates in the original image data in step S 136 so that a rectangular area in the original image data can be computed from the rectangular area in the compressed image data.
- step S 137 the rectangular area computed in step S 136 is defined as a character area in the original image data.
- step S 138 It is determined in step S 138 whether or hot a character touches the character box in the process in step S 135 .
- the box touching character range obtaining process is performed in step S 139 through S 143 .
- the character area is enlarged outward from the side touching a character in step S 139 , and the position at a predetermined distance outside the position of the character area computed in step S 137 is defined as the end of the character area.
- step S 140 the coordinate of the position of the side of the character box in the original image data is computed from the coordinate of the position of the side of the character box in the compressed image data by converting the coordinate of the position of the side of the character box contained in the character area computed in step S 139 into the coordinate in the original image.
- step S 141 a projecting process is performed in the horizontal direction for horizontal sides and in the vertical direction for vertical sides on the character box area in the original image data obtained based on the coordinate of the position of the character box in the original image data computed in step S 140 .
- step S 142 it is assumed that the area indicating projection values larger than a predetermined value refers to the coordinates of the character box in the original image.
- step S 143 the computed coordinate of the character area in the original image and the coordinate indicating the position of the character box in the character area are transmitted to the character completing process.
- step S 144 the computed coordinate of the character area in the original image is defined as a character area.
- step S 41 Described below are the correction feature extracting process in step S 41 and the correction character candidate extracting process in step S 42 shown in FIG. 12 .
- FIG. 33 shows an embodiment of a corrected character.
- a character is corrected with deletion lines.
- a character can be deleted with ‘x’ as shown in FIG. 33A ; with double horizontal lines as shown in FIG. 33B ; with diagonal lines as shown in FIG. 33C ; with random lines as shown in FIG. 33D ; and by painting black as shown in FIG. 33 E.
- the features of a deleted character can be extracted from the above described deleted character.
- the features can be the line density in a predetermined direction, an Euler number, and the density of black picture elements.
- the ‘line density in a predetermined direction’ is obtained by counting the changes from white picture elements into black picture elements (or black picture elements into white picture elements) while an image in a rectangular area is scanned in a predetermined direction.
- the predetermined direction refers to the direction vertical to the line predicted as a deletion line.
- FIG. 34A shows an example of counting the maximum line density about the character ‘ 6 ’ in the vertical direction.
- the maximum line density in the vertical direction is 3.
- the ‘line density in a predetermined direction’ of a deleted character tends to increase as compared with the ‘line density in a predetermined direction’ of a normal character.
- Computing the ‘line density in a predetermined direction’ extracts a candidate for a deleted character.
- An Euler number ‘E’ is obtained by subtracting the number H of holes in an image from the number C of the elements linked in the image.
- FIG. 34B shows an example of two elements linked in an image and only one hole in the image.
- the Euler number of a corrected character tends to be a negative number indicating a large absolute number while the Euler number of a normal character tends to be a number (2 ⁇ 1) indicating a small absolute number. Therefore, computing the Euler number extracts a candidate for a deleted character.
- the density D of black picture elements refers to a ratio of the area B (number of black picture elements) of an object image to the area S of the rectangle circumscribing the object image.
- the ‘density of black picture elements’ of a deleted character tends to be higher than the ‘density of black picture elements’ of a deleted character.
- Computing the ‘density of black picture elements’ extracts a candidate for a deleted character.
- FIG. 35 is a block diagram showing an embodiment of the configuration of the basic character recognizing unit 17 .
- a feature extracting unit 121 extracts the features of a character from an unknown input character pattern, and represents the extracted features by the feature vectors.
- a basic dictionary 122 stores the feature vectors of each character category.
- a collating unit 123 collates the feature vector of an unknown character pattern extracted by the feature extracting unit 121 with the feature vector of each character category stored in the basic dictionary 122 , and computes the distance D ij (i indicates the feature vector of the unknown character, and j indicates the feature vector of the category of the basic dictionary 122 ) between feature vectors in a feature space.
- the category j indicating the shortest distance D ij between the feature vectors as an unknown character i.
- the distance D ij between the feature vectors in the feature space can be computed using a Euclidean distance ⁇ (i ⁇ j) 2 , a city block distance ⁇
- FIG. 36 shows an example of computing a feature vector.
- the character ‘ 2 ’ is written in a column of 20 blocks (5 in vertical direction and 4 in horizontal direction).
- a black-painted block indicates ‘ 1 ’ and a white-painted block indicates ‘ 0 ’.
- the blocks are checked sequentially from the left-top block to the right-bottom block.
- the characters ‘ 1 ’ or ‘ 0 ’ are rearranged as feature vectors.
- FIG. 37 shows an example of computing the distance D ij between the feature vectors using the city block distance d (i, j).
- the city block distanced (i, j) between the i-th feature vector x i and the j-th feature vector x j is defined as follows.
- the basic dictionary 12 contains the feature vectors of the character: categories of ‘ 1 ’, ‘ 2 ’, ‘ 3 ’, and ‘ 4 ’.
- 11.
- 1.
- 3.
- 11.
- the city block distances d (i, j) between the feature vector and each of the feature vector 1 of the character category of ‘ 1 ’, the feature vector 2 of the character category of ‘ 2 ’, the feature vector 3 of the character category of ‘ 3 ’, and the feature vector 4 of the character category of ‘ 4 ’ indicates the minimum value.
- a local partial pattern is extracted as a character segment, and the change in position and angle of the character segment of an unknown character is compared with the change in position and angle of the character segment preliminarily stored in the segment dictionary.
- the character can be recognized by making the unknown character correspond to character categories.
- FIG. 38 shows the method of extracting a character segment.
- FIG. 38A shows a binary image pattern about the character ‘ 2 ’, and the portion with diagonal lines refers to a character represented by black picture elements.
- FIG. 38B shows the outline of the character extracted from the binary image pattern shown in FIG. 38 A.
- the dotted-line portion indicates the original binary image pattern.
- FIG. 38C shows the outline shown in FIG. 38B into the character segments S 1 and S 2 and the end portions T 1 and T 2 .
- the end portions T 1 and T 2 correspond to the start and end of the character ‘ 2 ’ shown in FIG. 38 A.
- FIG. 39 shows the method of detecting an end point.
- the end point is detected as a portion where the slope of the outline indicates a sudden change.
- the three points A, B, and C at predetermined intervals are set on the outline S.
- the area making an angle of ⁇ smaller than a predetermined value at the point A on the outline between the points B and C is detected as an end portion.
- the representative points X, Y, and Z are set, for example, at predetermined intervals on the character segment.
- the angles made at the consecutive representative points X, Y, and Z are obtained, and an accumulative value of angle changes from the first representative point to each of the other representative points on the character segment is obtained as the features at the representative points X, Y, and Z.
- FIG. 40 shows the method of detecting a change in angle.
- the representative points X, Y, and Z are set at predetermined intervals on the outline S.
- the vector XY from the representative point X to the representative point Y and the vector YZ from the representative point Y to the representative point Z are generated.
- the angle ⁇ 2 between the vectors XY and YZ is defined as a change in angle at the representative point Y.
- the change in angle at the representative point X on the outline S as an initial value of a change in angle refers to the angle ⁇ 1 made by the vector GX from the center of the gravity G of a character to the representative point X and the vector XY.
- the feature at each of the representative points X, Y, and Z is represented by an accumulative value of changes in angle from the representative point X having the initial value of a change in angle to each of the representative points Y and Z.
- the feature at the representative point Y is expressed as ⁇ 1 + ⁇ 2 .
- the representative point of a character segment of the unknown character is made to correspond to the representative point on the character segment stored in the segment dictionary. That is, the distance between the r accumulative value of changes in angle at the representative point on the character segment of the unknown character and the accumulative value of changes in angle of the representative point on the character segment stored in the segment dictionary is computed.
- the representative point on the character segment in the segment dictionary indicating the shortest distance is made to correspond to the representative point of a character segment of the unknown character.
- FIG. 41A shows the correspondence between the representative point of the character segment on an unknown character and the representative point on the character segment of the segment dictionary.
- the representative points a 1 through a 8 refer to the representative points on the character segment of an unknown character.
- the representative points b 1 through b 8 refer to the representative points on the character segment stored in the segment dictionary.
- Each of the representative points a 1 through a 8 on the character segment of the unknown character corresponds to one of the representative points b 1 through b 8 on the character segment stored in the segment dictionary.
- the representative point on the character segment of the unknown character corresponding to the reference point on the character segment stored in the segment dictionary is set as a check point.
- FIG. 41B shows the correspondence between the reference point and the check point.
- the reference points d 1 and d 2 of the character segment stored in the segment dictionary correspond respectively to the check points c 1 and c 2 of the character segment of the unknown character.
- the check information about the check points c 1 and c 2 of the character segment of the unknown character is computed.
- the check information can be, for example, absolute positional information about an individual check point as to where the check point exists in the entire character image; relative positional information about two check points indicating the distance, direction, etc. between the two check points; and information on two or more check points about the changes in angle, linearity, etc. among the check points.
- the character category of the character segment which satisfies the condition and is stored in the segment dictionary is output as a recognition result of the unknown character.
- the character pattern shown in FIG. 38A can be recognized as belonging to the character category of ‘ 2 ’ by computing the change in angle from the check point c 1 to the check point c 2 on the character segment shown in FIG. 41B along the character segment when a character image of the character segment indicating the change in angle of 60 degrees or more belongs to the character category of ‘ 2 ’ in the segment dictionary stored corresponding to the character segment.
- FIG. 42 is a flowchart showing the character recognizing process by the detail identifying method.
- a list, for example, to be character recognized is scanned by a scanner, and the read character image is binarized into a binary monochrome image in step S 150 .
- step S 151 a character segment is extracted from the binary image obtained in step S 150 .
- step S 152 a character segment which is not associated with the character segment of an unknown character is detected from a plurality of character segments stored in the segment dictionary.
- step S 153 the character segment retrieved from the segment dictionary is associated with the character segment of the unknown character.
- step S 154 a check point is selected from the representative points on the character segment of the unknown character, and the check information about the check point is computed.
- step S 155 the character segment retrieved from the segment dictionary is compared with the character segment of the unknown character according to the check information computed in step S 154 , and a candidate for the character corresponding to the unknown character is determined by determining whether or not the check information about the character segment retrieved from the segment dictionary matches the check information about the character segment of the unknown character.
- step S 156 when a candidate for the character is determined in the process of determining a candidate for the unknown character, the character category corresponding to the character segment retrieved in step S 153 is output as a recognition result. If a candidate for the character is not determined, control is passed to step S 157 , and it is determined whether or not there is an unprocessed character segment which is not associated with the character segment of the unknown character. If there is an unprocessed character segment in the segment dictionary, then control is returned to step S 152 and the above described process is repeated.
- box-touching character recognizing unit 13 shown in FIG. 9 .
- FIG. 43 shows the character completing process performed by the box-touching character recognizing unit 13 .
- label 1 is assigned to the binary image of the character pattern 131 representing ‘ 3 ’ and the character box 132 touching the character. Then, the character pattern 131 indicating ‘ 3 ’ is divided into 3 portions as shown in FIG. 43B by extracting the character box 132 from the binary image shown in FIG. 43A , and by removing the character box 132 . Thus, the three character strokes are generated with labels 1 , 2 , and 3 assigned.
- the geometric structure such as the distance, direction, etc. of the three labelled character strokes assigned labels 1 , 2 , and 3 is evaluated and the character is completed.
- the three character strokes assigned labels 1 , 2 , and 3 are linked, and a character completed pattern 132 indicating ‘ 3 ’ with label 1 is generated as shown in FIG. 43 C.
- a recognizing process is performed on the character restored in the character completing process as a candidate for a character to be recognized.
- the character is collated with the standard pattern entered in the character category dictionary, and a code of a character category indicating the smallest difference is output.
- FIG. 44 shows the re-completing process performed by the box-touching character recognizing unit 13 .
- the character stroke parallel with the character box touches the character box and is removed when the character box is removed, then the character stroke is interpolated.
- the box-touching character is extracted based on the linkage using the labels, and the character stroke parallel with the character box can be interpolated by detecting the matching in linkage between the completed character pattern completed by the character completing process and the box-touching character.
- a binary image of a character pattern 141 indicating ‘ 7 ’ touching a character box 142 as shown in FIG. 44A is assigned label 1 .
- the character box 142 is extracted from the binary image shown in FIG. 44A , and the character box 142 is removed.
- the character pattern 141 indicating ‘ 7 ’ is divided into 3 portions and three character strokes are generated with labels 1 , 2 and 3 assigned.
- the geometric structure such as the distance, direction, etc. of the three labelled character strokes assigned labels 1 , 2 , and 3 is evaluated and the character is completed.
- the three character strokes assigned labels 1 and 2 are linked, and a character completed pattern 142 formed of two character strokes assigned with labels 1 and 2 is generated as shown in FIG. 44 C.
- the character is completed only between the portion assigned label 1 and the portion assigned label 2 as shown in FIG. 44 B. However, the character is not completed between the portion assigned label 1 and the portion assigned label 3 as shown in FIG. 44 B. The character is completed in the re-completing process between the portion assigned label 1 and the portion assigned label 3 as shown in FIG. 44 B.
- a character stroke parallel with the character box is interpolated by preliminarily extracting the box-touching character based on the linkage using labels, and by detecting the matching in linkage between the pattern shown in FIG. 44 C and the box-touching character. That is, the patterns assigned labels 1 and 2 as shown in FIG. 44C have been linked before removing the character box as shown in FIG. 44 A. Therefore, the patterns assigned labels 1 and 2 as shown in FIG. 44C are linked using the character stroke parallel with the character box.
- the binary image ‘ 7 ’ divided into two character strokes assigned labels 1 and 2 as shown in FIG. 44C can be completed, and a re-completed pattern 143 indicating ‘ 7 ’ with label 1 is generated as shown in FIG. 44 D.
- the recognizing process is performed on the character restored in the character completing process as a candidate for a character to be recognized.
- the character is collated with the standard pattern entered in the character category dictionary, and a code of a character category indicating the smallest difference is output.
- the character completed pattern 142 shown in FIG. 44C is recognized as belonging to the category of the character .
- the character completed pattern 143 shown in FIG. 44D is recognized as belonging to the category of the character ‘ 7 ’. After it is determined that ‘ 7 ’ indicates a smaller difference than , the character is finally recognized as ‘ 7 ’ and the character code is output.
- recognizing process is performed by the box-touching character recognizing unit 13 shown in FIG. 7 by referring to the knowledge table 14 .
- FIG. 45 shows an example of recognizing a box-touching character by learning a pair of a character and its misread character and entering it in the knowledge table 14 .
- label 1 is assigned to the binary image of the character pattern 151 representing ‘ 2 ’ and the character box 152 touching the character. Then, the character pattern 151 indicating ‘ 2 ’ is divided into 2 partial patterns with labels 1 and 2 as shown in FIG. 45B by extracting the character box 152 from the binary image shown in FIG. 45A , and by removing the character box 152 .
- the two partial patterns with labels 1 and 2 shown in FIG. 45B are linked and the character completed pattern 153 is generated in the character completing process.
- the lower stroke of the character pattern 151 indicating ‘ 2 ’ touches the lower side of the character box 152 , and the touching portion of the character almost completely overlaps the character box 152 . Therefore, even the re-completing process cannot complete the lower stroke of the character pattern 151 indicating ‘ 2 ’, and, there is a high possibility that the character ‘ 2 ’ can be misread as ‘ 7 ’.
- the box-touching character is not partly written outside the character box, but completely overlaps the character box. Therefore, if it can be misread as any other character, the box-touching character should be correctly recognized by entering it through learning a pair of a character and its misread character.
- Described below is the method of recognizing a box-touching character by entering a pair of the character and its misread character.
- FIG. 46 is a block diagram showing the configuration for learning a pair of a character and its misread character in the box-touching character recognizing unit 13 shown in FIG. 7 .
- An automatic box-touching character generating unit 161 generates a box-touching character by making a character box overlap a learning character input as not touching the character box. By the method of changing the learning character relative to its character box, a plurality of box-touching characters can be generated for a single learning character.
- a learning character 168 indicating ‘ 2 ’ is input to the automatic box-touching character generating unit 161 , and a box-touching character 169 is generated with the lower stroke of the character ‘ 2 ’ overlapping the lower side of the character box.
- the information generated by the automatic box-touching character generating unit 161 is entered in a knowledge table 167 .
- the variation of a character relative to its character box can be, for example, a displacement, a variation in size, a variation in pose, etc.
- the variation of a character box can be, for example, a variation in pose, a variation in width of a character box, a variation in size, a convexity and concavity of a character box, etc.
- the following parameters indicates the amount of a change in each of the above described variations.
- the x axis indicates the vertical direction and the y axis indicates the horizontal axis.
- a box-touching character is generated by providing a learning character with an operation F (dx, dy, dsx, dsy, d ⁇ , w, fsx, fsy, f ⁇ , f ⁇ ).
- FIG. 47 shows an example of generating a box-touching character by assigning a character box 172 to a learning character 171 indicating ‘ 7 ’.
- a box-touching character can be generated by performing the changing operation F (dx, dy, dsx, dsy, d ⁇ , w, fsx, fsy, f ⁇ , f ⁇ ) for the learning character 171 and the character box 172 to make the learning character 171 and the character box 172 overlap each other.
- the changing operation F (dx, dy, dsx, dsy, d ⁇ , w, fsx, fsy, f ⁇ , f ⁇ ) is performed while, for example, fixing the position of the center of gravity of the character box 172 .
- FIG. 48 shows an example of each type of a box-touching character generated for the learning character ‘ 3 ’ with the size variation fsx in the x direction and the size variation fsy in the y direction fixed and the size of character box fixed.
- the character ‘ 3 ’ is partly outside the lower side of the character box (lower displacement).
- the character ‘ 3 ’ touches the upper, lower, left, and right sides of the character box ‘ 3 ’.
- the rectangle circumscribing ‘ 3 ’ equals the character box.
- (F) in FIG. 48 indicates an example of the type of ‘variation in concavity and convexity of a character box’, where each element f ⁇ [i] of the amount of change f ⁇ [L] is controlled.
- a character box removing unit 162 shown in FIG. 46 extracts only the character box from the box-touching character generated by the automatic box-touching character generating unit 161 , and outputs to a character completing unit 163 the image data on the obscure character obtained by removing the character box.
- the character completing unit 163 evaluates and completes the geometric structure such as the distance, direction, etc. of the labelled character strokes on the image data from which the character box has been removed by the character box removing unit 162 .
- FIG. 46 shows an example of generating a character completed pattern 170 by completing a box-touching character by the character completing unit 163 after removing the character box from the box-touching character 169 generated by the automatic box-touching character generating unit 161 .
- a re-completing unit 164 preliminarily extracts the box-touching character based on the linkage using labels in the area where the character completing unit 163 has not completed the image data, and completes a character stroke parallel with the character box by detecting the matching in linkage between the pattern completed by the character completing unit 163 and the box-touching character.
- the character completed pattern completed by the character completing unit 163 and the re-completed pattern completed by the re-completing unit 164 are input to a basic character recognizing unit 165 .
- the basic character recognizing unit 165 performs a character recognizing process on the character completed pattern completed by the character completing unit 163 and the re-completed pattern re-completed by the re-completing unit 164 ; Then, the basic' character recognizing unit 165 outputs the recognition result about each learning character to a character box touching state and recognition knowledge acquiring unit 166 .
- the character box touching state and recognition knowledge acquiring unit 166 compares a recognition result output from the basic character recognizing unit 165 with preliminarily provided solution data and obtains the recognition ratio for the entire sample data. Then, the character box touching state and recognition knowledge acquiring unit 166 enters in the knowledge table 167 this recognition ratio as reliability, and the combination of the misread. (mis-recognized) character and the correct character as a pair of a character and its misread character. The above described pair of a character and its misread character are entered in, for example, character codes.
- the character box touching state and recognition knowledge acquiring unit 166 also extracts a parameter indicating the feature of the state of a character and a character box touching the character, and enters the feature in the knowledge table 167 .
- the knowledge table 167 contains for each character category the recognition ratio, together with the pair of a character and its misread character, for the character in various touching states between the character and the character box.
- FIG. 49 shows an example of the knowledge table 167 generated through learning a character.
- the character box touching state and recognition knowledge acquiring unit, 166 enters in the knowledge table 167 the ‘amount of change’, ‘width of the sides of a character box’, a ‘character and its misread character’, and a reliability.
- the pair (L 1 , L 2 ) of a character and its misread character indicates that the character ‘L 1 ’ may be actually mis-recognized as ‘L 2 ’.
- the character codes for characters ‘L 1 ’ and ‘L 2 ’ are entered for the corresponding characters ‘L 1 ’ and ‘L 2 ’.
- dx ‘ ⁇ 3’ ⁇ ‘+3’
- the amount of change entered in the knowledge table 167 as ‘displacement at a lower position’ can be the displacement dx in the x direction, the displacement dy in the y direction, and other values.
- dx ‘ ⁇ 3’ ⁇ ‘+3’
- dy ‘ ⁇ 3’ ⁇ ‘+3’
- w 5
- a character recognizing method is followed on a pair (L 1 , L 2 ) of a character and its misread character whose reliability is equal to or lower than a predetermined threshold (for example, 90%) in a way that the reliability is equal to or higher than the predetermined threshold.
- the learned character recognizing method is entered in the knowledge table 167 .
- a circumscribing rectangle 180 circumscribing the character completed pattern completed by the character completing unit 163 or the re-completed pattern re-completed by the re-completing unit 164 is divided into m ⁇ n divided areas in m columns by n rows. Then, with diagonal lines shown in FIG. 51B , the upper half m/2 ⁇ n area of the circumscribing rectangle 180 is emphasized to recognize the character again.
- the knowledge table 167 shown in FIG. 49 contains, in the line of a pair of a character and its misread character ( 2 , 7 ), the ‘highlighted area’ as a re-recognizing method the ‘m/2 ⁇ n’ area as a re-recognition area, and ‘95%’ as the re-recognition reliability.
- FIG. 52A is an example that the lower portion of the character pattern indicating ‘ 2 ’ touches a character box 182 .
- the character completing unit 163 obtains a character completed pattern 183 similar to ‘ 7 ’ shown in FIG. 52B.
- a circumscribing rectangle 184 shown in FIG. 52C is computed corresponding to the character completed pattern 183 . If, as shown in FIG. 51 , the circumscribing rectangle 184 is divided into m ⁇ n areas, and the upper half m/2 ⁇ n partial area 185 is especially emphasized when the character is recognized, then there is a high probability that the character completed pattern 183 can be recognized as ‘ 2 ’.
- FIG. 53 is a flowchart showing the method of re-recognizing a character pattern by emphasizing a specific area.
- the data of a pair of a character and its misread character indicating a lower reliability is retrieved from the knowledge table 167 in step S 161 .
- a character pattern as binary learning data corresponds to the left character of the pair of the character and its misread character.
- the character completed pattern completed by the character completing unit 163 or the re-completed pattern re-completed by the re-completing unit 164 are input.
- the character completed pattern or the re-completed pattern is prescribed by an amount-of-change parameter entered in the knowledge table 167 , and can be represented by plural forms of patterns even in the same category.
- the character pattern as the learning data input in step S 161 , and the character completed pattern completed by the character completing unit 163 or the re-completed pattern re-completed by the re-completing unit 164 are divided into m ⁇ n areas in step S 162 .
- step S 163 a character recognizing process is performed on the X ⁇ Y partial pattern in the m ⁇ n area, and the recognition ratio z is obtained.
- the above described X ⁇ Y partial pattern is a re-recognition area where X and Y indicates the length in the X and Y directions respectively in the m ⁇ n area, and X ⁇ m and Y ⁇ n.
- the above described recognition ratio z indicates the probability that a correct solution can be output with characters recognized using the above described X ⁇ Y partial patterns.
- the character recognition result of the partial pattern of a character pattern as learning data is regarded as a solution.
- the character recognition result on plural partial patterns about the character completed pattern completed by the character completing unit 163 or the re-completed pattern re-completed by the re-completing unit 164 is compared with the character recognition result on the partial pattern of the character pattern as learning data.
- the recognition ratio z of the partial pattern about the character completed pattern completed by the character completing unit 163 or the re-completed pattern re-completed by the re-completing unit 164 is obtained.
- step S 164 it is discriminated whether or not the recognition ratio z is larger than the maximum recognition ratio max.
- the maximum recognition ratio max is a variable storing the maximum value of the recognition ratio z obtained while a partial pattern of X ⁇ Y varies. First, an initial value is set (for example, ‘0’)
- step S 165 If the recognition ratio z is larger than the maximum recognition ratio max, then control is passed to step S 165 to substitute the recognition ratio z for the maximum recognition ratio max, and control is passed to step S 166 to check whether or not the lengths X and Y is variable. If the recognition ratio z is equal to or smaller than the maximum recognition ratio max in step S 164 , then control is immediately passed to step S 166 .
- Changing the lengths X and Y is changing the size of the lengths X and Y, and also includes a position change in the m ⁇ n area of the partial pattern of X ⁇ Y.
- step S 166 If it is discriminated in step S 166 that the lengths X and Y are variable, control is returned to step S 163 , the lengths X and Y are changed, and a new partial pattern of X ⁇ Y is determined, and characters are recognized in the partial pattern.
- step S 163 through S 166 are repeated until it is determined in step S 166 that the lengths X and Y cannot be changed. If it is determined in step S 166 that the lengths X and Y cannot be changed, the maximum identification ratio max and the partial pattern of X ⁇ Y from which the maximum identification ratio max has been obtained are entered in the knowledge table 167 as a re-recognition reliability and a re-recognition area respectively. The ‘area emphasis’ is entered as a re-recognizing method in the knowledge table 167 .
- FIG. 53 is a flowchart showing an example of learning the method of re-recognizing a character using the area emphasizing method.
- the character re-recognizing method can also be learned by any other methods than the area emphasizing method.
- FIG. 54 is a block diagram showing the configuration with which a box-touching character is recognized according to the knowledge table 167 obtained through learning.
- a box-touching state detecting unit. 191 detects the touching state between a character box and a character when an unknown box-touching character is input.
- This example shows the lower portion of a box-touching character ‘pattern 201 (‘ 2 ’) partially overlapping the lower side of the character box as indicated by (A) shown in FIG. 54 and the lower portion of a box-touching character pattern 203 (‘ 2 ’) completely overlapping the lower side of the character box as indicated by (B) shown in FIG. 54 .
- the box-touching state detecting unit 191 detects the box-touching character pattern 201 and box-touching character pattern 203 .
- a character box removing unit 192 removes a character box from the box-touching character pattern detected by the box-touching state detecting unit 191 .
- a character completing unit 193 evaluates and completes the geometric structure such as the distance, direction, etc. of the labelled character strokes on the character pattern from which the character box has been removed by the character box removing unit 192 .
- a re-completing-unit 194 preliminarily extracts the box-touching character based on the linkage using labels in the area where the character completing unit 193 has not completed the image data, and completes a character stroke parallel with the character box by detecting the matching in linkage between the pattern completed by the character completing unit 193 and the box-touching character.
- a re-completed pattern 202 shows a pattern completed in the re-completing process performed by the re-completing unit 194 on the box-touching character pattern 201 indicated by (A) shown in FIG. 54.
- a re-completed pattern 204 shows a pattern which cannot be completed in the re-completing process performed by the re-completing unit 194 on the box-touching character pattern 203 indicated by (B) shown in FIG. 54 .
- the basic character recognizing unit 195 performs a character recognizing process on the character completed pattern completed by the character completing unit 193 and the re-completed pattern re-completed by the re-completing unit 194 .
- the character code of ‘ 2 ’ is output for the re-completed pattern 202 indicated by (A) shown in FIG. 54
- the character code of ‘ 7 ’ is output for the re-completed pattern 204 indicated by (B) shown in FIG. 54
- the character codes obtained as the recognition result are output to a character box touching state and recognition knowledge acquiring unit 196 .
- the character box touching state and recognition knowledge acquiring unit 196 obtains the type of variation according to the positional information about the rectangle circumscribing the character completed pattern completed by the character completing unit 193 or the re-completed pattern re-completed by the re-completing unit 194 , and according to the positional information and width information about the character box extracted from the box-touching character pattern 201 indicated by (A) shown in FIG. 54 or the box-touching character pattern 203 indicated by (B) shown in FIG. 54 .
- a change to a character relative to its character box such as the displacement, variation in size, variation in pose, etc. as shown in FIG. 47 , or a change to a character box such as a variation in pose, variation in width of a character box, convexity and concavity of a character box, etc. is obtained. Furthermore computed is the amount of change dx, dy, dsx, dsy, d ⁇ , w, fsx, fsy, f ⁇ , or f ⁇ for the type of each variation as obtained above.
- the knowledge table 167 is searched using, as key items, the computed variation type information, amount-of-change information, and a character code input from the basic character recognizing unit 195 . It is checked whether or not the knowledge table 167 stores a line containing the variation type information, amount-of-change information, and a pair of a character and its misread character matching the key items.
- the line matching the key item exists, then it is discriminated whether or not the reliability stored in the line is equal to or larger than a predetermined threshold. If it is smaller than the threshold, then the character completed pattern completed by the character completing unit 193 or the re-completed pattern re-completed by the re-completing unit 194 is output to a character re-recognizing unit 197 . The characters are recognized again by the re-recognizing method entered in the line.
- a box-touching character in unknown image data is re-recognized by a method other than the method by the basic character recognizing unit 195 using the character completed pattern completed by the character completing unit 193 , the re-completed pattern re-completed by the re-completing unit 194 , or binary image data of an unknown character. Then a character code obtained in the re-recognizing process is output.
- the character box touching state and recognition knowledge acquiring unit 196 obtains the type of variation and an amount of change according to the positional information about the rectangle circumscribing the re-completed pattern 204 , and the positional information and width information about the character box extracted from the box-touching character pattern 203 .
- the ‘displacement at a lower position’ is computed as the type of variation.
- the lines corresponding to the key items store a pair ( 2 and 7 ) of a character and its misread character, and the reliability of the character code ‘ 7 ’ recognized by the basic character recognizing unit 195 is 770, thereby referring to that the character ‘ 2 ’ is misread for ‘ 7 ’ at the probability of 23%.
- the character re-recognizing unit 197 re-recognizes the box-touching character pattern 203 contained in the unknown image data by a method other than the method followed by the basic character recognizing unit 195 .
- the character re-recognizing unit 197 refers to the line corresponding to the key item on the knowledge table 167 to specify the re-recognizing method.
- the character re-recognizing unit 197 is informed of the ‘area emphasizing method’ as a re-recognizing method, and of an upper half m/2 ⁇ n area 205 of the re-completed pattern 204 as a re-recognition area when the ‘area emphasizing process’ is performed. It is also informed of the re-recognition reliability of 95%.
- the character re-recognizing unit 197 re-recognizes only the upper half area 205 of the re-completed pattern 204 by the re-recognizing method entered in the knowledge table 167 .
- the character re-recognizing unit 197 is informed that the upper half area 205 of the re-completed pattern 204 matches a partial area 207 of a character pattern 206 corresponding to the character code ‘ 2 ’ at a probability of 95%, and matches a partial area 209 of a character pattern 208 corresponding to the character code ‘ 7 ’ at a probability of 5%, and outputs the character code ‘ 2 ’ as a recognition result of the character touching the character box of the box-touching character pattern 203 of an unknown character.
- FIG. 55 is a flowchart showing the operations of the character box touching state and recognition knowledge acquiring unit 196 .
- an amount of change to a character relative to its character box is computed based on the character box extracted from an unknown box-touching character pattern and the character pattern separated from the box-touching character pattern, and the knowledge table 167 is searched using the amount of change as a key item in step S 171 . Then, it is checked whether or not the knowledge table 167 stores a line containing the amount of change matching the computed amount of change.
- step S 172 If a line containing a matching amount of change exists, control is passed to step S 172 , and it is determined whether or not the line containing, the matching amount of change stores the line containing in the pair of a character and its misread character the character code (character recognition code) input from the basic character recognizing unit 195 .
- the top line on the knowledge table 167 is detected with the character ‘ 2 ’ indicating the displacement at a lower position as shown in FIG. 49 .
- step S 173 if the line containing the matching amount of change stores the line containing in the pair of a character and its misread character the character code input from the basic character recognizing unit 195 , the re-recognition reliability entered in the corresponding line on the knowledge table 167 is compared with the reliability computed by the basic character recognizing unit 195 . It is determined whether or not the re-recognition reliability entered in the corresponding line on the knowledge table 167 is larger than the reliability computed by the basic character recognizing unit 195 .
- the re-recognition reliability entered in the top line on the knowledge table 167 shown in FIG. 49 and the reliability computed by the basic character recognizing unit 195 are ‘95%’ and ‘77%’ respectively.
- the re-recognition reliability entered in the corresponding line on the knowledge table 167 is larger than the reliability computed by the basic character recognizing unit 195 .
- step S 174 If the re-recognition reliability entered in the corresponding line on the knowledge table 167 is larger than the reliability computed by the basic character recognizing unit 195 , control is passed to step S 174 and it is determined whether or not the re-recognition reliability entered in the corresponding line on the knowledge table 167 is larger than a predetermined threshold th 1 . If it is larger than the threshold th 1 , then control is passed to step S 175 , and the re-recognizing method and the re-recognizing area entered in the line on the knowledge table 167 and detected in step S 172 are referred to.
- step S 176 a re-recognition area shown on the knowledge table 167 is detected from the character completed pattern completed by the character completing unit 193 or the re-completed pattern re-completed by the re-completing’ unit 194 , and a character recognizing process is performed on the detected area by the re-recognizing method shown on the knowledge table 167 . Then, the character code obtained in the character recognizing process is output.
- the character code of ‘ 2 ’ is finally output by performing the character re-recognizing process by the area emphasizing method using the upper half ‘m/2 ⁇ n’ area on the completed pattern of the character ‘ 2 ’, which indicates the displacement at its lower position and is input by the basic character recognizing unit 195 .
- Described below is an embodiment of the character string recognizing unit 15 shown in FIG. 7 .
- the character string recognizing unit 15 does not determine in a heuristic manner the threshold when the character is integrated on the parameter as a feature value used when a character is detected one by one from a character string extracted in the layout analysis in step S 2 shown in FIG. 8 .
- the threshold is determined as a statistically reasonable value.
- a parameter value and a statistic data are obtained on the successful or unsuccessful integration of a character corresponding to the parameter.
- Each parameter is not individually evaluated, but is counted as an element in a multiple-dimensional space, and the discriminate phase is obtained by the multivariate analysis to discriminate the two groups (cases) in the multiple-dimensional space for both cases where the integration is successfully performed and unsuccessfully performed.
- sample data comprising P feature values indicating the features of a pattern is divided into two groups, that is, a first group as successfully retrieved and a second group as unsuccessfully retrieved.
- the discriminant phase between the first and second groups is generated in the P-dimensional space.
- the discriminant phase can be obtained by, for example, the discriminant analysis method. That is, when the discriminant phase is formed by a linear discriminant function, the coefficient vector of the discriminant function is expressed as follows. ⁇ ⁇ 1 ( ⁇ 1 ⁇ 2 ) (3) where
- the discriminant function having the coefficient vector in equation (3) is generated in a way that an equal distance can be set from each center of gravity of the first and second groups.
- the coefficient vector of the discriminant function can also be computed based on the standard that the ratio of the inter-group variation between the first and second groups to the intra-group variation can be the largest possible.
- the process of detecting a character from a character string is performed separately by the statistic process of integrating character patterns by referring to the positions, sizes, arrays, etc. of the circumscribing rectangles of character patterns, and by the non-statistic process of observing the forms of character patterns to correctly process the superscript strokes, separate-stroke characters, etc.
- the detection parameter refers to the position and ratio of vertical to horizontal size of a rectangle circumscribing a pattern, ratio of character size to an average character size, width of overlap between patterns, density of character string, etc.
- step S 181 the rectangle circumscribing the link pattern is retrieved.
- step S 182 It is checked in step S 182 whether or not there is another circumscribing rectangle to the right of the circumscribing rectangle retrieved in step S 181 . If there is no circumscribing rectangle to the right of the circumscribing rectangle retrieved in step S 181 , then the circumscribing rectangle retrieved in step S 181 is removed from the objects of the statistic process.
- step S 182 If it is determined in step S 182 that there is another circumscribing rectangle to the right of the circumscribing rectangle retrieved in step S 181 , then control is passed to step S 184 .
- step S 183 the average character size of the rectangle circumscribing a character string is computed.
- the exact average character size cannot be immediately computed because each character has not been detected yet.
- a provisional average character size is computed by temporarily integrating the rectangle circumscribing a link pattern.
- An average character size is computed after the temporary integration.
- the average character size of a rectangle circumscribing a character string tan also be obtained by generating a frequency histogram for each size of a circumscribing rectangle.
- the pose and density of a pattern, the size of an integrated pattern obtained by integrating adjacent patterns, and the distance between the patterns are used as detection parameters.
- the following values are used as the detection parameters as shown in FIG. 58 .
- This separate-stroke-character process is described below by referring to the flowchart shown in FIG. 59 .
- This separate-stroke-character process is to detect a separate-stroke character in a link pattern formed by two or more separate strokes such as , , etc.
- step S 191 it is determined whether or not a right-lifted pattern exists in link patterns. If there is no right-lifted pattern, the separate-stroke-character process is not performed.
- step S 191 If a right-lifted pattern is detected in step S 191 , then control is passed to step S 192 and it is determined whether or not there is a left-lifted pattern adjacent to the right of the above described right-lifted pattern, that is, a pattern of, for example, , or a pattern adjacent to the right of the above described right-lifted pattern and intersecting another pattern (right angle line density) two times when it is searched for in the vertical direction, that is, a pattern of, for example, . Unless the pattern refers to a pattern in the form of or , the separate-stroke character process is not performed in this case.
- step S 192 If it is determined in step S 192 that the pattern refers to a pattern in the form of or control is passed to step S 194 .
- step S 193 an average character size of a character string in a circumscribing rectangle is computed in step S 193 .
- step S 194 the values of the parameters p through r shown in FIG. 54 are computed instep S 194 .
- a candidate for a pattern with superscript strokes is checked to use the size of the adjacent patterns integrated, the distance between these patterns, and the ratio of the size of the characters to the size of an average character as detection parameters.
- a pattern of a candidate for a superscript stroke is extracted. For example, when two adjacent link patterns are extracted by a link pattern extracting unit 1 , and when the ratio of the size of the integrated pattern of the two adjacent patterns to an average character size of the circumscribing rectangle of a character string is equal to or smaller than a predetermined threshold, that is 1 ⁇ 4, the pattern is extracted as a candidate for a pattern with superscript strokes.
- step S 202 It is checked in step S 202 whether or not there is an adjacent circumscribing rectangle to the left of the pattern which is a candidate for a character with superscript stokes. Unless there is an adjacent circumscribing rectangle to the left of the pattern which is a candidate for a character with superscript strokes, then the candidate for a character with superscript strokes is removed from the objects to be processed in the superscript-stroke process.
- step S 202 If it is determined in step S 202 that there is an adjacent circumscribing rectangle to the left of the pattern which is a candidate for a character with superscript strokes, then control is passed to step S 204 .
- step S 203 in addition to the above described step S 201 and S 202 , an average character size of a rectangle circumscribing a character string is computed.
- step S 204 the values of the parameters p through r shown in FIG. 60 are computed in step S 204 .
- a discriminant phase is set to compute the reliability in detecting a character from an unknown handwritten character string using learning data. If the number of parameters is n, then two groups are generated in the n-dimensional space to store in each group the characters detected successfully and unsuccessfully.
- FIG. 62 is a flowchart showing the method of computing data on successful and unsuccessful detection.
- step S 211 it is visually determined about the preliminarily collected learning data whether or not the object circumscribing rectangle and the adjacent circumscribing rectangle can be integrated into a single character in step S 211 . If the object circumscribing rectangle and the adjacent circumscribing rectangle can be integrated into a single character, then control is passed to step S 212 . If the object circumscribing rectangle and the adjacent circumscribing rectangle cannot be integrated into a single character, then control is passed to step S 213 .
- step S 212 when the object circumscribing rectangle and the adjacent circumscribing rectangle can be integrated into a single character, the values of the parameters of the object circumscribing rectangle and the adjacent circumscribing rectangle are recorded.
- the parameters of the object circumscribing rectangle and the adjacent circumscribing rectangle can be the parameters a through g shown in FIG. 52 in the statistic process, and can be the parameters p through r shown in FIGS. 58 and 60 in the non-statistic process.
- step S 213 when the object circumscribing rectangle and the adjacent circumscribing rectangle cannot be successfully integrated into a single character, the values of the parameters of the object circumscribing rectangle and the adjacent circumscribing rectangle are recorded.
- the values of the detection parameters in the statistic process and non-statistic process are computed about an unknown character string.
- a distance from a discriminant phase obtained from the learning data is computed on the point in a multiple-dimensional space determined by the value of the parameter. The obtained distance is quantified as the detection reliability.
- H indicates the discriminant phase for use in discriminating the two groups, that is, successfully detected characters and unsuccessfully detected characters
- n indicates the unit normal vector of the discriminant phase H as shown in FIG. 63 .
- Whether the distance h from the discriminant phase H is positive or negative determines to which group, that is, successfully detected group or unsuccessfully detected group, the value of the parameter belongs, and determines to what extent the value of the parameter is away from the discriminant phase H.
- a successfully detected histogram 241 and an unsuccessfully detected histogram 242 are obtained for the entire parameters of the learning data in the multiple-dimensional space based on the distance h from the discriminant phase H.
- the histogram distributions 241 and 242 are normal distributions, the histogram distributions 241 and 242 are approximated at a normal distribution. In these normal distributions, partially overlapping areas normally exist.
- the present invention it is determined whether or not the patterns are to be integrated in consideration of the reliability of the detection of the adjacent pattern having a detection parameter positioned at the overlapping area.
- FIG. 65 is a flowchart showing an example of the method of computing the detection reliability.
- the distance h from the discriminant phase H to a point in a multiple-dimensional space determined by a plurality of parameters is computed by the above described equation (10) in step S 221 .
- step S 222 the histogram distribution of a plurality of parameter values obtained from the learning data is approximated using the normal distribution. That is, as shown in FIG. 66 , the histogram distribution of successfully detected patterns is approximated using a successfully detected pattern normal distribution 251 , and the histogram distribution of unsuccessfully detected patterns is approximated using an unsuccessfully detected pattern normal distribution 252 .
- step S 223 the overlap areas of the two groups is computed.
- the overlap area between the successfully detected pattern normal distribution 251 and the unsuccessfully detected pattern normal distribution 252 is computed as a 2-group overlap area 254 as shown in FIG. 62 .
- an area 253 other than the 2-group overlap area 254 in the successfully detected pattern normal distribution 251 is set as a successfully detected area.
- an area 255 other than the 2-group overlap area 254 in the unsuccessfully detected pattern normal distribution 252 is set as an unsuccessfully detected area.
- the position of the value of the parameter input for an unknown character in the histogram distribution is determined in step S 224 .
- step S 225 if the value of the parameter input for the unknown character is included in the 2-group overlap area 254 as a determination result of the position of the value of the parameter input for the unknown character in the histogram distribution, then control is passed to step S 226 . Then, the detection reliability is computed based on the value of the parameter input for the unknown character in the 2-group overlap area 254 .
- step S 225 If it is determined in step S 225 that the value of the parameter input for the unknown character is not included in the 2-group overlap area 254 , then control is passed to step S 226 , and it is determined whether or not the value of the parameter input for the unknown character is included in the successfully detected area 253 .
- control is passed to step S 228 , and the detection reliability is set to ‘1’. If it is determined that the value of the parameter input for the unknown character is not included in the successfully detected area 253 , then control is passed to step S 229 , and the detection reliability is set to ‘0’.
- the detection reliability is computed based on the distance from the discriminant phase to the value of the parameter input for the unknown character. If the distance from the discriminant phase to the value of the parameter input for the unknown character is included in the successfully detected area 253 , then the detection reliability is set to ‘1’. If the distance from the discriminant phase to the value of the parameter input for the unknown character is included in the unsuccessfully detected area 255 , then the detection reliability is set to ‘0’.
- FIG. 67 is a flowchart showing an example of computing the 2-group overlap area.
- an average value m and a distribution value v of a histogram 261 is computed in step S 231 about the histogram distribution of successfully detected patterns and unsuccessfully detected patterns obtained from the learning data.
- step S 232 The sum d of squares error between a normal distribution curve 262 and the histogram 261 about the histogram 261 is computed in step S 232 about the histogram distribution of successfully detected patterns and unsuccessfully detected patterns.
- step S 235 the area from a right end 267 of a normal distribution curve 263 to a left end 266 of a normal distribution curve 264 is set as a 2-group overlap area 265 .
- the recognizing process is not performed on a candidate for a detected character having high detection reliability, but is performed on a candidate for a detected character having low detection reliability.
- a character to be detected is selected in consideration of the detection reliability as well as the recognition reliability.
- a candidate for a character which partially appears a character, but entirely appears a wrong character string can be removed from characters to be detected.
- the entire detection reliability R is expressed as follows.
- ⁇ 1 indicates the detection reliability of adjacent patterns or detection determined portion
- ⁇ i indicates the recognition reliability
- j indicates a weight coefficient
- a character having a larger entire reliability R is selected as the final character to be detected from the candidates for a plurality of characters to be detected.
- FIG. 68 shows the case where each character is detected from a character string ‘ ’.
- the discriminant phase for the statistic and non-statistic processes and the normal distribution curve of a histogram value are individually obtained using learning data.
- parameters c, e, and f shown, in FIG. 55 are used as the parameters for use in determining successful or unsuccessful detection of a character string.
- the average value m of the histogram distribution indicating a successful detection of learning data shown in FIG. 67 is 128.942.
- the standard deviation is 34.77.
- the adaptability T is 0.12 according to equation (11). Assuming that the constant of proportionality k is 2, the distance from the center to the end of the distribution is 77.8 according to equation (12).
- the average value m of the histogram distribution indicating an unsuccessful detection of learning data shown in FIG. 67 is 71.129.
- the standard deviation is 36.26.
- the adaptability T is 0.35 according to equation (11). Assuming that the constant of proportionality k is 2, the distance from the center to the end of the distribution is 92.2 according to equation (10).
- the input pattern of an unknown character is read from an input image in step S 241 .
- step S 242 a link pattern is extracted using labels, and the label numbers ⁇ 1 > through ⁇ 6 > are assigned as shown in FIG. 68 to each of the extracted link patterns.
- step S 245 the detection reliability is quantified based on the statistic process in step S 243 and the non-statistic process in step S 244 .
- the detection reliability obtained when adjacent link patterns are integrated is computed based on the distance h from the discriminant phase to the point in the 3-dimensional space having the parameter values c, e, and f.
- the detection reliability obtained when the pattern assigned the label number ⁇ 1 > is integrated into the pattern assigned the label number ⁇ 2 > is 80.
- the detection reliability obtained when the pattern assigned the label number ⁇ 2 > is integrated into the pattern assigned the label number ⁇ 3 > is 12.
- the detection reliability obtained when the pattern assigned the label number ⁇ 3 > is integrated into the pattern;assigned the label number ⁇ 4 > is 28.
- the detection reliability obtained when the pattern assigned the label number ⁇ 4 > is integrated into the pattern assigned the label number ⁇ 5 > is 92.
- the detection reliability obtained when the pattern assigned the label number ⁇ 5 > is integrated into the pattern assigned the label number ⁇ 6 > is 5.
- step S 244 the detection reliability of the pattern ‘ ’ with the superscript strokes is computed based on the distance h from the discriminant phase to the point in the 3-dimensional space having the parameter values p through r in FIG. 60 .
- the detection reliability obtained when the pattern assigned the label number ⁇ 1 > is integrated into the superscript-character-pattern of a detection determined portion 271 comprising the patterns assigned the label numbers ⁇ 2 > and ⁇ 3 > is 85.
- FIG. 65 shows the method of computing the detection reliability in the non-statistic process in step S 244 .
- a pattern 272 is extracted as a candidate for superscript strokes.
- the pattern can be a candidate for superscript strokes when there are two adjacent link patterns, and when the ratio of the size of the integrated patterns to the average character size of the rectangle circumscribing the character string is lower than a predetermined threshold.
- step 252 it is determined whether or not there is a circumscribing rectangle 281 adjacent to the left of the pattern 272 which is a candidate for superscript strokes. If it is determined that there is a circumscribing rectangle 281 adjacent to the left of the pattern 272 which is a candidate for superscript strokes, then control is passed to step S 253 and the values of the parameters p through r shown in FIG. 50 are output.
- step S 254 the distance from a discriminant phase 293 to a point in the 3-dimensional space having the values of the parameters p through r is computed.
- the discriminant phase 293 is computed based on the learning pattern.
- the discriminant phase 293 can be obtained by equation (3) based on, for example, a histogram distribution 292 indicating successful detection of a character string of a learning pattern and a histogram distribution 291 indicating unsuccessful detection of the learning data.
- the distance h from the discriminant phase 293 is computed by substituting the values obtained by equations (16) through (18) for equation (19) as follows.
- the average value m of the histogram distribution 292 indicating a successful detection of learning data is 38.
- the standard deviation is 25.
- the adaptability T is 0.2 according to equation (11).
- the average value m of the histogram distribution 291 indicating an unsuccessful detection of learning data shown is ⁇ 34.
- the standard deviation is 28.
- the adaptability T is 0.3 according to equation (11).
- a 2-group overlap area 294 is positioned at the distance of ⁇ 22 through 38.8 from the discriminant phase.
- the detection reliability a is computed in step S 255 .
- detection determined portion 271 is generated by integrating the patterns assigned the label numbers ⁇ 2 > and ⁇ 3 >.
- step S 246 shown in FIG. 68 the reliability of the statistic and non-statistic processes is synthesized.
- the detection determined portion if any, is prioritized.
- the reliability of the detection determined portion 271 is synthesized by priority.
- the detection reliability obtained by integrating the pattern assigned the label number ⁇ 1 > into the pattern of the detection determined portion 271 is 85.
- the detection reliability obtained by integrating the pattern of the detection determined portion 271 into the pattern assigned the label number ⁇ 4 > is 30.
- the detection reliability obtained by integrating the pattern assigned the label number ⁇ 4 > into the pattern assigned the label number ⁇ 5 > is 92.
- the detection reliability obtained by integrating the pattern assigned the label number ⁇ 5 > into the pattern assigned the label number ⁇ 6 > is 5.
- Patterns are integrated if the detection reliability is higher than a predetermined threshold (for example, 90) or if the detection reliability is higher than a predetermined threshold (for example, 70) and the ratio of the reliability to the detection reliability of the adjacent detected pattern is higher than a predetermined value (for example, 5).
- a predetermined threshold for example, 90
- a predetermined threshold for example, 70
- a predetermined value for example, 5
- Patterns are not integrated if the detection reliability is lower than a predetermined threshold (for example, 8).
- the detection reliability obtained by integrating the pattern of the detection determined portion 271 into the pattern assigned the label number ⁇ 4 > is 30, and therefore, the pattern of the detection determined portion 271 is not integrated into the pattern assigned the label number ⁇ 4 >.
- the pattern assigned the label number ⁇ 4 > is integrated into the pattern assigned the label number ⁇ 5 >. Since the detection reliability obtained by integrating the pattern assigned the label number ⁇ 5 > into the pattern assigned the label number ⁇ 6 > is 5, the pattern assigned the label number ⁇ 5 > is not integrated into the pattern assigned the label number ⁇ 6 >.
- a circumscribing rectangle 275 corresponding to a detection determined portion 273 obtained by integrating the pattern assigned the label number ⁇ 4 > into the pattern assigned the label number ⁇ 5 > and a circumscribing rectangle 276 corresponding to the pattern assigned the label number ⁇ 6 > are generated.
- the detection reliability is 60 in the example shown in FIG. 68 .
- a detection candidate 1 and a detection candidate 2 are extracted when the patterns are completely integrated based on the detection reliability. Then, a recognizing process is performed on each character of the detection candidates 1 and 2 .
- the detection reliability ⁇ and ⁇ of a character in the detection candidates 1 and 2 is obtained for each character, and the sum of the detection reliability ⁇ and ⁇ is defined as the entire reliability R.
- the recognition reliability ⁇ obtained when a character recognizing process is performed on the pattern in a circumscribing rectangle 278 is 80
- the recognition reliability ⁇ obtained when a character recognizing process is performed on the pattern in a circumscribing rectangle 275 is 90
- the recognition reliability ⁇ obtained when a character recognizing process is performed on the pattern in a circumscribing rectangle 276 is 85.
- the entire reliability R is 345 by equation (13) with the weight coefficient j assumed to be 1.
- the recognition reliability ⁇ obtained when a character recognizing process is performed on the pattern in a circumscribing rectangle 281 is 83
- the recognition reliability ⁇ obtained when a character recognizing process is performed on the pattern in a circumscribing rectangle 282 is 55
- the recognition reliability ⁇ obtained when a character recognizing process is performed on the pattern in a circumscribing rectangle 276 is 85.
- the entire reliability R is 283.
- step S 248 the detection candidate 1 or the detection candidate 2 , whichever is larger in entire reliability R, is selected as a candidate for a successfully detected character. As a result, each of the characters , , and can be correctly detected in the character string .
- FIG. 70 is a block diagram showing an embodiment of the configuration of the obscure character recognizing unit 19 .
- a feature extracting unit 301 extracts a feature of a character from an obscure character and represent the extracted feature by a feature vector.
- An obscure-character dictionary 302 stores a feature vector of each category of obscure characters.
- a collating unit 303 collates the feature vector of a character pattern extracted by the feature extracting unit 301 with the feature vector of each category stored in the obscure-character dictionary 302 , and computes the distance D ij (i indicates a feature vector of an unknown character, and j indicates a feature vector of a category in the obscure-character dictionary 302 ) between the feature vectors in a feature space.
- the category j indicating the shortest distance D ij between the feature vectors is recognized as an unknown character i.
- the distance D ij between the feature vectors in the feature space can be computed using, for example, an Euclidean distance ⁇ (i ⁇ j) 2 , city block distance ⁇
- a table 1 relating to the first category j 1 , the second category j 2 , the distance between categories (D ij2 ⁇ D ij1 ), and the reliability is preliminarily generated.
- a table 2 relating to the distance D ij1 from the first category, the first category j 1 , and the reliability is preliminarily generated. The data having lower reliability in the tables 1 and 2 is stored in the intermediate process result table.
- the deformed character recognizing unit 21 shown in FIG. 7 can be designed similarly to the obscure character recognizing unit 19 except that the deformed character recognizing unit 21 uses a deformed-character dictionary storing feature vectors in each category of deformed characters.
- the deletion line recognizing unit 26 generates, for example, a histogram containing sums of the numbers of picture elements in the horizontal direction for the candidate for a corrected character extracted by the correction analysis in step S 4 in FIG. 8 , and removes the horizontal lines in the area by recognizing that the horizontal lines exist in the areas where the histogram value exceeds a predetermined value.
- a character is recognized by completing an obscure portion with the horizontal lines removed and then collating the completed pattern with the dictionary.
- a pattern is recognized as a character, then a candidate for a corrected character is regarded as a character with deletion lines. If a pattern is rejected, then a candidate for a corrected character is regarded as a normal character.
- a character ‘ 5 ’ which is a candidate for a corrected character, is input as being corrected with double horizontal lines.
- the input pattern is recognized as a corrected character after detecting double horizontal lines which indicate the horizontal histogram value equal to or larger than the threshold N and recognizing the completed pattern as the category of ‘ 5 ’ after the double horizontal lines are removed.
- a character ‘ 5 ’ is input as a candidate for a corrected character, and horizontal line indicating the horizontal histogram value equal to or larger than the threshold N is detected. If the horizontal line is removed from the character ‘ 5 ’ and the pattern is rejected, then input pattern is not recognized as a corrected character.
- the unique-character analyzing unit 23 clusters handwritten characters recognized as belonging to the same category into a predetermined number of clusters. When clusters belonging to different categories indicates a short distance from each other, the character category of the cluster containing a smaller number of elements is amended to the character category of the cluster containing a larger number of elements, thereby a handwritten character which has been misread for a wrong character category can be correctly read.
- FIG. 72 shows the clustering process using a feature vector of a handwritten character recognized as belonging to the character category of ‘ 4 ’.
- FIG. 72 shows the handwritten characters which has been determined as belonging to the recognition result category of ‘ 4 ’ because the pattern indicates a shorter distance from the feature vector of the character category of ‘ 4 ’ stored in the recognizing dictionary.
- the handwritten character ‘ 2 ’ is mis-recognized as belonging to the recognition result category of ‘ 4 ’.
- the handwritten character determined to belong the character category of ‘ 4 ’ is individually regarded as a cluster.
- the distance of the feature vectors between the handwritten characters regarded as clusters is computed, and the characters indicating a short distance from each other in feature vector are integrated into one cluster. As a result, the number of the clusters is decreased by 1 from 11 to 10 as in the example shown in FIG. 72 .
- the number of clusters can be reduced by computing the distance of feature vectors between clusters and integrating the closest feature vectors, resulting in one cluster in the eleventh clustering process.
- the center-of-gravity method is used when clusters contain plural elements.
- the representative vector x m of the lusters is represented by an average of the feature vector x i of the elements of the clusters as follows.
- Clusters containing plural elements are compared with each other by computing the city block distance between the representative vectors x m .
- the clustering aborting condition can be satisfied when
- FIG. 73 is a flowchart showing the clustering process
- step S 261 only the feature vector of the handwritten character recognized as belonging to a specific character category is extracted in step S 261 .
- Each of the extracted feature vectors of the handwritten characters is regarded as one cluster.
- step S 262 the clustering aborting condition is set to abort the clustering processes.
- steps S 263 two clusters closest to each other in all clusters are selected about a specific character.
- step S 264 It is determined in step S 264 whether or not the cluster aborting conditions set in step S 262 are satisfied. If the cluster aborting conditions set in step S 262 are not satisfied, then control is passed to step S 265 , the two clusters selected in step S 263 are integrated, control is returned to step S 263 , and clusters are repeatedly integrated.
- step S 264 If it is determined in step S 264 that the clustering aborting conditions are satisfied after repeating the cluster integrating processes, then control is passed to step S 266 , and it is determined whether or not the clustering processes have been performed on all character categories. If all clustering processes have not been performed on all character categories, control is returned to step 261 , and the clustering process is performed on the unprocessed character categories.
- step S 266 If it is determined in step S 266 that the clustering processes have been performed on all character categories, then control is passed to step S 267 , and the clustering results are stored in the memory.
- FIG. 74 shows the process of correctly reading a handwritten character ‘ 2 ’, which is mis-recognized for belonging to the character category of ‘ 4 ’, as the character category ‘ 2 ’.
- FIG. 74 shows the handwritten character determined to belong to the recognition result category of ‘ 2 ’, and the handwritten character determined to belong to the recognition result category of ‘ 4 ’.
- the handwritten character ‘ 3 ’ is mis-recognized for belonging to the recognition result category of ‘ 2 ’, and the handwritten character ‘ 2 ’ is mis-recognized for belonging to the recognition result category of ‘ 4 ’.
- the handwritten character ‘ 4 ’ is rejected for not belonging to any recognition result category.
- Clusters g, h, and i are generated for the three rejected handwritten characters ‘ 4 ’.
- clusters a, b, and c belonging to the recognition result category of ‘ 2 ’ or clusters d, e, and f belonging to the recognition result category of ‘ 4 ’ whichever contains a smaller number of characters is extracted as a candidate for a mis-read cluster.
- Cluster b is extracted as the cluster closest to misread candidate cluster a. It is determined whether or not the distance between mis-read candidate cluster a and cluster b is shorter than a predetermined value. Since the distance between mis-read candidate cluster a and cluster b is not shorter than a predetermined value, mis-read candidate cluster a is rejected.
- Cluster b is extracted as the cluster closest to mis-read candidate cluster d. It is determined whether or not the distance between mis-read candidate cluster d and cluster b is shorter than a predetermined value. Since the distance between mis-read candidate cluster d and cluster b is shorter than a predetermined value, mis-read candidate cluster d is integrated into cluster b to generate cluster j. It is determined that cluster j belongs to the recognition result category of ‘ 2 ’ to which cluster b containing the larger number of elements belongs. Thus, the handwritten character ‘ 2 ’ which has been mis-read for ‘ 4 ’ and determined to belong to mis-read candidate cluster d can be correctly read.
- Cluster a is extracted as the cluster closest to cluster g. It is determined whether or not the distance between cluster g and cluster a is shorter than a predetermined value. Since the distance between cluster g and cluster a is shorter than a predetermined value, cluster g is integrated into cluster e.
- Cluster e is extracted as the cluster closest to cluster h. It is determined whether or not the distance between cluster h and cluster e is shorter than a predetermined value, cluster h is integrated into cluster e. After integrating clusters g and h into cluster e, cluster k belongs to the recognition result category of ‘ 4 ’ to which cluster e containing the larger number of elements belongs. Therefore, the handwritten character ‘ 4 ’ which has been rejected as being unrecognizable can be correctly read.
- Cluster e is extracted as the cluster closest to cluster i. It is determined whether or not the distance between cluster i and cluster e is shorter than a predetermined value. Since the distance between cluster i and cluster e is not shorter than a predetermined value, cluster i is not integrated into cluster e.
- FIG. 75 is a flowchart showing the character category recognition result amending process.
- step S 271 the data of the clustering result obtained in the clustering process shown in FIG. 73 is read from the memory in step S 271 .
- step S 272 the distance between clusters is computed and compared for all clusters in all categories obtained in the clustering process shown in FIG. 73 .
- step S 273 it is determined whether or not the distance between clusters is shorter than a predetermined threshold. If the distance between any clusters is shorter than a predetermined threshold, then control is passed to step S 274 , and the clusters are integrated. If the distance between any clusters is not shorter than a predetermined threshold, the clusters are rejected.
- condition of integrating clusters can be represented as follows with the constant set to 1.5.
- xam indicates the representative vector of cluster A
- xbm indicates the representative vector of cluster B
- step S 275 the character categories are determined in all clusters integrated in step S 274 .
- step S 276 it is determined whether or not the character categories of the integrated clusters are different from each other. If the character categories of the integrated clusters are different from each other, then control is passed to step S 277 , and the character category containing a smaller number of elements is amended to the character category of the cluster containing a larger number of elements. Then, control is passed to step S 278 . If the character categories of the clusters match, then step 5277 is skipped and control is passed to step S 278 .
- step S 278 a character category is output for a character in the cluster.
- FIG. 76 shows an example of the form input to the pattern recognizing apparatus according to an embodiment of the present invention.
- the form shown in FIG. 76 contains a free-pitch 167 box having the box number 1 ; a one-character box having the box number 2 , 3 , or 4 ; a block character box having the box number 5 ; and an irregular table having the box number 6 .
- the free-pitch box having the box number 1 contains the box-touching character ‘ 5 ’ corrected with double horizon lines; the box-touching characters ‘ 3 ’ and ‘ 2 ’; the box-touching obscure character ‘ 7 ’; the unique characters ‘ 4 ’ and ‘ 6 ’; and the partly-outside-box unique character ‘ 4 ’.
- the one-character box having the box number 2 contains ‘ 5 ’.
- the one-character box having the box number 3 contains ‘ 3 ’.
- the one-character box having the box number 4 contains the partly-outside-box characters ‘’ corrected with double horizontal lines.
- the character box having the box number 5 - 1 contains the unique characters ‘’ corrected with double horizontal lines
- the character box having the box number 5 - 2 contains the box-touching character ‘
- the character box having the box number 5 - 3 contains the unique characters ‘ 4 ’.
- the character box having the box number 6 - 1 - 1 contains the partly-outside-box characters ‘ 3 ’, ‘ 2 ’, and ‘ 1 ’
- the character box having the box number 6 - 1 - 2 contains the character 6 , 3 , and 8
- the character boxes having the box numbers 6 - 1 - 3 , 6 - 14 - 1 , 6 - 14 - 2 , 6 - 14 - 3 , 6 - 2 - 1 , 6 - 2 - 2 , and 6 - 2 - 3 are kept blank.
- the entire irregular table having the character box number 6 is corrected with the mark ‘X’.
- the environment recognizing system 11 shown in FIG. 7 performs the process shown in FIGS. 9 through 12 , thereby extracting the state of an input image from the form shown in FIG. 76 .
- the free-pitch character box having the box number 1 , one-character boxes having the box numbers 2 , 3 , and 4 , the block character box having the box number 5 , and the irregular table having the box number 6 are extracted from the form shown in FIG. 76 by performing the layout analysis shown in FIG. 10 .
- eight patterns are extracted as candidates for characters from the free-pitch box having the box number 1 .
- a pattern is extracted as a candidate for a character from each of the one character boxes having the box numbers 2 , 3 , and 47
- Three patterns are extracted as candidates for characters from the block character box having the box number 5 .
- Three patterns are extracted as candidates for characters from the character box having the box number 6 - 1 - 1 .
- Three patterns are extracted as candidates for characters from the character box having the box number 6 - 1 - 2 .
- No patterns are extracted as candidates for characters from the character boxes having the box numbers 6 - 1 - 3 , 6 - 1 - 4 - 1 , 6 - 1 - 4 - 2 , 6 - 1 - 4 - 3 , 6 - 2 - 1 , 6 - 2 - 2 , and 6 - 2 - 3 .
- the text extracting method shown in FIGS. 18 and 19 is used.
- the ruled line extracting method shown in FIGS. 20 and 26 is used.
- the box extracting method shown in FIGS. 27 and 28 is used.
- the first second, fifth, and eighth patterns extracted from the free-pitch box having the box number 1 are regarded as candidates for box-touching characters.
- the patter extracted from the one character box having the box number 4 , the pattern extracted from the character box having the box number 5 - 2 , the first pattern extracted from the character box having the box number 6 - 1 - 1 are also regarded as candidates for box-touching characters.
- FIG. 11 Through the quality analysis shown in FIG. 11 , obscure, deformed, and high-quality characters are detected in the form shown in FIG. 76 .
- the quality of images is normal, but obscure, deformed, or high-quality characters are not detected.
- a candidate for a corrected character is extracted from the list shown in FIG. 76 through the correction analysis shown in FIG. 12 .
- the first pattern extracted from the free-pitch box having the box number 1 the patterns extracted from the one character boxes having the box numbers 2 and 4 , the pattern extracted from the character box having the box number 5 - 1 , and the pattern extracted from the irregular table having the box number 6 are regarded as candidates for corrected characters.
- the feature extracting method shown in FIG. 34 is used, for example.
- the environment recognizing system 11 generates an intermediate process result table containing the state extracted from the form in the processes shown in FIGS. 9 through 12 for each of the candidates for the characters extracted from the input image.
- FIG. 77 shows the intermediate process result table containing the state extracted from the form in the processes in FIGS. 9 through 12 .
- the column of the character box having the box number 1 contains ‘free-pitch’ for the type of box and ‘ 8 ’ for the number of characters.
- the column for the first pattern in the box having the box number 1 contains ‘YES’ indicating the existence of a box-touching character, ‘YES 2 ’ indicating the existence of deletion lines, and ‘NORMAL’ indicating the quality.
- the column for the second pattern in the box having the box number 1 contains ‘YES’ indicating the existence of a box-touching character, ‘NO’ indicating the non-existence of deletion lines, and ‘NORMAL’ indicating the quality.
- the column for the eighth pattern in the box having the box number 1 contains ‘YES’ indicating the existence of a box-touching character, ‘NO’ indicating the non-existence of deletion lines, and ‘NORMAL’ indicating the quality.
- ‘YES 1 ’ indicating the existence of the deletion lines refers to the existence of a candidate for deletion lines for a plurality of characters. ‘YES 2 ’ indicating the existence of the deletion lines refers to the existence of a candidate for deletion lines for a single character.
- the column of the character box having the box number 2 contains ‘one-character’ for the type of box, ‘NO’ indicating the non-existence of a box-touching character, ‘YES 2 ’ indicating the existence of deletion lines, ‘NORMAL’ indicating the quality, and ‘ 1 ’ indicating the number of characters.
- the column of the character box having the box number 3 contains ‘one-character’ for the type of box, ‘NO’ indicating the non-existence of a box-touching character, ‘NO’ indicating the non-existence of deletion lines, ‘NORMAL’ indicating the quality, and ‘ 1 ’ indicating the number of characters.
- the column of the character box having the box number 4 contains ‘one-character’ for the type of box, ‘YES’ indicating the existence of a box-touching character, ‘YES 2 ’ indicating the existence of deletion lines, ‘NORMAL’ indicating the quality, and ‘ 1 ’ indicating the number of characters.
- the column of the character box having the box number 5 contains ‘INDIVIDUAL CHARACTER BOXES’ indicating the type of box, ‘ 3 ’ indicating the number of characters.
- the column of the character box having the box number 5 - 1 contains ‘NO’ indicating the non-existence of a box-touching character, ‘YES 2 ’ indicating the existence of deletion lines, ‘NORMAL’ indicating the quality, and ‘ 1 ’ indicating the number of characters.
- the column of the character box having the box number 5 - 2 contains ‘YES’ indicating the existence of a box-touching character, ‘NO’ indicating the non-existence of deletion lines, ‘NORMAL’ indicating the quality, and ‘ 1 ’ indicating the number of characters.
- the column of the character box having the box number 6 contains ‘TABLE’ indicating the type of character box.
- the column of the character box having the box number 6 - 1 - 1 contains ‘FREE-PITCH’ indicating the type of character box, ‘YES’ indicating the existence of a box-touching character, ‘YES 1 ’ indicating the existence of deletion lines, and ‘NORMAL’ indicating the quality.
- the column of the character box having the box number 6 - 2 - 2 contains ‘FREE-PITCH’ indicating the type of character box, ‘NO’ indicating the non-existence of a box-touching character, ‘YES 1 ’ indicating the existence of deletion lines, and ‘NORMAL’ indicating the quality.
- the environment recognizing system 11 performs the process shown in FIG. 13 based on the state extracted from the form in the process shown in FIGS. 9 through 12 .
- the environment recognizing system 11 determines which process is to be called by referring to the process order control rules, the process performed by the basic character recognizing unit 17 , character string recognizing unit 15 , box-touching character recognizing unit 13 , obscure character recognizing unit 19 , or deformed character recognizing unit 21 of the character recognizing unit 12 shown in FIG. 7 , or the process performed by the deletion line recognizing unit 26 or noise recognizing unit 28 of the non-character recognizing unit 25 .
- the determined process is entered in the column ‘CALLING PROCESS’ of the intermediate process result table shown in FIG. 77 .
- the determined order is entered in the column ‘PROCESS ORDER’ of the intermediate process result table shown in FIG. 77 .
- the process order control rule can be specified as follows.
- FIG. 78 shows an example of the process order table.
- the process order table stores the following procedures.
- FIG. 79 shows an example of entering in the column ‘CALLING PROCESS’ the process to be called based on the state of an input image entered in the intermediate process result table shown in FIG. 77 , and of entering in the column ‘PROCESS ORDER’ the order of performing the process entered in the column ‘CALLING PROCESS’.
- the column of the character box having the box number 1 contains ‘FREE-PITCH’ for the type of box.
- the column for the first pattern in the box having the box number 1 contains ‘YES’ indicating the existence of a box-touching character, and ‘YES 2 ’ indicating the existence of deletion lines.
- the column for the eighth pattern in the box having the box number 1 contains ‘YES’ indicating the existence of a box-touching character, ‘NO’ indicating the non-existence of deletion lines, and ‘NORMAL’ as ‘QUALITY’. Therefore, according to (A 1 ) in the process order control rule, ‘BLACK-CHARACTER-BOX/FREE-PITCH’ is entered in the column ‘CALLING PROCESS’, (B 2 ) of the process order table is referred to according to (A 3 ) of the process order control rule, and ‘BLACK-CHARACTER-BOX ⁇ FREE-PITCH’ is entered in the column ‘PROCESS ORDER’.
- the column for the second pattern in the box having the box number 1 contains ‘YES’ indicating the existence of a box-touching character, ‘NO’ indicating the non-existence of deletion lines, and ‘NORMAL’ as ‘QUALITY’. Therefore, according to (A 1 ) in the process order control rule, ‘BLACK-CHARACTER-BOX/FREE-PITCH’ is entered in the column ‘CALLING PROCESS’, (B 2 ) of the process order table is referred to according to (A 3 ) of the process order control rule, and ‘BLACK-CHARACTER-BOX ⁇ FREE-PITCH’ is entered in the column ‘PROCESS ORDER’.
- the column of the character box having the box number 2 contains ‘ONE CHARACTER’ for the type of box, ‘NO’ indicating the non-existence of a box-touching character, ‘YES 2 ’ indicating the existence of deletion lines, and ‘NORMAL’ indicating the quality. Therefore, according to (A 1 ) in the process order control rule, ‘DELETION LINE (YES 2 )’ is entered in the column ‘CALLING PROCESS’, and ‘ONE-CHARACTER DELETION LINE’ is entered in the column ‘PROCESS ORDER’ according to (A 1 ) in the process order control rule.
- the column of the character box having the box number 3 contains ‘ONE CHARACTER’ for the type of box, ‘NO’ indicating the non-existence of a box-touching character, ‘NO’ indicating the non-existence of deletion lines, and ‘NORMAL’ indicating the quality. Therefore, according to (A 2 ) in the process order control rule, ‘BASIC’ is entered in the column ‘CALLING PROCESS’, and ‘BASIC’ is entered in the column ‘PROCESS ORDER’ according to (A 1 ) in the process order control rule.
- the column of the character box having the box number 4 contains ‘ONE CHARACTER’ for the type of box, ‘YES’ indicating the existence of a box-touching character, ‘YES 2 ’ indicating the existence of deletion lines, and ‘NORMAL’ indicating the quality. Therefore, according to (A 1 ) in the process order control rule, ‘BLACK-CHARACTER-BOX/DELETION LINE (YES 2 )’ is entered in the column ‘CALLING PROCESS’, (B 3 ) of the process order table is referred to and ‘BLACK-CHARACTER-BOX ⁇ ONE-CHARACTER DELETION LINE’ is entered in the column ‘PROCESS ORDER’ according to (A 3 ) in the process order control rule.
- the column of the character box having the box number 5 contains ‘INDIVIDUAL CHARACTER BOXES’ for the type of box.
- the column of the character box having the box number 5 - 1 contains ‘NO’ indicating the nonexistence of a box-touching character, ‘YES 2 ’ indicating the existence of deletion lines, and ‘NORMAL’ indicating the quality. Therefore, according to (A 1 ) in the process order control rule, ‘DELETION LINE (YES 2 )’ is entered in the column ‘CALLING PROCESS’, and ‘ONE-CHARACTER DELETION LINE’ is entered in the column ‘PROCESS ORDER’ according to (A 1 ) of the process order control rule.
- the column of the character box having the box number 5 - 2 contains ‘YES’ indicating the existence of a box-touching character, ‘NO’ indicating the nonexistence of deletion lines, and ‘NORMAL’ indicating the quality. Therefore, according to (AI) in the process order control rule, ‘BLACK-CHARACTER-BOX’ is entered in the column ‘CALLING PROCESS’, and ‘BLACK-CHARACTER-BOX’ is entered in the column ‘PROCESS ORDER’ according to (AZ) of the process order control rule.
- the column of the character box having the box number 5 - 3 contains’ NO’ indicating the non-existence of a box-touching character, ‘NO’ indicating the nonexistence of deletion lines, and ‘NORMAL’ indicating the quality. Therefore, according to (A 2 ) in the process order control rule, ‘BASIC’ is entered in the column ‘CALLING PROCESS’, and ‘BASIC’ is entered in the column ‘PROCESS ORDER’ according to (A 1 ) in the process order control rule.
- the column of the character box having the box number 6 contains ‘TABLE’ indicating the type of character box.
- the column of the character box having the box number 6 - 1 - 1 contains ‘FREE-PITCH’ indicating the type of character box, ‘YES’ indicating the existence of a box-touching character, ‘YES 1 ’ to (A 1 ) in the process order control rule, ‘BLACK-CHARACTER-BOX/FREE-PITCH/DELETION LINE (YES 1 )’ is entered in the column ‘CALLING PROCESS’, (B 5 ) of the process order table is referred to according to (A 3 ) of the process order control rule, and ‘PLURAL-CHARACTER DELETION LINE ⁇ BLACK-CHARACTER-BOX ⁇ FREE-PITCH’ is entered in the column ‘PROCESS ORDER’.
- the column of the character box having the box number 6 - 2 - 2 contains ‘FREE-PITCH’ indicating the type of character box, ‘NO’ indicating the non-existence of a box-touching character, ‘YES 1 ’ indicating the existence of deletion lines, and ‘NORMAL’ indicating the quality. Therefore, according to (A 1 ) in the process order control rule, ‘FREE-PITCH/DELETION LINE (YES 1 )’ is entered in the column ‘CALLING PROCESS’, (B 6 ) of the process order table is referred to according to (A 3 ) of the process order control rule, and ‘PLURAL-CHARACTER DELETION LINE ⁇ FREE-PITCH’ is entered in the column ‘PROCESS ORDER’.
- the first recognizing process shown in FIG. 78 is performed by referring to the process execution rule based on the intermediate process result table shown in FIG. 79 with the data entered in the columns ‘CALLING PROCESS’ and ‘PROCESS ORDER’.
- the completed recognizing process is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table, and the reliability obtained in the recognizing process is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- the column ‘PROCESS ORDER’ on the intermediate process result table is updated according to (B 7 ) through (B 9 ) on the process order table shown in FIG. 78 . If the next process is specified according to the process execution rule, the process is entered in the column ‘PROCESS INSTRUCTION’ on the intermediate process result table.
- the process execution rule can be described as follows.
- FIG. 80 shows an example of performing a recognizing process by referring to the process execution rule based on the intermediate process result table shown in FIG. 79 ; entering the reliability obtained in the recognizing process in the column ‘RELIABILITY’ on the intermediate process result table; updating the column ‘PROCESS ORDER’ on the intermediate process result table based on the process execution rule; and entering data in the column ‘PROCESS INSTRUCTION’ on the intermediate process result table.
- the box-touching character recognizing unit 13 recognizes a character by completing or re-completing the character for the pattern from which its character box is removed as shown in FIGS. 43 and 44 . If a pattern cannot be recognized at acceptable reliability even after the character completing or re-completing process, then the knowledge table 14 is referred to and a character re-completing process is performed on the learning character shown in FIGS. 46 through 55 to successfully recognize a box-touching character.
- the recognition reliability of the first pattern extracted from the free-pitch box having the box number 1 shown in FIG. 76 is computed as 20% in the character recognizing process performed by the box-touching character recognizing unit 13 , then the first pattern extracted from the free-pitch box having the box number 1 as shown in FIG. 76 is regarded as a non-character and ‘REJECT’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘20%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- ‘BLACK-CHARACTER-BOX’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table, and the column ‘PROCESS ORDER’ on the intermediate process result table is updated to ‘ONECHARACTER DELETION LINE ⁇ FREE-PITCH’.
- the second pattern extracted from the free-pitch box having the box number 1 shown in FIG. 76 is recognized as the character category 3 with the recognition reliability of 60% in the character recognizing process performed by the box-touching character recognizing unit 13 , then ‘ 3 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, arid ‘60%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- ‘BLACK-CHARACTER-BOX’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table, and the column ‘PROCESS ORDER’ on the intermediate process result table is updated to ‘FREE-PITCH’.
- the eighth pattern extracted from the free-pitch box having the box number 1 shown in FIG. 76 is recognized as the character category 4 with the recognition reliability of 95% in the character recognizing process performed by the box-touching character recognizing unit 13 , then ‘ 4 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and '95V is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- ‘BLACK-CHARACTER-BOX’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table, and the column ‘PROCESS ORDER’ on the intermediate process result table is updated to ‘FREE-PITCH’.
- the deletion line recognizing unit 26 removes horizontal lines indicating the histogram value equal to or larger than a predetermined value from a pattern extracted as a candidate for a corrected character as shown in FIG. 71 . If the pattern from which the horizontal lines are removed is recognized as a character, then the pattern extracted as the candidate for the corrected character can be recognized as a corrected character by recognizing the removed horizontal lines as deletion lines. If the pattern from which the horizontal lines indicating the histogram value equal to or larger than the predetermined value are removed is rejected, then the removed horizontal lines are not regarded as deletion lines but a portion of a character. Thus, the pattern extracted as a candidate for a corrected character is recognized as a normal character.
- the recognition reliability of the pattern extracted from the one-character box having a box number 2 shown in FIG. 76 is computed as 10%.
- the pattern extracted from the one-character box having a box number 2 shown in FIG. 76 is not regarded as a corrected character, ‘10%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table, and ‘BASIC’ is entered in the column ‘PROCESS INSTRUCTION’ on the intermediate process result table.
- ‘DELETION LINE’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table, and ‘BASIC’ is entered in the column ‘PROCESS ORDER’ on the intermediate process result table.
- the basic character recognizing unit 17 computes the distance between feature vectors in a feature space by extracting the feature of an input unknown character as shown in FIG. 35 , representing the feature of the unknown character by a feature vector, and collating the vector with the feature vector of each category preliminarily stored in the basic dictionary.
- the character category indicating the shortest distance between the feature vectors is recognized as an unknown character.
- the basic character recognizing unit 17 computes the deformation of an unknown character by computing the number of convexity and concavity on the outline of the unknown character. If the unknown character indicates large deformation and reduces the recognition ratio, then the knowledge table 18 is referred to and a character recognizing process is performed by the detail identifying method shown in FIGS. 37 through 42 :
- the pattern extracted from the one-character box having the box number 3 shown in FIG. 76 is recognized as the character category 3 with the recognition reliability of 95% in the character recognizing process performed by the basic character recognizing unit 17 , then ‘ 3 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘95%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- the recognition reliability of the pattern extracted from the one-character box having the box number 4 shown in FIG. 76 is computed as 15% in the character recognizing process performed by the box-touching character recognizing unit 13 , then the pattern extracted from the one-character box having the box number 4 as shown in FIG. 76 is regarded as a non-character and ‘REJECT’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘15%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- ‘BLACK-CHARACTER-BOX’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table, and the column ‘PROCESS ORDER’ on the intermediate process result table is updated to ‘ONECHARACTER DELETION LINE’.
- the recognition reliability of the pattern extracted from the one-character box having a box number 5 - 1 shown in FIG. 76 is computed as 950 .
- the pattern extracted from the box having a box number 5 - 1 shown in FIG. 76 is regarded as a corrected character, ‘95%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table, and ‘DELETION LINE’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table.
- the lower stroke of the pattern extracted from the character box having the box number 5 - 2 shown in FIG. 76 touches the character box. Since the pattern cannot be recognized with high reliability in the character completing process shown in FIG. 43 or the character re-completing process shown in FIG. 44 , a pair of a character and its misread character ( 2 , 7 ) can be obtained by referring to the knowledge table 167 shown in FIG. 49 , and a character can be re-recognized by the area emphasizing method shown in FIG. 51 .
- the pattern extracted from the box having the box number 5 - 2 shown in FIG. 76 is recognized as the character category 2 with the recognition reliability of 95% in the character recognizing process performed by the box-touching character recognizing unit 13 , then ‘ 2 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘95%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- the pattern extracted from the box having the box number 5 - 3 shown in FIG. 76 is recognized as the character category 6 with the recognition reliability of 90% in the character recognizing process performed by the basic character recognizing unit 17 , then ‘ 6 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘90%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- the recognition reliability of the deletion line extracted from the table having a box number 6 - 1 - 1 shown in FIG. 76 is computed as 98%.
- the pattern extracted from the box having a box number 6 - 1 - 1 shown in FIG. 76 is regarded as a corrected character, ‘DELETION LINE’ is entered in the column ‘CHARACTER CODE’, ‘98%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table, and ‘DELETION LINE’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table.
- the recognition reliability of the deletion line extracted from the table having the box number 6 - 2 - 2 shown in FIG. 76 is computed as 98%.
- the pattern extracted from the box having the box number 6 - 2 - 2 shown in FIG. 76 is regarded as a corrected character, ‘DELETION LINE’ is entered in the column ‘CHARACTER CODE’, ‘98%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table, and ‘DELETION LINE’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table.
- the intermediate process result table shown in FIG. 80 is generated by performing the above described processes. Since the process to be called is entered in the column ‘PROCESS ORDER’ on the intermediate process result table shown in FIG. 80 , the processes are continued according to the process execution rule (C 1 ).
- FIG. 81 shows the result obtained by continuing the recognizing process based on the intermediate process result table shown in FIG. 80 .
- ‘ONE-CHARACTER DELETION LINE’ is specified for the first pattern in the column ‘PROCESS ORDER’ of the character box having the box number 1 on the intermediate process result table shown in FIG. 80 , the process of the deletion line recognizing unit 26 shown in FIG. 1 is performed on the first pattern extracted from the free-pitch character box having the box number 1 as shown in FIG. 76 corresponding to ‘ONE-CHARACTER DELETION LINE’ according to the process execution rule (C 1 ), and the corrected character recognizing process is performed.
- the recognition reliability of the first pattern extracted from the free-pitch character box having the box number 1 shown in FIG. 76 is computed as 96%.
- the first pattern extracted from the free-pitch box having the box number 1 shown in FIG. 76 is regarded as a corrected character, ‘DELETION LINE’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘96%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table, and ‘BLACK-CHARACTER-BOX/DELETION LINE’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table.
- ‘FREE-PITCH’ is specified in the column ‘PROCESS ORDER’ of the second pattern-in the character box having the box number 1 on the intermediate process result table shown in FIG. 79 . Therefore, the process of the character string recognizing unit 15 shown in FIG. 7 is performed corresponding to the ‘FREE-PITCH’ on all patterns extracted from the free-pitch box having the box number 1 to recognize a character with the detection reliability of a character taken into account according to the process execution rule (C 4 ) when the columns ‘PROCESS ORDER’ of all patterns in the character box having the box number 1 contain ‘FREE-PITCH’ after extracting the second pattern from the free-pitch box having the box number 1 shown in FIG. 76 and waiting for the columns ‘PROCESS ORDER’ of all other patterns in the character box having the box number 1 to contain ‘FREE-PITCH’.
- ‘FREE-PITCH’ is specified in the column ‘PROCESS ORDER’ of the eighth pattern in the character box having the box number 1 on the intermediate process result table shown in FIG. 79 . Therefore, the process of the character string recognizing unit 15 shown in FIG. 7 is performed corresponding to the ‘FREE-PITCH’ on all patterns extracted from the free-pitch box having the box number 1 to recognize a character with the detection reliability of a character taken into account according to the process execution rule (C 4 ) when the columns ‘PROCESS ORDER’ of all patterns in the character box having the box number 1 contain ‘FREE-PITCH’ after extracting the eighth pattern from the free-pitch box having the box number 1 shown in FIG. 76 and waiting for the columns ‘PROCESS ORDER’ of all other patterns in the character box having the box number 1 to contain ‘FREE-PITCH’.
- the character recognizing process is performed by the character string recognizing unit 15 on all patterns extracted from the free-pitch box having the box number 1 shown in FIG. 76 when the columns ‘PROCESS ORDER’ of all patterns in the character box having the box number 1 contain ‘FREE-PITCH’.
- the recognizing process is performed by the character string recognizing unit 15 on the second through eighth patterns extracted from the free-pitch box having the box number 1 shown in FIG. 76 after removing the first pattern extracted from the free-pitch box in the character box having the box number 1 shown in FIG. 76 from the process objects of the character string recognizing unit 15 because ‘TERMINATION’ is entered in the column ‘PROCESS INSTRUCTION’ of the first pattern in the character box having the box number 1 on the intermediate process result table shown in FIG. 81 .
- the character string recognizing unit 15 computes the recognition reliability based on the distance from the discriminant phase when a character is detected as shown in, for example, FIGS. 56 through 69 , and defines that the maximum product of character detection reliability and character recognition reliability refers to a detected character.
- the second pattern extracted from the free-pitch box having the box number 1 shown in FIG. 76 is recognized as a character category ‘ 3 ’ with recognition reliability of 95%.
- ‘ 3 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘95%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- ‘BLACK-CHARACTER-BOX/FREE-PITCH’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table, and the column ‘PROCESS ORDER’ on the intermediate process result table becomes blank.
- ‘PERSONAL HANDWRITING FEATURES’ is entered in the column ‘PROCESS INSTRUCTION’ on the intermediate process result table.
- the eighth pattern extracted from the free-pitch box in the character box having the box number 1 shown in FIG. 76 is recognized as a character category ‘ 4 ’ with recognition reliability of 98%.
- ‘ 4 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘98%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- ‘BLACK.-CHARACTER-BOX/FREE-PITCH’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table, and the column ‘PROCESS ORDER’ on the intermediate process result table becomes blank.
- ‘PERSONAL HANDWRITING FEATURES’ is entered in the column ‘PROCESS INSTRUCTION’ on the intermediate process result table.
- the third pattern extracted from the free-pitch box in the character box having the box number 1 shown in FIG. 76 is recognized as a character category ‘ 2 ’.
- the fourth and fifth patterns extracted from the free-pitch box in the character box having the box number 1 shown in FIG. 76 are integrated into a single character in the recognizing process by the character string recognizing unit 15 and recognized as a character category ‘ 7 ’.
- the sixth pattern extracted from the free-pitch box in the character box having the box number 1 shown in FIG. 76 is recognized as a character category ‘ 4 ’.
- the seventh pattern extracted from the free-pitch box in the character box having the box number 1 shown in FIG. 76 is recognized as a character category ‘ 6 ’.
- the pattern extracted from the one-character box in the character box having the box number 2 shown in FIG. 76 is recognized as a character category ‘ 5 ’ with recognition reliability of 97%.
- ‘ 5 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘97%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- ‘DELETION LINE (YES 2 )/BASIC’ is entered in the column ‘CALLING PROCESS’ on the intermediate process result table.
- ‘DELETION LINE/BASIC’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table.
- the column ‘PROCESS ORDER’ on the intermediate process result table becomes blank. Therefore, according to the process execution rule (C 4 ), ‘PERSONAL HANDWRITING FEATURES’ is entered in the column ‘PROCESS INSTRUCTION’.
- the recognition reliability of the pattern extracted from the one-character box having the box number 4 shown in FIG. 76 is computed as 95% in the deletion line recognizing process performed by the deletion line recognizing unit 26 , then the pattern extracted from the one-character box having the box number 4 as shown in FIG. 76 is regarded as a corrected character and ‘95%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table, and ‘BLACK-CHARACTER-BOX/DELETION LINE’ is entered in the column ‘COMPLETION OF PROCESS’ on the intermediate process result table.
- the intermediate process result table shown in FIG. 81 is generated by performing the above described processes. Since the ‘PERSONAL HANDWRITING FEATURES’ is entered in a column ‘PROCESS INSTRUCTION’ on the intermediate process result table shown in FIG. 81 , the processes are continued according to the process execution rule (C 5 ).
- FIG. 82 shows the result obtained by continuing the recognizing process based on the intermediate process result table shown in FIG. 81 .
- the unique character analyzing unit 23 clusters the characters handwritten by the same writer into categories as shown in FIGS. 72 through 75 .
- the second cluster which is close to the first cluster of written characters obtained by the clustering process, belongs to a category different from that of the first cluster, and has a smaller number of elements, is integrated into the first cluster so that the category of the handwritten characters belonging to the second cluster can be amended to the category of the first clusters.
- the second pattern extracted from the free-pitch box having the character box number 1 shown in FIG. 76 is recognized as character category 3 with the recognition reliability of 97%.
- ‘ 3 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘97%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- the eighth pattern extracted from the free-pitch box having the character box number 1 shown in FIG. 76 is recognized as character category 4 with the recognition reliability of 98%.
- ‘ 4 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘98%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- the pattern extracted from the one-character box having the character box number 2 shown in FIG. 76 is recognized as character category 5 with the recognition reliability of 97%.
- ‘ 5 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘97%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- ‘DELETION LINE/BASIC/PERSONAL HANDWRITING FEATURES’ is entered in the column ‘COMPLETION OF PROCESS’, and ‘TERMINATION’ is entered in the column ‘PROCESS INSTRUCTION’.
- the pattern extracted from the one-character box having the box number 3 shown in FIG. 76 is recognized as character category 3 with the recognition reliability of 97%.
- ‘ 3 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘97%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- the pattern extracted from the character box having the box number 5 - 2 shown in FIG. 76 is recognized as character category 2 with the recognition reliability of 97%.
- ‘ 2 ’ is entered 15 in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘97%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- the pattern extracted from the character box having the box number 5 - 3 shown in FIG. 76 is recognized as character category 4 with the recognition reliability of 96%.
- ‘ 4 ’ is entered in the column ‘CHARACTER CODE’ on the intermediate process result table, and ‘96%’ is entered in the column ‘RELIABILITY’ on the intermediate process result table.
- the intermediate process result table shown in FIG. 82 is generated by performing the above described processes. Since ‘TERMINATION’ is entered in the column ‘PROCESS INSTRUCTION’ on the intermediate process result table shown in FIG. 82 for all process objects, all processes terminate according to the process execution rule (,C 6 ).
- the character recognizing unit 12 and non-character recognizing unit 25 perform appropriate recognizing processes to process the state of the input image recognized by the environment recognizing system 11 according to the embodiments of the present invention.
- the environment recognizing system 11 when the environment recognizing system 11 extracts a character touching the ruled line, it uses the box-touching character recognizing unit 13 for exclusively performing a recognizing process on a box-touching character.
- the environment recognizing system 11 extracts a free-pitch character string, it uses the character string recognizing unit 15 for exclusively performing a recognizing process on a free-pitch character string.
- the environment recognizing system 11 extracts an obscure character, it uses the obscure character recognizing unit 19 for exclusively performing a recognizing process on an obscure character.
- the environment recognizing system 11 extracts a deformed character, it uses the deformed character recognizing unit 21 for exclusively performing a recognizing process on a deformed character.
- the environment recognizing system 11 extracts a non-character, it uses the non-character recognizing unit 25 for exclusively performing a recognizing process on a non-character.
- the reliability on the recognition result from the character recognizing unit 12 or non-character recognizing unit 25 is computed. For a character or non-character with low reliability, the environment recognizing system 11 , character recognizing unit 12 , and non-character recognizing unit 25 mutually feed back data to re-perform other processes. When high reliability is obtained or there are no executable processes to be performed, the entire process terminates.
- a recognizing process can be performed with the features and identifying methods to be used when characters are recognized adaptively amended depending on the environment in which characters are written. Therefore, a high-precision character recognition can be realized corresponding to various environments of documents and forms.
- character recognition results can be confirmed with high reliability by outputting only character codes as recognition results, simultaneously outputting environment recognition results and character recognition results, and outputting character recognition results when the environment recognition results and character recognition results match each other.
- non-character recognizing unit 25 can be provided exclusively for performing a non-character recognizing process independent of I a character recognizing process, the reliability in the character and non-character recognizing processes can be improved.
- the recognition reliability can be improved by increasing the volume of the dictionary and knowledge in each recognizing process.
- the pattern recognizing apparatus correctly rejects deletion lines apparently indicating the deletion of a character to prevent it from being mis-read, but does not mistakenly reject a character other than a character with deletion lines, thereby recognizing a character with high reliability without a heavy load to a user.
- the third embodiment shown in FIG. 83 comprises an image input unit 411 for receiving an image, detecting a character pattern from the input image, and pre-processing the detected pattern; a deletion line character discriminating unit 412 for discriminating a character with deletion lines of either simple deletion lines formed simply by horizontal lines indicating the deletion of the character or complicated deletion lines formed by complicated strokes; or a normal character without deletion lines; and an identifying unit 413 for identifying a character.
- ‘Inputting an image’ refers to inputting an original image representing a character.
- the character pattern on the form is transmitted to the read area of an opto-electrical converter to read the character pattern, convert it into an electrical signal, and output a digital signal through binarization, etc.
- Detecting a character refers to detecting only a character portion from the form data of the input digital image and separating each character from others.
- Pre-processing refers to removing noises, standardizing, the position, size, pose, line-thinness, etc. of a character.
- Discriminating a character with deletion lines can be a pre-process.
- a ‘character with deletion lines’ refers to a character provided with deletion lines indicating the deletion of the character.
- deletion lines can be simple deletion lines or complicated deletion lines.
- a ‘simple deletion line’ refers to a deletion line formed by a horizontal line assumed to indicate the deletion of a character.
- the simple deletion line can be one or more horizontal lines.
- a ‘horizontal line’ is not always a correctly-horizontal line, but includes a line drawn in the horizontal direction with permitted allowance in slope.
- horizontal deletion lines are discriminated by a contiguous projection.
- a ‘complicated deletion line’ refers to a complicated line or stroke drawn on a character to be deleted. To determine whether or not a line is a complicated deletion line, the amount of complexity is extracted by the pattern recognizing apparatus shown in FIG. 84 .
- Identifying a character refers to extracting a feature, inputting the obtained feature vector, and computing the state of matching between the feature vector and a standard feature vector, that is, a preliminarily stored identifying dictionary.
- the category of the standard feature vector indicating the highest similarity is output as a recognition result.
- a discriminating process is performed on a character with a simple deletion line or a complicated deletion line so that a normal character can be identified.
- deletion lines as a predetermined horizontal line, an apparent deletion line simple and a little irregular from the predetermined horizontal line, and an apparent deletion line complicated and quite different from the simple deletion line can be discriminated as deletion lines without fail.
- FIG. 84 is a block diagram showing the configuration of the pattern recognizing apparatus according to the fourth embodiment of the present invention.
- the deletion line character discriminating unit 412 comprises a complicated deletion line character discrimination process unit 414 for discriminating a character with a complicated deletion line and a simple deletion line character discriminating unit 415 for discriminating a character with a simple deletion line by determining whether or not a candidate for a character with a simple deletion line can be identified as a character when the candidate for the simple deletion line is removed from the candidate for the character with the simple deletion line when the plural-character deletion line character determination process unit 414 determines that the pattern is not a character with a complicated deletion line.
- a character with a simple deletion line can be discriminated only when it is determined that the character is not provided with a complicated deletion line after first determining whether or not the character is a character with a complicated deletion line. Therefore, the pattern recognizing apparatus avoids misreading a character for a character without a deletion line when it cannot be identified as a character after removing a simple deletion line from it, thereby discriminating with high reliability a candidate for a character with a deletion line.
- FIG. 85 is a block diagram showing the configuration of the pattern recognizing apparatus according to the fifth embodiment of the present invention.
- the deletion line character discriminating unit 412 comprises a character form information extracting unit 416 for extracting character form information comprising at least one of the picture element number histogram, obtained by the contiguous projection in which the black picture elements in the object horizontal scanning line and the contiguous lines in a predetermined range above and below the object line are added up and counted in the horizontal direction, and the amount of complexity indicating the complexity of drawings from the detected and pre-processed character pattern; a deletion line character candidate discriminating unit 417 for discriminating a candidate, for a character with a simple deletion line or a character with a complicated deletion line according to the character form information extracted by the character form information extracting unit 416 ; and a deletion line character determining unit 418 for determining a character with a simple deletion line when the character can be identified as a character by the identifying unit 413 after removing a candidate for a deletion line from the candidate for a character with a simple deletion line.
- a character form information extracting unit 416 for extracting character form information comprising
- the ‘character form information’ comprises at least one of the ‘picture-element-number histogram obtained by a contiguous projection performed in the horizontal direction’ and ‘amount of complexity’.
- a ‘predetermined range’ refers to a range of contiguous lines above and below the object horizontal line.
- the range can contain the three lines, that is, the object horizontal line and the lines immediately above and below the object line according to the embodiment described later.
- the ‘predetermined range’ is set narrow, then only lines in the horizontal direction can be recognized as simple deletion lines. If it is set larger, then even a line making an angle with the horizontal line can be recognized as a simple deletion line. However, if it is set too large, the peak of the picture element histogram becomes dull and the horizontal lines cannot be correctly determined. Therefore, the range is appropriately set depending on experiences and experiments.
- the number of black picture elements is counted by adding up the black picture elements in the lines in a predetermined range containing an object horizontal scanning line and the contiguous lines above and below the object line.
- An ‘amount of complexity’ can be represented by, for example, the line density in a predetermined direction, Euler number, number of black picture elements, etc.
- a character with a simple deletion line is discriminated by the picture-element-number histogram obtained by a contiguous projection in the horizontal direction, and a character with a complicated deletion line is discriminated by the amount of complexity.
- the above described amount of complexity or the picture-element-number histogram obtained by the contiguous projection in the horizontal direction can be obtained objectively, easily, rapidly, and with high reliability. Therefore, a character with a deletion line can be easily discriminated at a high speed and with high reliability.
- FIG. 86 is a block diagram showing the configuration of the pattern recognizing apparatus according to the sixth embodiment of the present invention.
- the complicated deletion line character discrimination process unit 414 comprises a complexity amount extracting unit 419 for extracting the amount of complexity indicating the complexity of drawings, and a complicated deletion line character discriminating unit 420 for determining a character with a complicated deletion line based on the extracted amount of complexity.
- the simple deletion line character discriminating unit 415 comprises a 25 picture element number histogram computing unit 421 for computing the picture element number histogram by the contiguous projection in which the black picture elements in the object horizontal scanning line and the contiguous lines in a predetermined range above and below the object line are added up and counted in the horizontal direction; a simple deletion line character candidate discriminating unit 422 for discriminating a candidate for a character with a simple deletion line based on the computed number of picture elements, and a simple deletion line character determining unit 423 for determining a character with a simple deletion line when the character can be identified as a character by the identifying unit 413 after removing a candidate for a deletion line from the candidate for a character with a simple deletion line.
- a character with a deletion line can be discriminated based on the picture element number histogram obtained by the contiguous projection in the horizontal direction and the amount of complexity.
- the character with a deletion line can be discriminated with high reliability.
- FIG. 87 is a block diagram showing the configuration of the pattern recognizing apparatus according to the seventh embodiment of the present invention.
- the simple deletion line character determining unit 423 comprises a deletion line removing unit 424 for removing a candidate for a deletion line from a discriminated candidate for a character with a simple deletion line and transmitting the result to the identifying unit 413 ; a storage unit 426 for storing an original image of a candidate for a character with a simple deletion line before removing the candidate for a deletion line; and a deletion line character determining unit 425 for defining a candidate for a character with a simple deletion line as a character with a simple deletion line when the candidate for the character with a simple deletion line can be identified as a character even after removing the deletion line from the character, and for defining the candidate for a character with a simple deletion line stored in the storage unit 426 as a normal character and transmitting it to the identifying unit 413 when the candidate for the character with a simple deletion line cannot be identified as a character after removing the deletion line from the character.
- the original image before removing by the deletion line removing unit 424 a candidate for a deletion line from a candidate for a character with a deletion line is temporarily stored in the storage unit 426 .
- the candidate, from which the candidate for a deletion line is removed, for a character with a deletion line cannot be identified as a character, then the original image of the candidate for a character with a deletion line is read from the storage unit 426 and transmitted to the identifying unit 413 for identification of a character.
- a character can be identified at a high speed with a simple configuration according to the present invention.
- FIG. 88 is a flowchart showing the operations of the pattern recognizing apparatus according to the eighth embodiment of the present invention:
- an image is input, a character pattern is detected from the input image, and the result is pre-processed in step S 301 . Then, a discriminating process is performed in step S 302 on the detected and pre-processed character pattern as to whether the current character is a character with a deletion line provided with either a simple deletion line formed by only a horizontal line or a complicated deletion line drawn by a complicated stroke to indicate the deletion of the character or a normal character without a deletion line. Thus, a normal character without a deletion line can be identified as a character in step S 303 .
- the current character is a character with a deletion line provided with either a simple deletion line or a complicated line or a normal character without a deletion line.
- a normal character can be identified as a character.
- FIG. 89 is a flowchart showing the operations of the pattern recognizing apparatus according to the ninth embodiment of the present invention.
- an image is input, a character pattern is detected from the input image, and the result is pre-processed in step S 304 . Then, a discriminating process is performed in step S 305 to discriminate a character with a complicated deletion line having a complicated form on the detected and pre-processed character pattern. If it is determined that the current character is not a character with a complicated deletion line, then it is determined whether the current character is a character with a simple deletion line formed by only a horizontal line to indicate the deletion of the character or a normal character without a deletion line in step S 306 . Thus, a normal character without a deletion line can be identified as a character in step S 307 .
- a discriminating process when a discriminating process is performed on a character with a deletion line, it is first performed on a character with a complicated deletion line. Only if it is determined that the current character is not a character with a complicated deletion line, the discriminating process is performed on a candidate for a character with a simple deletion line.
- a candidate for a character with a simple deletion line is defined as a character with a simple deletion line when the character can be identified as a character even after removing a candidate for a deletion line from the candidate for a character. If the candidate for the character cannot be identified as a character, it is defined as a normal character.
- the determination as to whether or not the candidate is a character with a deletion line can be made correctly with high reliability.
- FIG. 90 is a flowchart showing the operations of the pattern recognizing apparatus according to the tenth embodiment of the present invention.
- an image is input, a character pattern is detected from the input image, and the detected result is pre-processed in step S 311 .
- the amount of complexity of the detected and pre-processed character is computed in step S 312 .
- a discriminating process about a character with a complicated deletion line represented by a complicated stroke is performed in step S 313 . If it is determined that the character is not provided with a complicated deletion line, then the picture element number histogram is obtained by the contiguous projection in which the black picture elements in the object horizontal scanning line and the contiguous lines in a predetermined range above and below the object line are added up and counted in the horizontal direction in step S 314 .
- the discriminating process is performed about the candidate for a character with a simple deletion line in step S 315 . If a candidate for a deletion line is removed from a candidate for a character with a simple deletion line and the result can be identified as a character, then the candidate is defined as a character with a simple deletion line. If it cannot be identified as a character, the candidate is defined as a normal character in step S 316 , and it is determined in step S 317 whether or not the defined normal character can be identified as a character.
- a discriminating process is performed about a character with a simple deletion line or a character with a complicated deletion line by extracting the amount of complexity or performing the contiguous projection.
- FIG. 91 is a flowchart showing the operations of the pattern recognizing apparatus according to the eleventh embodiment of the present invention.
- the amount of complexity such as an Euler number, the line density, the density of black picture elements, etc. is computed in step S 321 about the detected and pre-processed character.
- a determination as to whether or not the current character is provided with a complicated deletion line is made using a predetermined threshold in step S 322 on the computed amount of complexity. If it is determined that the current character is a character with a complicated deletion line, then a reject code is output in step S 323 . If it is not determined that the current character is provided with a complicated deletion line, then a horizontal direction contiguous projection histogram is computed in step S 314 , and it is determined whether or not a candidate for a character with a simple deletion line exists in step S 351 .
- step S 352 If there is a candidate for a deletion line, it is removed as a candidate for a character with a simple deletion line in step S 352 . If the candidate for a deletion line is removed from the candidate for a character with a simple deletion line, and the result can be identified as a character, the candidate for a character with a simple deletion line is defined as a character with a simple deletion line. If the result cannot be identified as a character, it is determined whether or not the current character is a character with a simple deletion line in step S 351 when it is defined as determines a normal character a normal character (step S 353 and S 354 ).
- the defined normal character is checked whether or not it can be identified as a character in step S 361 and S 262 . If it cannot be identified as a character, then a reject code is output in step S 363 . If it can be identified as a character, the result of the recognition is output (step S 364 ).
- FIG. 92 is a block diagram showing the configuration of the pattern recognizing apparatus according to the twelfth embodiment of the present invention.
- the apparatus comprises an optical reading unit 430 for optically reading transmitted character pattern entered in a form, converting the, character pattern into an electrical signal, and outputting a digital signal through binarization, etc.; a transmitting unit 431 for transmitting the form to the optical reading area of the optical reading unit 430 ; a dictionary 432 storing the standard feature vector of a character; a display unit for displaying a character on a screen; an output unit 433 for printing a character on paper; an operating unit 434 for performing various operations; and a CPU and memory 435 having various functions for character recognition.
- the CPU and memory 435 comprises a character detecting and preprocessing unit 436 for detecting and pre-processing a character from an input image; a complicated deletion line character discrimination process unit 437 for performing a discriminating process on a character with a complicated deletion line represented by a complicated stroke; a simple deletion line character discriminating unit 438 for performing a discriminating process on, a candidate for a character with a simple deletion line formed only by a horizontal line indicating the deletion of a character when it is determined by the complicated deletion line character determination process unit 437 that the character is not provided with a complicated deletion line; an identifying unit 439 for identifying a character; and a result output directing unit 440 for rejecting the character if it is discriminated as a character with a deletion line, outputting a character identification result if the character is discriminated as a normal character, and giving a direction to output the rejection if the character is discriminated as a normal character but cannot be identified as a character.
- the character detecting and preprocessing unit 436 comprises a character detecting unit 441 for detecting only the character portion from the form image of an input digital image and separating characters from each other, and a pre-processing unit 442 for removing a noise from a detected character signal and standardizing the position, size, etc. of a character.
- the optical reading unit 430 , transmitting unit 431 , and character detecting and preprocessing unit 436 correspond to an image input unit.
- the complicated deletion line character discrimination process unit 437 comprises a complexity amount extracting unit 443 for extracting the amount of complexity from a character pattern; and a complicated deletion line character discriminating unit 444 for discriminating whether or not the character pattern is a character with a complicated deletion line based on the extracted amount of complexity.
- the ‘amount of complexity’ can be the line density in a predetermined direction, an Euler number, black picture element density, etc.
- the ‘line density in a predetermined direction’ refers to-a value obtained by counting the portions changing from white picture elements into black picture elements (or black picture elements into white picture elements) when an image in a rectangle is scanned in a predetermined direction. For example, if a character pattern is formed by a character ‘ 2 ’ with a deletion line 501 as indicated by (A) in FIG. 93 , the line density in the vertical direction as indicated by (D) in FIG. 93 is 6.
- the ‘predetermined direction’ normally refers to the vertical or horizontal direction to a character.
- FIG. 93 shows examples of characters with complicated deletion line.
- a complicated deletion line character discriminating unit 444 performs a discriminating process based on the amount of complexity such as the line density, Euler number, or black picture element density, etc. of the extracted feature or an appropriate combination of them.
- the simple deletion line character discriminating unit 438 shown in FIG. 92 comprises a picture element histogram computing unit 445 for computing the picture element number histogram computing unit 445 for computing the picture elements in the object horizontal scanning line and the contiguous lines in a predetermined range above and below the object line are added up and counted in the horizontal direction; a simple deletion line character candidate discriminating unit 446 for performing a discriminating process about a candidate for a character with a simple deletion line based on the computed number of picture elements; and a simple deletion line character determining unit 447 for determining a character with a simple deletion line if the identifying unit 439 identifies a character when a candidate for a deletion line is removed from the discriminated candidate for a character with a simple deletion line.
- the ‘contiguous projection in the horizontal direction’ refers to adding up and counting in the horizontal direction the black picture elements in the object horizontal scanning line and the contiguous lines in a predetermined range above and below the object line.
- a predetermined range contains 3 lines.
- the simple deletion line character candidate discriminating unit 446 determines a candidate for a character with a simple horizontal deletion line having a peak whose picture element number histogram exceeds a predetermined threshold, and recognizes the corresponding line as a candidate for a deletion line.
- the candidate for a character with a deletion line determined by the simple deletion line character candidate discriminating unit 446 and stripped of the candidate for a deletion line by the deletion line removing unit 450 can be identified as a character, then it is defined as a character with a deletion line by the simple deletion line character determining unit 448 . If it cannot be identified as a character, then it is defined as a normal character (character without a deletion line). A candidate for a deletion line is removed by the deletion line removing unit 450 by an existing image processing method.
- the simple deletion line character determining unit 447 comprises a deletion line removing unit 450 for removing a candidate for a deletion line from the discriminated candidate for a character with a simple deletion line and transmitting the result to the identifying unit 439 ; the simple deletion line character determining unit 448 for determining the candidate for a character with a simple deletion line as a character with a simple deletion line when the discriminated candidate for a character with a simple deletion line is recognized as a character, and determining it as a normal character and transmitting it to the identifying unit 439 when it is not recognized as a character; and a character candidate storage unit 449 for storing the candidate for a character with a deletion line removed by the deletion line removing unit 450 .
- the identifying unit 439 comprises a feature extracting unit 451 for extracting the feature value of each character pattern and compressing data; and a dictionary collating unit 452 for collating a character pattern with a dictionary, that is, a standard feature vector of each character type.
- the transmitting unit 431 transmits a form containing entered characters to the reading area of the optical reading unit 430 and has an OCR input the form in step SJ 1 .
- step SJ 2 the optical reading unit 430 converts a character pattern on the form into an electrical signal in the opto-electrical conversion, and outputs it as a digital signal through binarization, etc.
- the form can contain entered characters and their character boxes in the same color.
- step SJ 3 the character detecting unit 441 detects a character portion from the digital signal and separates characters from one another.
- the preprocessing unit 442 standardizes the position, size, slope line thinness, etc.
- step SJ 4 a complicated deletion line character discrimination process unit 437 performs a discriminating process about a character with a complicated deletion line.
- step SJ 5 it does not recognize the current character pattern as a character with a complicated deletion line.
- the deletion line existence determining unit A shown in FIG. 95 shows the processes in step SJ 4 and SJ 5 indicating the contents of the process performed by the complicated deletion line character discrimination process unit 437 .
- step SJ 41 the complexity amount extracting unit 443 extracts the amount of complexity such as an Euler number, line density, black picture element density, etc. to determine whether or not the current character pattern is a character with a complicated deletion line.
- step SJ 42 the complicated deletion line character discriminating unit 444 determines using a predetermined threshold whether the object character pattern is a normal character or a character with a complicated deletion line.
- the threshold of the line density by the scanning in the horizontal direction is 2
- the threshold of the line density by scanning in the vertical direction is 3
- the threshold of an Euler number is ⁇ 1
- the threshold of the black picture element density is 0.6
- a rejection code is output as a recognition result in step SJ 5 .
- step SJ 6 If the character pattern is not discriminated as a character with a complicated deletion line, control is passed to step SJ 6 as shown in FIG. 94 , and the simple deletion line character discriminating unit 438 performs a process of discriminating whether or not a horizontal line, that is, a single deletion line exists. If the pattern is discriminated as a character with a deletion line, then a rejection code is output as a recognition result in step SJ 7 .
- the discriminating process is indicated by the deletion line existence determining unit B in FIG. 95 .
- the picture element histogram computing unit 445 generates a horizontal direction contiguous projection histogram in step SJ 61 .
- the simple deletion line character candidate discriminating unit 446 determines that there is a candidate for a deletion line in step SJ 63 and discriminates the pattern as a candidate for a character with a simple deletion line. If such a peak does not exist, it is determined that the pattern does not contain a candidate for a deletion line and the pattern is discriminated as a normal character. If a character size is standardized in a pre-process, N can be a fixed value. If it is not standardized in a pre-process, it is recommended that N is variable depending on the width of the rectangle circumscribing the object character. In this case, the ratio of the threshold N to the width of the circumscribing rectangle should be appropriately given as a fixed value.
- step SJ 63 if the simple deletion line character candidate discriminating unit 446 discriminates the pattern as a candidate for a character with a deletion line, then control is passed to step SJ 64 , and the character candidate storage unit 449 stores a candidate for a character with a deletion line (before removing a deletion line).
- the deletion line removing unit 450 detects and deletes a horizontal simple deletion line of the candidate for a character with a deletion line.
- a candidate for a deletion line is removed by an existing method, for example, a method of extracting a line through an n line run length, etc.
- the feature extracting process is performed by the identifying unit 439 on a character pattern from which a deletion line has been removed in step SJ 8 .
- step SJ 9 the extracted feature is collated with the dictionary.
- the matching level is computed by referring to a given standard feature vector, and a character type with a feature vector indicating the highest similarity is output as a recognition result.
- the simple deletion line character determining unit 448 determines that the candidate for a character with a deletion line is a character without a deletion line in step SJ 66 , transmits the original image of the candidate for a character with a deletion line temporarily stored in the character candidate storage unit 449 to the identifying unit 439 which identifies a character in step SJ 67 .
- the result output directing unit 440 instructs the output unit 433 to output the identification result.
- step SJ 65 If it is determined that the pattern has been identified as a character after the collating process in step SJ 65 , then control is passed to step SJ 68 and the simple deletion line character determining unit 448 discriminates the candidate for a character with a deletion line as a character with a deletion line, and the result output directing unit 440 outputs the recognition result as ‘REJECT’ to the output unit 433 in step SJ 69 .
- the upper line in FIG. 71 shows an example of recognizing a candidate for a deletion line about a candidate for a character with a simple deletion line, identifying as ‘ 5 ’ the pattern stripped of the candidate for a deletion line, and determining the pattern as a character with a deletion line.
- the lower line in FIG. 71 shows an example of recognizing a candidate for a deletion line in the candidate for a character with a simple deletion line, rejecting the pattern stripped of the candidate for a deletion line, and therefore discriminating the candidate for the original character with a simple deletion line as a normal character.
- the present embodiment realizes the discrimination of a character with a complicated deletion line with high reliability at a high speed.
- a candidate for a character with a simple deletion line can be discriminated by counting the number of picture elements in the contiguous projection method.
- a candidate for a character with a deletion line including a line drawn in a roughly horizontal direction as well as a predicted horizontal simple deletion line can be easily and quickly discriminated with high reliability with a simple configuration.
- misreading a character with a deletion line can be reduced by performing a deletion line existence determining process according to the aspect of the present embodiment. Additionally, a character deleted by a deletion line can be clearly rejected.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Artificial Intelligence (AREA)
- Character Discrimination (AREA)
- Character Input (AREA)
Abstract
Description
A 1=(1-1)−(1-2)
A 2=(2-1)−(2-2)+(2-3)
P(i)=p(i−j)+. . . +p(i)+. . . +p(i+j) (1)
-
- where dx (position indicated by black dot in
FIG. 47 ) and dy (position indicated by x inFIG. 47 ) respectively indicate the difference in position of the center of gravity of the character and the character box between the x and y directions.
- where dx (position indicated by black dot in
-
- where dsx and dsy indicate the size of a character in the x and y directions respectively.
-
- where dα indicates the angle of the pose of the character to the vertical line.
-
- where fα indicates the angle of the pose of the character box to the vertical line.
-
- where w indicates the width of the character box.
-
- where fsx and fsy indicate the size of the character in the x and y directions respectively.
-
- where fδ is a parameter for controlling, for example, the concavity and convexity of a character box with the deterioration, etc. of the quality of a printed character box on a facsimile taken into account. Assuming that the circumference of the character box is represented by L, fδ is represented by the array fδ [L] in size L. In this array, each element fδ [i] (i=1, 2, 3, . . . ) is an integer in the range of −β−+β determined by a random number.
Σ−1(μ1−μ2) (3)
where
-
- Σ indicates the variance co-variance matrix of the first and second groups;
- μ1 indicates the mean vector of the first group; and
- μ2 indicates the mean vector of the second group.
-
- 1) distance a between the right side of the circumscribing
rectangle 211 and the left side of the circumscribingrectangle 212 - 2) distance b between the left side of the circumscribing
rectangle 211 and the right side of the circumscribingrectangle 212 - 3) ratio c of the distance a between the right side of the circumscribing
rectangle 211 and the left side of the circumscribingrectangle 212 to the distance b between the left side of the circumscribingrectangle 211 and the right side of the circumscribingrectangle 212 - 4) ratio d of the distance b between the left side of the circumscribing
rectangle 211 and the right side of the circumscribingrectangle 212 to an average width MX of circumscribing rectangles - 5) angle a made between the lower side of the circumscribing
rectangle 213 and the line connecting the mid-point of the lower side of the circumscribingrectangle 213 to the mid-point of the lower side of the circumscribingrectangle 214 - 6) angle f made between the lower side of the circumscribing
rectangle 213 and the line connecting the right-bottom vertex of the circumscribingrectangle 213 to the left-bottom vertex of the circumscribingrectangle 214 - 7) when the circumscribing
rectangle 215 overlaps the circumscribingrectangle 216;
ratio g of the distance p between the right side of the circumscribingrectangle 215 and the left side of the circumscribingrectangle 216 to the distance q between the left side of the circumscribingrectangle 215 and the right side of the circumscribingrectangle 216.
- 1) distance a between the right side of the circumscribing
c=a/b (4)
d=b/MX (5)
g=p/q (6)
N(=0.8)<P<M(=1.2)
-
- 8) ration p of the distance a between the right side of the circumscribing
rectangle 221 and the left side of the circumscribingrectangle 222 to the distance b between the left side of the circumscribingrectangle 221 and the right side of the circumscribingrectangle 222 - 9) ratio q of the distance b between the left side of the circumscribing
rectangle 221 and the right side of the circumscribingrectangle 222 to an average width MX of circumscribing rectangles - 10) ratio r of the product of the area c of the circumscribing
rectangle 21 and the area d of the circumscribingrectangle 22 to the square of the product of the average width MX of circumscribing rectangles and the average height MY of the circumscribing rectangles.
- 8) ration p of the distance a between the right side of the circumscribing
p=a/b (7)
q=b/MX (8)
r=(c×d)/(MX×MY)2 (9)
-
- 11) ratio p of the distance a between the right side of the circumscribing
rectangle 231 and the left side of the circumscribingrectangle 232 to the distance b between the left side of the circumscribingrectangle 231 and the right side of the circumscribingrectangle 232 - 12) ratio q of the distance b between the left side of the circumscribing
rectangle 231 and the right side of the circumscribingrectangle 232 to an average width MX of circumscribing rectangles - 13) ratio r of the product of the area c of the circumscribing
rectangle 231 and the area d of the circumscribingrectangle 232 to the square of the product of the average width MX of circumscribing rectangles and the average height MY of the circumscribing rectangles is used as a detection parameter.
- 11) ratio p of the distance a between the right side of the circumscribing
h=OP·n (10)
where OP is a vector from the
T=d/S (11)
where S indicates the area of the
-
- in step S234, the distance L from the center to the end of the
normal distribution curve 262 is computed by the following equation (12)
L=k×(1+T)×v 1/2
where k indicates a constant of proportionality, and v1/2 equals a standard deviation.
- in step S234, the distance L from the center to the end of the
0.84×0+0.43×1+0.33×2−145.25=0 (14)
α=(h−w 1)/(w 2 −w 1)×10 (15)
where
-
- w1 indicates the leftmost position of the 2-group overlap area
- w2 indicates the rightmost position of the 2-group overlap area
p=a/b=0.1 (16)
q=b/MX=1.3 (17)
r=(c×d)/(MX×MY)2=0.3 (18)
where
-
- a indicates the distance between the right side of the circumscribing
rectangle 281 and the left side of the circumscribingrectangle 272; - b indicates the distance between the left side of the circumscribing
rectangle 281 and the right side of the circumscribingrectangle 272; - c indicates the area of the circumscribing
rectangle 281; - d indicates the area of the circumscribing
rectangle 272; - MX indicates an average width of a circumscribing rectangle, and
- MY indicates an average height of a circumscribing rectangle.
- a indicates the distance between the right side of the circumscribing
0.17×x 0+0.75×x 1+0.64×x 2+30.4=0 (19)
w 1=38−2×(1+0.2)×25=−22 (21)
w 2=−34+2×(1+0.3)×28=38.8 (22)
α=(29.6−(−22))/(38.8−(−22))×100=85 (23)
-
- (1) the number of clusters reaches a predetermined value (for example, 3);
- (2) the distance between, clusters exceeds a predetermined threshold when the clusters care integrated; and
- (3) the increase ratio of the distance between clusters exceeds a predetermined threshold when the clusters are integrated.
where xai (i=1, 2, . . . , M) indicates the feature vector in cluster A.
| character with | ||
| feature/tendency | normal character | deletion line |
| maximum line density | small | large |
| Euler number | small abstract | negative value large |
| value (2˜−1) | absolute value | |
| back picture element | small | large |
| density | ||
Claims (2)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US10/000,525 US6850645B2 (en) | 1996-01-09 | 2001-12-04 | Pattern recognizing apparatus |
Applications Claiming Priority (7)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| JP00173096A JP3345246B2 (en) | 1996-01-09 | 1996-01-09 | Character recognition device and character recognition method |
| JP08-001730 | 1996-01-09 | ||
| JP25521796 | 1996-09-27 | ||
| JP08-255217 | 1996-09-27 | ||
| US08/778,621 US6104833A (en) | 1996-01-09 | 1997-01-03 | Pattern recognizing apparatus and method |
| US09/363,020 US6335986B1 (en) | 1996-01-09 | 1999-07-29 | Pattern recognizing apparatus and method |
| US10/000,525 US6850645B2 (en) | 1996-01-09 | 2001-12-04 | Pattern recognizing apparatus |
Related Parent Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US09/363,020 Division US6335986B1 (en) | 1996-01-09 | 1999-07-29 | Pattern recognizing apparatus and method |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20030113016A1 US20030113016A1 (en) | 2003-06-19 |
| US6850645B2 true US6850645B2 (en) | 2005-02-01 |
Family
ID=26335003
Family Applications (4)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US08/778,621 Expired - Lifetime US6104833A (en) | 1996-01-09 | 1997-01-03 | Pattern recognizing apparatus and method |
| US09/363,020 Expired - Lifetime US6335986B1 (en) | 1996-01-09 | 1999-07-29 | Pattern recognizing apparatus and method |
| US10/000,525 Expired - Fee Related US6850645B2 (en) | 1996-01-09 | 2001-12-04 | Pattern recognizing apparatus |
| US10/000,506 Expired - Fee Related US6687401B2 (en) | 1996-01-09 | 2001-12-04 | Pattern recognizing apparatus and method |
Family Applications Before (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US08/778,621 Expired - Lifetime US6104833A (en) | 1996-01-09 | 1997-01-03 | Pattern recognizing apparatus and method |
| US09/363,020 Expired - Lifetime US6335986B1 (en) | 1996-01-09 | 1999-07-29 | Pattern recognizing apparatus and method |
Family Applications After (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US10/000,506 Expired - Fee Related US6687401B2 (en) | 1996-01-09 | 2001-12-04 | Pattern recognizing apparatus and method |
Country Status (3)
| Country | Link |
|---|---|
| US (4) | US6104833A (en) |
| KR (1) | KR100248917B1 (en) |
| CN (1) | CN1156791C (en) |
Cited By (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20040161152A1 (en) * | 2001-06-15 | 2004-08-19 | Matteo Marconi | Automatic natural content detection in video information |
| US20060008122A1 (en) * | 2004-04-02 | 2006-01-12 | Kurzweil Raymond C | Image evaluation for reading mode in a reading machine |
| US20060075392A1 (en) * | 2004-10-05 | 2006-04-06 | International Business Machines Corporation | System and method for reverse engineering of pattern string validation scripts |
| US20060214707A1 (en) * | 2005-03-28 | 2006-09-28 | Tooru Ishizaki | Information processing apparatus and method |
| US7130487B1 (en) * | 1998-12-15 | 2006-10-31 | Matsushita Electric Industrial Co., Ltd. | Searching method, searching device, and recorded medium |
| US20080013833A1 (en) * | 2006-07-11 | 2008-01-17 | Alexander M Scot | Method and apparatus for identifying and/or removing combs from scanned images |
| US20080175483A1 (en) * | 2007-01-18 | 2008-07-24 | Fujitsu Limited | Pattern separating extraction program, pattern separating extraction device, and pattern separating extraction method |
| US20100115016A1 (en) * | 2008-11-04 | 2010-05-06 | Sean Miceli | Thresholding of Image Difference Maps |
| US20100215276A1 (en) * | 2009-02-25 | 2010-08-26 | Fujitsu Limited | Storage medium storing character recognition program, character recognition method, and character recognition apparatus |
| US7809195B1 (en) * | 2008-09-18 | 2010-10-05 | Ernest Greene | Encoding system providing discrimination, classification, and recognition of shapes and patterns |
| US20100284611A1 (en) * | 2009-05-07 | 2010-11-11 | Siliconfile Technologies Inc. | Image sensor and image sensing method for character recognition |
| US20110211759A1 (en) * | 2010-02-26 | 2011-09-01 | Samsung Electronics Co., Ltd. | Character recognition apparatus and method based on character orientation |
| US20120045129A1 (en) * | 2010-08-17 | 2012-02-23 | Fujitsu Limited | Document image processing method and apparatus |
| US20200311883A1 (en) * | 2019-04-01 | 2020-10-01 | Fuji Xerox Co., Ltd. | Image processing apparatus and non-transitory computer readable medium |
Families Citing this family (87)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US6519046B1 (en) * | 1997-03-17 | 2003-02-11 | Fuji Photo Film Co., Ltd. | Printing method and system for making a print from a photo picture frame and a graphic image written by a user |
| JPH1196301A (en) * | 1997-09-22 | 1999-04-09 | Hitachi Ltd | Character recognition device |
| JP3584458B2 (en) * | 1997-10-31 | 2004-11-04 | ソニー株式会社 | Pattern recognition device and pattern recognition method |
| CN1237476C (en) * | 1999-02-19 | 2006-01-18 | 株式会社理光 | Image detection method, image processing control method, image processing device |
| JP4412762B2 (en) * | 1999-05-31 | 2010-02-10 | 富士通株式会社 | Dotted line extraction method, apparatus and recording medium |
| US6563502B1 (en) | 1999-08-19 | 2003-05-13 | Adobe Systems Incorporated | Device dependent rendering |
| JP3604993B2 (en) * | 2000-03-16 | 2004-12-22 | シャープ株式会社 | Image encoding device, image encoding method, image decoding device, and image decoding method |
| WO2001093188A1 (en) * | 2000-06-02 | 2001-12-06 | Japan Science And Technology Corporation | Method for processing document, recorded medium on which document processing program is recorded and document processor |
| US7027649B1 (en) * | 2000-11-27 | 2006-04-11 | Intel Corporation | Computing the Euler number of a binary image |
| US7598955B1 (en) | 2000-12-15 | 2009-10-06 | Adobe Systems Incorporated | Hinted stem placement on high-resolution pixel grid |
| US6681060B2 (en) * | 2001-03-23 | 2004-01-20 | Intel Corporation | Image retrieval using distance measure |
| JP4846924B2 (en) * | 2001-05-31 | 2011-12-28 | キヤノン株式会社 | Pattern recognition device |
| JP4100885B2 (en) * | 2001-07-11 | 2008-06-11 | キヤノン株式会社 | Form recognition apparatus, method, program, and storage medium |
| JP3863775B2 (en) * | 2001-12-25 | 2006-12-27 | 株式会社九州エレクトロニクスシステム | Image information compression method, image information compression apparatus, and image information compression program |
| CA2375355A1 (en) * | 2002-03-11 | 2003-09-11 | Neo Systems Inc. | Character recognition system and method |
| US7016529B2 (en) * | 2002-03-15 | 2006-03-21 | Microsoft Corporation | System and method facilitating pattern recognition |
| JP2004145072A (en) * | 2002-10-25 | 2004-05-20 | Riso Kagaku Corp | Handwritten character font creation device and handwritten character font creation program |
| US7227993B2 (en) * | 2003-01-27 | 2007-06-05 | Microsoft Corporation | Learning-based system and process for synthesizing cursive handwriting |
| US20040146200A1 (en) * | 2003-01-29 | 2004-07-29 | Lockheed Martin Corporation | Segmenting touching characters in an optical character recognition system to provide multiple segmentations |
| US20040151377A1 (en) * | 2003-02-04 | 2004-08-05 | Boose Molly L. | Apparatus and methods for converting network drawings from raster format to vector format |
| US7624277B1 (en) * | 2003-02-25 | 2009-11-24 | Microsoft Corporation | Content alteration for prevention of unauthorized scripts |
| US7606915B1 (en) | 2003-02-25 | 2009-10-20 | Microsoft Corporation | Prevention of unauthorized scripts |
| JP4144395B2 (en) * | 2003-03-25 | 2008-09-03 | ブラザー工業株式会社 | Image processing apparatus and image forming apparatus having the same |
| US7006107B2 (en) * | 2003-05-16 | 2006-02-28 | Adobe Systems Incorporated | Anisotropic anti-aliasing |
| US7002597B2 (en) * | 2003-05-16 | 2006-02-21 | Adobe Systems Incorporated | Dynamic selection of anti-aliasing procedures |
| JP2004348591A (en) * | 2003-05-23 | 2004-12-09 | Canon Inc | Document search method and apparatus |
| CN1303563C (en) * | 2003-09-29 | 2007-03-07 | 摩托罗拉公司 | Method and system for compressing hand-written character template |
| CN100433050C (en) * | 2004-01-08 | 2008-11-12 | 日本电气株式会社 | Mobile communication system, mobile terminal and fixed station equipment, character recognition equipment, method and program |
| US7602995B2 (en) * | 2004-02-10 | 2009-10-13 | Ricoh Company, Ltd. | Correcting image distortion caused by scanning |
| JP3861157B2 (en) | 2004-02-27 | 2006-12-20 | 国立大学法人広島大学 | Reference data optimization device and pattern recognition system |
| JP2005251115A (en) * | 2004-03-08 | 2005-09-15 | Shogakukan Inc | Associative search system and associative search method |
| US7580039B2 (en) * | 2004-03-31 | 2009-08-25 | Adobe Systems Incorporated | Glyph outline adjustment while rendering |
| US7333110B2 (en) * | 2004-03-31 | 2008-02-19 | Adobe Systems Incorporated | Adjusted stroke rendering |
| US7719536B2 (en) * | 2004-03-31 | 2010-05-18 | Adobe Systems Incorporated | Glyph adjustment in high resolution raster while rendering |
| US7602390B2 (en) * | 2004-03-31 | 2009-10-13 | Adobe Systems Incorporated | Edge detection based stroke adjustment |
| US7639258B1 (en) | 2004-03-31 | 2009-12-29 | Adobe Systems Incorporated | Winding order test for digital fonts |
| JP4398777B2 (en) * | 2004-04-28 | 2010-01-13 | 株式会社東芝 | Time series data analysis apparatus and method |
| US7561738B2 (en) * | 2004-09-22 | 2009-07-14 | Microsoft Corporation | Symbol grouping and recognition in expression recognition |
| US7724981B2 (en) * | 2005-07-21 | 2010-05-25 | Ancestry.Com Operations Inc. | Adaptive contrast control systems and methods |
| JP4920928B2 (en) * | 2005-08-08 | 2012-04-18 | キヤノン株式会社 | Image processing apparatus, control method therefor, and program |
| US20070127824A1 (en) * | 2005-12-07 | 2007-06-07 | Trw Automotive U.S. Llc | Method and apparatus for classifying a vehicle occupant via a non-parametric learning algorithm |
| US7583841B2 (en) * | 2005-12-21 | 2009-09-01 | Microsoft Corporation | Table detection in ink notes |
| US7664325B2 (en) * | 2005-12-21 | 2010-02-16 | Microsoft Corporation | Framework for detecting a structured handwritten object |
| US7734092B2 (en) * | 2006-03-07 | 2010-06-08 | Ancestry.Com Operations Inc. | Multiple image input for optical character recognition processing systems and methods |
| US7873215B2 (en) * | 2007-06-27 | 2011-01-18 | Seiko Epson Corporation | Precise identification of text pixels from scanned document images |
| JP2009070226A (en) * | 2007-09-14 | 2009-04-02 | Ricoh Co Ltd | Matching device |
| RU2436156C1 (en) * | 2007-11-20 | 2011-12-10 | Люмэкс Ас | Method of resolving conflicting output data from optical character recognition system (ocr), where output data include more than one character image recognition alternative |
| US8634645B2 (en) * | 2008-03-28 | 2014-01-21 | Smart Technologies Ulc | Method and tool for recognizing a hand-drawn table |
| US7471826B1 (en) | 2008-03-31 | 2008-12-30 | International Business Machines Corporation | Character segmentation by slices |
| JP5586917B2 (en) * | 2009-10-27 | 2014-09-10 | キヤノン株式会社 | Information processing apparatus, information processing method, and program |
| WO2011079432A1 (en) * | 2009-12-29 | 2011-07-07 | Nokia Corporation | Method and apparatus for generating a text image |
| JP5591578B2 (en) * | 2010-04-19 | 2014-09-17 | 日本電産サンキョー株式会社 | Character string recognition apparatus and character string recognition method |
| US9285983B2 (en) * | 2010-06-14 | 2016-03-15 | Amx Llc | Gesture recognition using neural networks |
| US8737702B2 (en) * | 2010-07-23 | 2014-05-27 | International Business Machines Corporation | Systems and methods for automated extraction of measurement information in medical videos |
| US8947736B2 (en) * | 2010-11-15 | 2015-02-03 | Konica Minolta Laboratory U.S.A., Inc. | Method for binarizing scanned document images containing gray or light colored text printed with halftone pattern |
| US9319556B2 (en) | 2011-08-31 | 2016-04-19 | Konica Minolta Laboratory U.S.A., Inc. | Method and apparatus for authenticating printed documents that contains both dark and halftone text |
| JP6008076B2 (en) * | 2011-10-13 | 2016-10-19 | セイコーエプソン株式会社 | Projector and image drawing method |
| KR102073024B1 (en) * | 2012-11-30 | 2020-02-04 | 삼성전자주식회사 | Apparatus and method for editing memo in a user terminal |
| JP2014127188A (en) * | 2012-12-27 | 2014-07-07 | Toshiba Corp | Shaping device and method |
| JP2014215877A (en) * | 2013-04-26 | 2014-11-17 | 株式会社デンソー | Object detection device |
| KR101532652B1 (en) * | 2013-05-22 | 2015-06-30 | 재단법인대구경북과학기술원 | Image Recognition Calculating Apparatus and the Method |
| EP2818948B1 (en) * | 2013-06-27 | 2016-11-16 | ABB Schweiz AG | Method and data presenting device for assisting a remote user to provide instructions |
| KR101526490B1 (en) * | 2013-10-25 | 2015-06-08 | 경희대학교 산학협력단 | Visual data processing apparatus and method for Efficient resource management in Cloud Computing |
| KR20150049700A (en) * | 2013-10-30 | 2015-05-08 | 삼성전자주식회사 | Method and apparautus for controlling input in portable device |
| US9928572B1 (en) | 2013-12-20 | 2018-03-27 | Amazon Technologies, Inc. | Label orientation |
| CN104978782B (en) * | 2014-04-11 | 2017-09-19 | 南京视察者图像识别科技有限公司 | Passenger flow analysing method based on touch screen technology |
| JP6045752B2 (en) * | 2014-05-14 | 2016-12-14 | 共同印刷株式会社 | Two-dimensional code, two-dimensional code analysis system, and two-dimensional code creation system |
| CN104156694B (en) * | 2014-07-18 | 2019-03-19 | 百度在线网络技术(北京)有限公司 | A kind of method and apparatus of target object in image for identification |
| JP2016153933A (en) * | 2015-02-20 | 2016-08-25 | 株式会社リコー | Image processor, image processing system, image processing method, program, and recording medium |
| JP6390501B2 (en) * | 2015-04-15 | 2018-09-19 | 京セラドキュメントソリューションズ株式会社 | Learning support apparatus and learning support method |
| JP6705237B2 (en) * | 2016-03-18 | 2020-06-03 | 株式会社リコー | Image compression apparatus, image forming apparatus, image compression method and program |
| JP2017215807A (en) * | 2016-05-31 | 2017-12-07 | 富士ゼロックス株式会社 | Program and information processing apparatus |
| US10372980B2 (en) * | 2016-11-16 | 2019-08-06 | Switch, Inc. | Electronic form identification using spatial information |
| JP2018112839A (en) * | 2017-01-10 | 2018-07-19 | 富士通株式会社 | Image processing program, image recognition program, image processing apparatus, image recognition apparatus, image recognition method, and image processing method |
| US11004205B2 (en) | 2017-04-18 | 2021-05-11 | Texas Instruments Incorporated | Hardware accelerator for histogram of oriented gradients computation |
| JP2019139592A (en) * | 2018-02-13 | 2019-08-22 | シャープ株式会社 | Character recognition device and character recognition method |
| US10565443B2 (en) * | 2018-02-16 | 2020-02-18 | Wipro Limited | Method and system for determining structural blocks of a document |
| GB2571530B (en) * | 2018-02-28 | 2020-09-23 | Canon Europa Nv | An image processing method and an image processing system |
| US12190570B2 (en) | 2019-07-11 | 2025-01-07 | Nec Corporation | Object detection device, learning method, and recording medium |
| WO2021054850A1 (en) * | 2019-09-17 | 2021-03-25 | Публичное Акционерное Общество "Сбербанк России" | Method and system for intelligent document processing |
| US12112518B2 (en) | 2019-11-08 | 2024-10-08 | Nec Corporation | Object detection device, learning method, and recording medium |
| CN110968667B (en) * | 2019-11-27 | 2023-04-18 | 广西大学 | Periodical and literature table extraction method based on text state characteristics |
| JP7484198B2 (en) * | 2020-02-03 | 2024-05-16 | 富士フイルムビジネスイノベーション株式会社 | Document processing device and program |
| WO2022064610A1 (en) | 2020-09-24 | 2022-03-31 | 日本電気株式会社 | Object detection device, trained model generation method, and recording medium |
| CN114495108B (en) * | 2021-12-29 | 2025-04-25 | 南京中科创达软件科技有限公司 | Character detection method, device, electronic device and readable medium |
| US11961316B2 (en) * | 2022-05-10 | 2024-04-16 | Capital One Services, Llc | Text extraction using optical character recognition |
| CN118567950B (en) * | 2024-06-04 | 2025-01-28 | 四川云互未来科技有限公司 | A server hosting operation environment monitoring method and system |
Citations (17)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS6057477A (en) | 1983-09-08 | 1985-04-03 | Fuji Xerox Co Ltd | Pattern recognizer |
| US5197107A (en) | 1988-04-26 | 1993-03-23 | Fujitsu Limited | Character recognition apparatus |
| US5253303A (en) | 1988-02-15 | 1993-10-12 | Nec Corporation | Character recognizing method and apparatus thereof |
| JPH06309501A (en) | 1993-04-20 | 1994-11-04 | Fujitsu Ltd | Character recognizing device |
| JPH06309498A (en) | 1993-02-25 | 1994-11-04 | Fujitsu Ltd | Image extraction method |
| JPH0773270A (en) | 1993-09-03 | 1995-03-17 | Oki Electric Ind Co Ltd | Optical character reading system |
| US5481621A (en) | 1992-05-28 | 1996-01-02 | Matsushita Electric Industrial Co., Ltd. | Device and method for recognizing an image based on a feature indicating a relative positional relationship between patterns |
| JPH0852219A (en) | 1994-02-24 | 1996-02-27 | Cardiovascular Dynamics Inc | Baloon catheter,multiple zone baloon catheter and method by using those |
| US5524065A (en) * | 1992-02-07 | 1996-06-04 | Canon Kabushiki Kaisha | Method and apparatus for pattern recognition |
| JPH08171609A (en) | 1994-12-16 | 1996-07-02 | Fujitsu Ltd | High-speed character string extraction device |
| US5581633A (en) | 1993-06-11 | 1996-12-03 | Fujitsu Limited | Method and apparatus for segmenting a character and for extracting a character string based on a histogram |
| US5621818A (en) | 1991-07-10 | 1997-04-15 | Fuji Xerox Co., Ltd. | Document recognition apparatus |
| US5684891A (en) | 1991-10-21 | 1997-11-04 | Canon Kabushiki Kaisha | Method and apparatus for character recognition |
| US5745599A (en) | 1994-01-19 | 1998-04-28 | Nippon Telegraph And Telephone Corporation | Character recognition method |
| US5754686A (en) | 1994-02-10 | 1998-05-19 | Canon Kabushiki Kaisha | Method of registering a character pattern into a user dictionary and a character recognition apparatus having the user dictionary |
| US5949906A (en) * | 1994-12-08 | 1999-09-07 | Kabushiki Kaisha Toshiba | Apparatus and method for extracting character string |
| US6512848B2 (en) * | 1996-11-18 | 2003-01-28 | Canon Kabushiki Kaisha | Page analysis system |
Family Cites Families (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CA2078423C (en) * | 1991-11-19 | 1997-01-14 | Per-Kristian Halvorsen | Method and apparatus for supplementing significant portions of a document selected without document image decoding with retrieved information |
| US5761340A (en) * | 1993-04-28 | 1998-06-02 | Casio Computer Co., Ltd. | Data editing method and system for a pen type input device |
| JPH07160827A (en) * | 1993-12-09 | 1995-06-23 | Matsushita Electric Ind Co Ltd | Handwriting stroke editing apparatus and method |
| JPH08221576A (en) * | 1994-12-12 | 1996-08-30 | Toshiba Corp | Straight line detection method in character string, straight line removal method and address area discrimination device |
| JP2817646B2 (en) * | 1995-02-01 | 1998-10-30 | 日本電気株式会社 | Document editing device |
-
1997
- 1997-01-03 US US08/778,621 patent/US6104833A/en not_active Expired - Lifetime
- 1997-01-08 CN CNB971020116A patent/CN1156791C/en not_active Expired - Fee Related
- 1997-01-09 KR KR1019970000322A patent/KR100248917B1/en not_active Expired - Fee Related
-
1999
- 1999-07-29 US US09/363,020 patent/US6335986B1/en not_active Expired - Lifetime
-
2001
- 2001-12-04 US US10/000,525 patent/US6850645B2/en not_active Expired - Fee Related
- 2001-12-04 US US10/000,506 patent/US6687401B2/en not_active Expired - Fee Related
Patent Citations (18)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS6057477A (en) | 1983-09-08 | 1985-04-03 | Fuji Xerox Co Ltd | Pattern recognizer |
| US5253303A (en) | 1988-02-15 | 1993-10-12 | Nec Corporation | Character recognizing method and apparatus thereof |
| US5197107A (en) | 1988-04-26 | 1993-03-23 | Fujitsu Limited | Character recognition apparatus |
| US5621818A (en) | 1991-07-10 | 1997-04-15 | Fuji Xerox Co., Ltd. | Document recognition apparatus |
| US5684891A (en) | 1991-10-21 | 1997-11-04 | Canon Kabushiki Kaisha | Method and apparatus for character recognition |
| US5524065A (en) * | 1992-02-07 | 1996-06-04 | Canon Kabushiki Kaisha | Method and apparatus for pattern recognition |
| US5481621A (en) | 1992-05-28 | 1996-01-02 | Matsushita Electric Industrial Co., Ltd. | Device and method for recognizing an image based on a feature indicating a relative positional relationship between patterns |
| US5793932A (en) | 1992-05-28 | 1998-08-11 | Matsushita Electric Industrial Co., Ltd. | Image recognition device and an image recognition method |
| JPH06309498A (en) | 1993-02-25 | 1994-11-04 | Fujitsu Ltd | Image extraction method |
| JPH06309501A (en) | 1993-04-20 | 1994-11-04 | Fujitsu Ltd | Character recognizing device |
| US5581633A (en) | 1993-06-11 | 1996-12-03 | Fujitsu Limited | Method and apparatus for segmenting a character and for extracting a character string based on a histogram |
| JPH0773270A (en) | 1993-09-03 | 1995-03-17 | Oki Electric Ind Co Ltd | Optical character reading system |
| US5745599A (en) | 1994-01-19 | 1998-04-28 | Nippon Telegraph And Telephone Corporation | Character recognition method |
| US5754686A (en) | 1994-02-10 | 1998-05-19 | Canon Kabushiki Kaisha | Method of registering a character pattern into a user dictionary and a character recognition apparatus having the user dictionary |
| JPH0852219A (en) | 1994-02-24 | 1996-02-27 | Cardiovascular Dynamics Inc | Baloon catheter,multiple zone baloon catheter and method by using those |
| US5949906A (en) * | 1994-12-08 | 1999-09-07 | Kabushiki Kaisha Toshiba | Apparatus and method for extracting character string |
| JPH08171609A (en) | 1994-12-16 | 1996-07-02 | Fujitsu Ltd | High-speed character string extraction device |
| US6512848B2 (en) * | 1996-11-18 | 2003-01-28 | Canon Kabushiki Kaisha | Page analysis system |
Non-Patent Citations (3)
| Title |
|---|
| An Japanese and English language copy of Japanese Office Action. |
| U.S. patent application Ser. No. 08/778,621, Naoi, et al., filed Jan. 1997. |
| U.S. patent application Ser. No. 09/363,020, Naoi et al., filed Jul. 29, 1999. |
Cited By (23)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7130487B1 (en) * | 1998-12-15 | 2006-10-31 | Matsushita Electric Industrial Co., Ltd. | Searching method, searching device, and recorded medium |
| US20040161152A1 (en) * | 2001-06-15 | 2004-08-19 | Matteo Marconi | Automatic natural content detection in video information |
| US20060008122A1 (en) * | 2004-04-02 | 2006-01-12 | Kurzweil Raymond C | Image evaluation for reading mode in a reading machine |
| US8249309B2 (en) * | 2004-04-02 | 2012-08-21 | K-Nfb Reading Technology, Inc. | Image evaluation for reading mode in a reading machine |
| US20060075392A1 (en) * | 2004-10-05 | 2006-04-06 | International Business Machines Corporation | System and method for reverse engineering of pattern string validation scripts |
| US7793207B2 (en) * | 2005-03-28 | 2010-09-07 | Canon Kabushiki Kaisha | Converting text data into binary data using external link information |
| US20060214707A1 (en) * | 2005-03-28 | 2006-09-28 | Tooru Ishizaki | Information processing apparatus and method |
| US20080013833A1 (en) * | 2006-07-11 | 2008-01-17 | Alexander M Scot | Method and apparatus for identifying and/or removing combs from scanned images |
| US7756340B2 (en) * | 2006-07-11 | 2010-07-13 | Pegasus Imaging Corporation | Method and apparatus for identifying and/or removing combs from scanned images |
| US8615129B2 (en) * | 2007-01-18 | 2013-12-24 | Fujitsu Limited | Pattern separating extraction device, and pattern separating extraction method |
| US20080175483A1 (en) * | 2007-01-18 | 2008-07-24 | Fujitsu Limited | Pattern separating extraction program, pattern separating extraction device, and pattern separating extraction method |
| US7809195B1 (en) * | 2008-09-18 | 2010-10-05 | Ernest Greene | Encoding system providing discrimination, classification, and recognition of shapes and patterns |
| US20100115016A1 (en) * | 2008-11-04 | 2010-05-06 | Sean Miceli | Thresholding of Image Difference Maps |
| US8239435B2 (en) * | 2008-11-04 | 2012-08-07 | Seiko Epson Corporation | Thresholding of image diffences maps using first and second two-dimenstional array wherein respective euler number is determined |
| US20100215276A1 (en) * | 2009-02-25 | 2010-08-26 | Fujitsu Limited | Storage medium storing character recognition program, character recognition method, and character recognition apparatus |
| US8401298B2 (en) * | 2009-02-25 | 2013-03-19 | Fujitsu Limited | Storage medium storing character recognition program, character recognition method, and character recognition apparatus |
| US20100284611A1 (en) * | 2009-05-07 | 2010-11-11 | Siliconfile Technologies Inc. | Image sensor and image sensing method for character recognition |
| US20110211759A1 (en) * | 2010-02-26 | 2011-09-01 | Samsung Electronics Co., Ltd. | Character recognition apparatus and method based on character orientation |
| US8761514B2 (en) * | 2010-02-26 | 2014-06-24 | Samsung Electronics Co., Ltd | Character recognition apparatus and method based on character orientation |
| US20120045129A1 (en) * | 2010-08-17 | 2012-02-23 | Fujitsu Limited | Document image processing method and apparatus |
| US8345977B2 (en) * | 2010-08-17 | 2013-01-01 | Fujitsu Limited | Document image processing method and apparatus |
| US20200311883A1 (en) * | 2019-04-01 | 2020-10-01 | Fuji Xerox Co., Ltd. | Image processing apparatus and non-transitory computer readable medium |
| US11715018B2 (en) * | 2019-04-01 | 2023-08-01 | Fujifilm Business Innovation Corp. | Image processing apparatus and non-transitory computer readable medium |
Also Published As
| Publication number | Publication date |
|---|---|
| KR100248917B1 (en) | 2000-03-15 |
| CN1162795A (en) | 1997-10-22 |
| CN1156791C (en) | 2004-07-07 |
| US20030113016A1 (en) | 2003-06-19 |
| US6687401B2 (en) | 2004-02-03 |
| US20020061135A1 (en) | 2002-05-23 |
| KR19980023917A (en) | 1998-07-06 |
| US6104833A (en) | 2000-08-15 |
| US6335986B1 (en) | 2002-01-01 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US6850645B2 (en) | Pattern recognizing apparatus | |
| US6151423A (en) | Character recognition with document orientation determination | |
| KR100324847B1 (en) | Address reader and mails separater, and character string recognition method | |
| KR100249055B1 (en) | Character recognition apparatus | |
| US5995659A (en) | Method of searching and extracting text information from drawings | |
| US7233697B2 (en) | Character recognition device and a method therefor | |
| US5325447A (en) | Handwritten digit normalization method | |
| US20050238257A1 (en) | Form search apparatus and method | |
| JP3345224B2 (en) | Pattern extraction device, pattern re-recognition table creation device, and pattern recognition device | |
| JP3485020B2 (en) | Character recognition method and apparatus, and storage medium | |
| Lehal et al. | Feature extraction and classification for OCR of Gurmukhi script | |
| US20050123199A1 (en) | Method for optical recognition of a multi-language set of letters with diacritics | |
| US5926564A (en) | Character recognition method and apparatus based on 0-1 pattern representation of histogram of character image | |
| US6256408B1 (en) | Speed and recognition enhancement for OCR using normalized height/width position | |
| US6549662B1 (en) | Method of recognizing characters | |
| US5272766A (en) | OCR system for recognizing user-specified custom fonts in addition to standard fonts using three-layer templates | |
| JP3917349B2 (en) | Retrieval device and method for retrieving information using character recognition result | |
| JPH09319824A (en) | Form recognition method | |
| JP2007058882A (en) | Pattern recognition device | |
| JP4176175B2 (en) | Pattern recognition device | |
| US20200242389A1 (en) | Information processing apparatus and non-transitory computer readable medium | |
| JP3372005B2 (en) | Character recognition device | |
| JP2917427B2 (en) | Drawing reader | |
| JP4810853B2 (en) | Character image cutting device, character image cutting method and program | |
| Amin | Recognition of printed Arabic text using machine learning |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Free format text: PAYER NUMBER DE-ASSIGNED (ORIGINAL EVENT CODE: RMPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| FPAY | Fee payment |
Year of fee payment: 4 |
|
| FPAY | Fee payment |
Year of fee payment: 8 |
|
| REMI | Maintenance fee reminder mailed | ||
| LAPS | Lapse for failure to pay maintenance fees | ||
| STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
| STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
| FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20170201 |