JPS61156484A - Picture image processor - Google Patents

Picture image processor

Info

Publication number
JPS61156484A
JPS61156484A JP59276463A JP27646384A JPS61156484A JP S61156484 A JPS61156484 A JP S61156484A JP 59276463 A JP59276463 A JP 59276463A JP 27646384 A JP27646384 A JP 27646384A JP S61156484 A JPS61156484 A JP S61156484A
Authority
JP
Japan
Prior art keywords
character
picture image
picture
image data
distribution
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
JP59276463A
Other languages
Japanese (ja)
Inventor
Kenichi Oota
健一 太田
Masatoshi Okutomi
正敏 奥富
Tetsuo Sueda
末田 哲夫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Canon Inc
Original Assignee
Canon Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Canon Inc filed Critical Canon Inc
Priority to JP59276463A priority Critical patent/JPS61156484A/en
Publication of JPS61156484A publication Critical patent/JPS61156484A/en
Pending legal-status Critical Current

Links

Landscapes

  • Character Input (AREA)

Abstract

PURPOSE:To carry out appropriate binarization processing by carrying out pre-processing for recognizing information read by reading means with the aid of multivalued digital data. CONSTITUTION:Picture images in a character frame position detection area are taken out sequentially out of a multivalued picture image data 31 which is an output data from an A/D converter 23, and transferred to adders 32 and 35. When picture taking out is of a luster scan system in an x direction, density of a y direction projection distribution is available by adding density distribution in one line, and addition completes by reciprocating scanning between the adder 32 and a latch 33, thereby storing said distribution into a memory 34. As for the x direction, scanning reciprocates between the adder 35 and a memory 36. A threshold value deciding circuit 39 decides a threshold value from a detected frame position and the picture image data 31, and send said value to a comparator 38, thereby obtaining a secondary value picture image 30 from the picture image data 31 and the comparator 38. Thus, a character area to be recognized is extracted, and the appropriate binarization processing with respect to a character image can be carried out.

Description

【発明の詳細な説明】 〔技術分計〕 本発明は画像処理装置特に光学的文字認識装置における
前処理装置特に文字枠とその内部に書き込まれた文字、
記号とを識別し、文字、記号のみを抽出する文字抽出装
置に関するものである。
[Detailed Description of the Invention] [Technical Summary] The present invention relates to a pre-processing device for an image processing device, particularly an optical character recognition device, particularly a character frame and characters written inside the character frame.
The present invention relates to a character extraction device that identifies characters and symbols and extracts only characters and symbols.

〔従来技術〕[Prior art]

従来光学的文字読み取り装置において、読み取るべき帳
票上の文字枠は、読み取り装置の分光感度では帳票下地
と区別できないような色の、いわゆるドロップアウトカ
ラーで印刷することにより、文字部のみを検出するとい
った方法が一般的であった。しかしこの方法においては
、文字枠に上記のような制約があるため、一般的な筆記
用具である鉛筆、ボールペンやサインペン或いはPPC
等により、文字枠を作成し、帳票を自家製造することは
不可能である。また、文字枠にドロップアウトカラーを
用いない場合としては、文字及び文字枠を2値画像に変
換した後に適当なフィルタリングを行うことにより、文
字枠を検出する方法も提案されている。(特開昭59−
77577号公報)しかし、この方法では文字及び文字
枠の両方に対して一定値の閾値による2値化を行うため
、第1図のように、文字と文字枠の濃度差が大きい場合
や文字ごとの濃度のバラツキがある場合には、適正な2
値化閾値がうまく設定できず、文字の切れやつぶれ等が
生じやすいという欠点があった。
In conventional optical character reading devices, the character frame on the form to be read is printed in a so-called dropout color, which is a color that cannot be distinguished from the form base using the spectral sensitivity of the reading device, so that only the text is detected. The method was common. However, in this method, due to the above-mentioned restrictions on the character frame, general writing instruments such as pencils, ballpoint pens, felt-tip pens, or PPC
etc., it is impossible to create character frames and create forms in-house. Furthermore, in the case where a dropout color is not used for the character frame, a method has also been proposed in which the character frame is detected by converting the characters and the character frame into a binary image and then performing appropriate filtering. (Unexamined Japanese Patent Publication No. 59-
(No. 77577) However, since this method performs binarization using a fixed threshold value for both characters and character frames, as shown in Figure 1, when there is a large density difference between characters and character frames, or when each character If there are variations in the concentration of
There was a drawback that the digitization threshold could not be set well, and characters were likely to be cut off or blurred.

〔目的〕〔the purpose〕

以上の点に鑑み、本願発明の目的は文字枠等にドロップ
アウトカラーを使用するといった制約なしで、認識すべ
き文字領域を抽出し、該文字に対して適正な2値化処理
を行うことが可能な画像処理装置を提供することにある
In view of the above points, an object of the present invention is to extract a character area to be recognized and perform appropriate binarization processing on the character without the restriction of using dropout colors for character frames etc. The purpose of the present invention is to provide a capable image processing device.

又、2値化する以前に、前処理段階で、確実に認識対象
物とノイズ、文字枠との識別が可能な画像処理装置を提
供することを目的としている。
Another object of the present invention is to provide an image processing apparatus that can reliably distinguish between a recognition target object, noise, and character frames in a preprocessing stage before binarization.

〔実施例〕〔Example〕

次に、図面を参照して本発明の詳細な説明する。第2図
に本発明の適用が可能のシステムの1例を示す。1はリ
ーダであり、CCD等で原稿を読み取り、該読み取った
画像データを編集処理し、プリンタ3に出力したり、外
部機器へネットワークを介して、伝達したりすることが
可能である。なお、2はOCR及び画像処理の制御部で
ある。又、4は、ディスプレイを有したコンピュータで
、5のキーボードからの入力により、機器の制御指令等
を行うものである。
Next, the present invention will be described in detail with reference to the drawings. FIG. 2 shows an example of a system to which the present invention can be applied. Reference numeral 1 denotes a reader, which is capable of reading a document with a CCD or the like, editing the read image data, outputting it to the printer 3, or transmitting it to an external device via a network. Note that 2 is a control unit for OCR and image processing. Further, numeral 4 is a computer having a display, and is used to issue control commands for equipment through input from the keyboard 5.

なお本願の適用が可能な機器は他に1元ディスク等のフ
ァイル、マイクロフィルム、レーザビームプリンタ、各
機器を制御するワークステーションを有したシステムの
中の1つであってもよい。
Note that the device to which the present application can be applied may also be one of the systems having a file such as a single-source disk, a microfilm, a laser beam printer, and a workstation for controlling each device.

第3図に本発明の適用が可能な機器のブロック図を示す
。読み取られるべき原稿21は光源で照明され、充電変
換素子22上に結像され、ラスタスキャンにより、時系
列の電気信号に変換され、23のA−D変換器により多
値のデジタル信号となり、24.25の走査方向及び副
走査方向の加算器及び相関器へ別々に送られる。
FIG. 3 shows a block diagram of equipment to which the present invention can be applied. The original 21 to be read is illuminated with a light source, imaged on the charging conversion element 22, converted into a time-series electric signal by raster scanning, converted into a multi-value digital signal by an A-D converter 23, and converted into a multi-value digital signal by an A-D converter 24. .25 scan direction and sub-scan direction adders and correlators separately.

相関器25からは文字領域情報が26の2値化回路へ送
られ、画像の2値化、つづいて27の認識回路で、認識
が行われる。本発明は、上述の2値化を行う直前までの
前処理に係るものである。
The character area information is sent from the correlator 25 to a binarization circuit 26, where the image is binarized and then recognized by a recognition circuit 27. The present invention relates to preprocessing immediately before performing the above-mentioned binarization.

第4図に、第3図に示した加算器24、相関器25につ
いて、更に詳細に説明する。なお、第5図にその手順を
示す。
In FIG. 4, the adder 24 and correlator 25 shown in FIG. 3 will be explained in more detail. The procedure is shown in FIG.

第3図に示したA/D変換器23からの出力データであ
る多値画像データ31から後述の第6図に示すように、
文字枠位置探査領域の画像を順次とり出し、32.35
の加算器へ転送する。画像取り出しがX方向のラスター
スキャン方式である場合を考えると、y方向投影投炭分
布は、1ライン中の濃度分布を加算して行えばよいので
、32の加算器と33のラッチの間を往復することによ
り加算が完了し、それを34のメモリに収容すればよい
。X方向については、1ラインスキヤンの間のデータを
それぞれ別々のメモリに収容して、ラインが変わる毎に
、それをよみ出して、データを加算することになるので
、35.36の間を往復することKなる。
As shown in FIG. 6, which will be described later, from the multivalued image data 31 which is the output data from the A/D converter 23 shown in FIG.
Sequentially extract images of the character frame position detection area, 32.35
transfer to the adder. Considering the case where image retrieval is performed using a raster scan method in the The addition is completed by going back and forth, and it is sufficient to store it in 34 memories. Regarding the X direction, the data for one line scan is stored in separate memories, and each time the line changes, it is read out and the data is added, so it is necessary to go back and forth between 35.36 It becomes K to do.

以上により得られた投影濃度分布は、37のマイクロプ
ロセッサで相関演算を施こされ、枠位置が検出される。
The projected density distribution obtained in the above manner is subjected to correlation calculation by 37 microprocessors, and the frame position is detected.

この部分については後述する。This part will be described later.

検出された枠位置と31の画像データから、上記とは独
立に設けられた39の1闇値決定回路が闇値を決定し、
38の比較器に送りこまれ、31と38から2値画像3
0が得られる。
Based on the detected frame position and 31 image data, 39 1 darkness value determination circuits provided independently from the above determine the darkness value,
It is sent to the comparator of 38, and binary image 3 is generated from 31 and 38.
0 is obtained.

第5図には以上の手順が示されており、X方向について
のみ説明する。
The above procedure is shown in FIG. 5, and only the X direction will be explained.

まず、枠探査領域が指定されると(ステップ2)、x方
向投影濃度分布を作成しくステップ3)、次に、濃度分
布のピークを検出するため、ステップ4で、初期化し、
ステップ5において、tをパラメータとして増加させ、
ステップ6でマスク演算を行う(後述の相関演算)。そ
して、ステップ7.8において、最大値を求め、枠を検
出する。
First, when the frame search area is specified (step 2), an x-direction projected density distribution is created (step 3), and then, in order to detect the peak of the density distribution, in step 4, initialization is performed.
In step 5, increase t as a parameter,
In step 6, a mask calculation is performed (correlation calculation to be described later). Then, in step 7.8, the maximum value is determined and a frame is detected.

更に第3図について、詳細に説明する。Further, FIG. 3 will be explained in detail.

読み取るべき原稿21は第6図の如く、あらかじめ、相
関器25に大きさ及び線幅が記憶された複数の文字枠3
1内に、文字、記号32が書きこまれているようなもの
であるとする。A−D変換器23からの多値デジタル信
号は24の加算器で順次加算されX軸方向及びy軸方向
への投影濃度分布を作成する。第7図に投影濃度分布の
一例を示す。濃度分布は、文字枠の少くとも1つが入り
2つ以上は入らないような領域、例えば第7図の点@4
1で囲まれた領域についてとるようにする。第7図の曲
I!42がX軸投影濃度分布、43がy軸投影濃度分布
を示して〜する。第7図中の長さ44.45は、あらか
じめ既知の文字枠のX方向、y方向の大きさとほぼ等し
くなっている。第3図25の相関器は上記投影濃度分布
と、与えられたマスクパターンとの相関演算を行い、文
字枠位置を検出するもので、以下にこの相関器の作用(
第4図のMPU内)について説明する。
As shown in FIG. 6, the original 21 to be read has a plurality of character frames 3 whose sizes and line widths are stored in the correlator 25 in advance.
Assume that a character or symbol 32 is written in 1. The multivalued digital signals from the A-D converter 23 are sequentially added by 24 adders to create projected density distributions in the X-axis direction and the y-axis direction. FIG. 7 shows an example of the projected density distribution. The density distribution is a region where at least one character frame is included and two or more are not included, for example, point @4 in Figure 7.
Let us take the area surrounded by 1. Song I in Figure 7! Reference numeral 42 indicates an X-axis projected density distribution, and 43 indicates a y-axis projected density distribution. The length 44.45 in FIG. 7 is approximately equal to the previously known size of the character frame in the X and Y directions. The correlator shown in FIG. 325 detects the character frame position by calculating the correlation between the projected density distribution and the given mask pattern.The function of this correlator (
(inside the MPU in FIG. 4) will be explained.

相関とは、この場合、ある−次元の関数fと、与えられ
たマスクパターンgの檀を有限区間にわたって積分し、
gを少しずつずらしながら積分値の最大値を探し出すと
いった操作を指す。
In this case, correlation is the integration of a certain -dimensional function f and a given mask pattern g over a finite interval,
Refers to the operation of finding the maximum value of the integral value while shifting g little by little.

X方向の枠位置を検出するには、fはX方向投影濃度分
布f(x)、gは例えば、第8図に示すようなマスクパ
ターンであり、点jを原点にとれば、 となる。ここで2つの凸部の間隔ωと凸部の大きさhを
X方向の文字枠の大きさ及び、文字枠の線幅にほぼ等し
くとっておけばfとgの相関値は第7図のt=aにおい
てまず最大となる。従って、この最大値を検出すれば枠
位置が求められることになる。y方向についても同様の
手続きをすれば、同様にして第7図す点が求められるの
で2次元的な枠位置A点が検出されることになる。又、
枠の長さがわかっているので、B点も容易に検出できる
To detect the frame position in the X direction, f is the X direction projected density distribution f(x), g is a mask pattern as shown in FIG. 8, for example, and if point j is taken as the origin, then the following equation is obtained. Here, if the interval ω between the two convex parts and the size h of the convex parts are set approximately equal to the size of the character frame in the X direction and the line width of the character frame, the correlation value between f and g will be as shown in Figure 7. It first becomes maximum at t=a. Therefore, by detecting this maximum value, the frame position can be determined. If the same procedure is carried out in the y direction, the points shown in FIG. 7 can be found in the same way, so that the two-dimensional frame position point A can be detected. or,
Since the length of the frame is known, point B can also be easily detected.

次に文字枠を抽出した後、文字領域の抽出方法について
説明する。
Next, a method for extracting a character area after extracting a character frame will be described.

前述のA、B点からそれぞれ枠の内側に投影濃度分布を
たどり、最初の極小点(第9図矢印)をもって文字抽出
範囲とするのである。又、他の2点についても同様であ
る。文字抽出範囲は、第9図一点鎖線で囲まれた領域と
なる。以上説明したように文字領域の判別までA−D変
換後の多値画像のままで行うので、該領域に対して、2
値化閾値を決めれば、適正な2値化を行うことができ、
またその後、バタンマツチング方法の精度を高めること
ができ非常に精度の高い画像処理装置を提供することが
可能となった。
The projected density distribution is traced inside the frame from the aforementioned points A and B, respectively, and the first minimum point (arrow in FIG. 9) is taken as the character extraction range. The same applies to the other two points. The character extraction range is the area surrounded by the dashed line in FIG. As explained above, since the character area is determined using the multivalued image after A-D conversion, two
By determining the digitization threshold, proper binarization can be performed.
Further, since then, the accuracy of the slam matching method has been improved, and it has become possible to provide an image processing device with extremely high accuracy.

前記第8図に示したマスクパターンのω、hには多少の
冗長性があるのである範囲内の値であれば、充分な機能
を発揮する。またその形も例えば第10図(a) 、 
(b)のようにいくつかの変形が考えられ、第3図のも
のだけに限定されるわけではない。また、第7図の投影
濃度分布を得る際、領域41内の全ての点を抽出する必
要はなく、例えばX軸方向投影濃度分布を作成するとき
は、第11図の如く、走査ラインをy方向に何本かおき
にとる、即ち図中81の一点鎖線上の濃度分布のみをX
方向投影濃度分布として積算する、といったことも可能
である。
Since there is some redundancy in ω and h in the mask pattern shown in FIG. 8, a sufficient function can be achieved if the values are within a certain range. The shape is also as shown in Figure 10(a), for example.
Several modifications are possible, as shown in (b), and the invention is not limited to the one shown in FIG. Furthermore, when obtaining the projected density distribution shown in FIG. 7, it is not necessary to extract all points within the region 41. For example, when creating the projected density distribution in the In other words, only the concentration distribution on the dashed line 81 in the figure
It is also possible to integrate as a directional projection density distribution.

以上、本発明によれば、所定の枠内に書きこまれた文字
等の画像を抽出する際に、ドロップアウトカラー等の細
工をする必要がなく、例えば、PPCにより自家作製し
た帳票を用いることも可能となる。また多値画像のまま
処理を行うため、2値画像処理で不適正なしきい値を設
定した場合におこりがちな帳票の汚れ、かすれ等の誤動
作を減少させ安定して文字を抽出することができ、その
後の2値化、認識等の処理も容易となる。
As described above, according to the present invention, when extracting images such as characters written in a predetermined frame, there is no need to use tricks such as drop-out colors, and, for example, a form made in-house using PPC can be used. is also possible. In addition, since processing is performed as a multivalued image, characters can be extracted stably by reducing malfunctions such as smearing and blurring of forms that tend to occur when an inappropriate threshold is set in binary image processing. , subsequent processing such as binarization and recognition becomes easier.

〔効果〕〔effect〕

以上、詳述したように本願発明により、文字枠にドロッ
プアウトカラー等を使用することなく、認識すべき文字
領域を抽出し、該文字画像に対して、適正な2値化処理
を行うことが可能となった。
As detailed above, according to the present invention, it is possible to extract the character area to be recognized without using dropout colors or the like in the character frame, and to perform appropriate binarization processing on the character image. It has become possible.

本願発明により、画像データの2値化処理以前に多値画
像のまま前処理を行うので、適正な2値化を容易に行う
ことが可能となった。
According to the present invention, since preprocessing is performed on the multivalued image before the image data is binarized, it is possible to easily perform proper binarization.

【図面の簡単な説明】[Brief explanation of the drawing]

第1図は、文字枠濃度と文字濃度との差が大きいか、或
いは文字同士の濃度にバラツキがあるような帳票の一例
を示す図、第2図は本発明の適用が可能な装置の例を示
す図、第3図は本発明適用の装置の構成ブロック図、第
4図は加惇器、相関器の説明図、第5図は文字枠の識別
の70−チャート、第6図は帳票上の文字及び文字枠の
配列例及び座標軸の説明図、第7図は投影濃度分布を説
明する図、第8図は相関処理に用いるマスクパターンを
示す図、第9図は文字領域の切り出し範囲を説明する図
、第10図は相関処理に用いるマスクパターンの他の例
を示す図である。第11図は走査ラインをy方向に何本
かおきにとった場合の説明図。
FIG. 1 is a diagram showing an example of a document in which there is a large difference between the character frame density and character density, or in which there is variation in the density between characters, and FIG. 2 is an example of an apparatus to which the present invention can be applied. 3 is a configuration block diagram of the device to which the present invention is applied, FIG. 4 is an explanatory diagram of the adder and correlator, FIG. 5 is a 70-chart for character frame identification, and FIG. 6 is a form diagram. Figure 7 is a diagram explaining the projected density distribution, Figure 8 is a diagram showing the mask pattern used for correlation processing, Figure 9 is the cutout range of the character area. FIG. 10 is a diagram illustrating another example of a mask pattern used for correlation processing. FIG. 11 is an explanatory diagram when scanning lines are taken every few lines in the y direction.

Claims (1)

【特許請求の範囲】[Claims] 枠情報と画像情報の読取りが可能な読取り手段、上記、
読取り手段によつて読取られた情報を認識するための前
処理を多値デイジタルデータで行うことを特徴とする画
像処理装置。
A reading means capable of reading frame information and image information;
An image processing apparatus characterized in that preprocessing for recognizing information read by a reading means is performed on multivalued digital data.
JP59276463A 1984-12-28 1984-12-28 Picture image processor Pending JPS61156484A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP59276463A JPS61156484A (en) 1984-12-28 1984-12-28 Picture image processor

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP59276463A JPS61156484A (en) 1984-12-28 1984-12-28 Picture image processor

Publications (1)

Publication Number Publication Date
JPS61156484A true JPS61156484A (en) 1986-07-16

Family

ID=17569792

Family Applications (1)

Application Number Title Priority Date Filing Date
JP59276463A Pending JPS61156484A (en) 1984-12-28 1984-12-28 Picture image processor

Country Status (1)

Country Link
JP (1) JPS61156484A (en)

Similar Documents

Publication Publication Date Title
US7751648B2 (en) Image processing apparatus, image processing method, and computer program
KR100512831B1 (en) Image processing method, apparatus and program storage medium
US5748809A (en) Active area identification on a machine readable form using form landmarks
US20070242882A1 (en) Image processing apparatus for identifying the position of a process target within an image
US20140119593A1 (en) Determining pose for use with digital watermarking, fingerprinting and augmented reality
KR101064845B1 (en) Passive embedded interaction coding
EP1300779B1 (en) Form recognition system, form recognition method, program and storage medium
JP2006025129A (en) System and method for image processing
JP2010056691A (en) Device and method for processing image
JPS62254282A (en) Method and apparatus for separating overlapped pattern
JP2008028716A (en) Image processing method and apparatus
JP2985935B2 (en) Handwritten character / graphic reader
JPS61156484A (en) Picture image processor
JP2008219800A (en) Writing extraction method, writing extracting device, and writing extracting program
JP2020047138A (en) Information processing apparatus
JP2894111B2 (en) Comprehensive judgment method of recognition result in optical type character recognition device
JP3756660B2 (en) Image recognition method, apparatus and recording medium
JPS5949671A (en) Optical character reader
JP2023034823A (en) Image processing apparatus, and control method, and program for image processing apparatus
JP3162451B2 (en) Preprocessing method for character recognition
JP2006107290A (en) Image processor and method thereof
JP2925270B2 (en) Character reader
KR20220005243A (en) Sharing and recognition method and device of handwritten scanned document
JP2778436B2 (en) Character segmentation device
JP2005208979A (en) Featured value extracting device and method and document filing device