JPH03269782A - Character area extracting method for character reader - Google Patents
Character area extracting method for character readerInfo
- Publication number
- JPH03269782A JPH03269782A JP2069965A JP6996590A JPH03269782A JP H03269782 A JPH03269782 A JP H03269782A JP 2069965 A JP2069965 A JP 2069965A JP 6996590 A JP6996590 A JP 6996590A JP H03269782 A JPH03269782 A JP H03269782A
- Authority
- JP
- Japan
- Prior art keywords
- character
- area
- character area
- value
- picture
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims description 6
- 239000011159 matrix material Substances 0.000 claims abstract description 16
- 239000000284 extract Substances 0.000 claims description 7
- 238000010586 diagram Methods 0.000 description 5
- 238000000605 extraction Methods 0.000 description 5
- 238000003384 imaging method Methods 0.000 description 5
- OKTJSMMVPCPJKN-UHFFFAOYSA-N Carbon Chemical compound [C] OKTJSMMVPCPJKN-UHFFFAOYSA-N 0.000 description 1
- 229910052799 carbon Inorganic materials 0.000 description 1
- 238000009499 grossing Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
Abstract
Description
【発明の詳細な説明】
〔産業上の利用分野〕
この発明は、CCDカメラ等の撮像装置を用いて撮像さ
れた、文字を含む多値画像から文字を読取る文字読取装
置における文字領域抽出方法に関する。[Detailed Description of the Invention] [Industrial Application Field] The present invention relates to a character area extraction method in a character reading device that reads characters from a multivalued image containing characters captured using an imaging device such as a CCD camera. .
[従来の技術]
従来の文字読取装置における文字領域抽出方法としては
、先ず、第4図(a)に示す画像に対して平滑化等のフ
ィルタ処理を施して第4図(b)に示すように、局所的
なノイズ20を除去し、次いで第4図(C)に示すよう
に、画像濃度値のヒストグラム分布を求め、これをもと
に適切な濃度値で画像を2値化することにより、第4図
(d)に示すように背景領域と文字領域を分離して文字
領域を抽出する方法が一般的である。[Prior Art] As a method for extracting a character area in a conventional character reading device, first, filter processing such as smoothing is performed on the image shown in FIG. 4(a), and the image shown in FIG. 4(b) is Then, as shown in FIG. 4(C), the local noise 20 is removed, the histogram distribution of image density values is obtained, and the image is binarized with appropriate density values based on this. A common method is to separate the background area and character area and extract the character area, as shown in FIG. 4(d).
しかしながら、上記従来の文字領域抽出方法にあっては
、例えば第5図(a)に示すように、背景画像が特徴的
なテクスチャパターンを有している場合であっても、そ
のテクスチャパターンの情報を用いず、第5図(b)に
示す濃度値のヒストグラム分布のみで背景領域と文字領
域とを分離しようとするため、得られた文字領域内には
、ノイズが現れたり、或いは抽出された文字が切れたり
して、適切な文字領域の抽出を行うことができないとい
う未解決の課題があった。However, in the conventional character area extraction method described above, even if the background image has a characteristic texture pattern, as shown in FIG. Since the background region and the character region are separated using only the histogram distribution of the density values shown in Fig. 5(b) without using the There was an unresolved problem that characters were cut off and it was not possible to extract an appropriate character area.
そこで、この発明は、上記従来例の未解決の課題に着目
してなされたものであり、特徴的なテクスチャパターン
を有する背景領域と文字領域とからなる多値画像から、
文字領域を正確に抽出することができる文字読取装置に
おける文字領域抽出方法を提供することを目的としてい
る。Therefore, the present invention has been made by focusing on the unresolved problems of the above-mentioned conventional example.
It is an object of the present invention to provide a character area extraction method for a character reading device that can accurately extract character areas.
上記目的を遠戚するために、この発明に係る文字読取装
置における文字領域抽出方法は、背景領域と文字領域と
からなる多値画像から文字領域を抽出して文字を読取る
文字読取装置において、前記多値画像の各点に対して、
画像の濃度値によってできる画像面上の直交する切口曲
線の長さを単位長さで計測したときの曲線被覆数から求
めた2次元フラクタル行列のトレースの値と行列式の値
を演算し、該演算値によって前記背景領域と文字領域と
を分離して文字領域を抽出するようにしたことを特徴と
している。In order to achieve the above object, a method for extracting a character region in a character reading device according to the present invention is provided. For each point in the multivalued image,
The trace value and determinant value of a two-dimensional fractal matrix obtained from the number of curves covered when the length of orthogonal cut curves on the image plane formed by the image density values are measured in unit length, and the values of the trace and determinant of the two-dimensional fractal matrix are calculated. The present invention is characterized in that the background area and the character area are separated from each other based on a calculated value, and the character area is extracted.
この発明においては、文字領域の抽出の際に、単に画像
の濃度値のみを基準として文字領域を抽出するのではな
く、2次元フラクタル行列のトレースの値と行列式の値
を演算し、両演算値から背景領域と文字領域とのテクス
チャパターンの違いを利用して文字領域を正確に抽出す
る。In this invention, when extracting a character area, instead of simply extracting the character area based on only the density value of the image, the value of the trace of the two-dimensional fractal matrix and the value of the determinant are calculated, and both calculations are performed. To accurately extract a character area from a value by using the difference in texture pattern between a background area and a character area.
以下、この発明の実施例を図面に基づいて説明する。 Embodiments of the present invention will be described below based on the drawings.
第1図はこの発明の一実施例を示す概略構成図である。FIG. 1 is a schematic diagram showing an embodiment of the present invention.
図中、1は文書を走査して画像データを出力するCCD
カメラ等の撮像装置であって、この撮像装置1から出力
される画像データが文字読取装置2に入力される。In the figure, 1 is a CCD that scans documents and outputs image data.
It is an imaging device such as a camera, and image data output from this imaging device 1 is input to a character reading device 2.
この文字読取装置2は、入力される画像データに基づい
て2次元フラクタル行列のトレース値と行列式の値とを
演算し、両演算値をもとに背景領域と文字領域とのテク
スチャパターンの違いを利用して文字領域を抽出する。This character reading device 2 calculates a trace value and a determinant value of a two-dimensional fractal matrix based on input image data, and based on both calculated values, the difference in texture pattern between a background area and a character area. Extract the character area using .
すなわち、今、撮像装置1から第2図(a)に示すよう
な特徴的なテクスチャパターンを有する背景領域3と文
字領域4とからなる多値画像5が文字読取装置2に入力
されているものとする。この第2図(a)において、x
、yの直交座標系を考え、さらにZ軸として画像の各点
における濃度値をとると、第2図中)に示すように濃度
値に応じた曲面の画像面6を有する多値画像を得ること
ができる。That is, a multivalued image 5 consisting of a background area 3 and a character area 4 having a characteristic texture pattern as shown in FIG. 2(a) is now input from the imaging device 1 to the character reading device 2. shall be. In this FIG. 2(a), x
, y, and taking the density value at each point of the image as the Z axis, we obtain a multivalued image having a curved image plane 6 according to the density value, as shown in Figure 2). be able to.
この第2図(ロ)において、多値画像上の任意の画素P
(x、y)を通り画像面6上で直交する切り口曲線7
x、7yを考え、この切り口曲線7x。In this FIG. 2 (b), any pixel P on the multivalued image
Cut curve 7 passing through (x, y) and orthogonal on the image plane 6
Considering x and 7y, this cut curve is 7x.
7yの長さを夫々単位長さUで計測したときの曲線被覆
数のうちX軸方向の値をX(u)、y軸方向の値をY
(u)とすると、下記(1)式のベクトル値関数V (
u)を得ることができる。When the length of 7y is measured with unit length U, the value in the X-axis direction of the number of curves covered is X(u), and the value in the y-axis direction is Y
(u), the vector-valued function V (
u) can be obtained.
V (u) =tCX (u)、 Y (u) )
−−(1)このとき、
(V(u)) =u−’ (V(1):l ・・・
−(2)を満たすような(2,2)行列Hが存在すれば
、■は2次元フラクタル量であり、行列Hは■に伴うフ
ラクタル行列であると定義することができる。V (u) = tCX (u), Y (u))
--(1) At this time, (V(u)) = u-' (V(1):l...
If there exists a (2,2) matrix H that satisfies -(2), it can be defined that ■ is a two-dimensional fractal quantity, and matrix H is a fractal matrix associated with ■.
したがって、行列Hを推定するには、上記(2)式%式
%())
(3)
で表される二次元ベクトルW(s)が得られる。Therefore, in order to estimate the matrix H, a two-dimensional vector W(s) expressed by the above equation (2) is obtained.
この(3)式の右辺が微分方程式W’ =−HWの解で
あることから、行列Hの各成分(h++、h+z。Since the right side of this equation (3) is the solution of the differential equation W' = -HW, each component of the matrix H (h++, h+z.
h2°hgz)を計算によって求めることができる。h2°hgz) can be obtained by calculation.
この算出された各成分hll+ hl□、 hz+
、 hz□O値から行列Hのトレースの値Tと行列式
の(iDとを算出する。Each of the calculated components hll+ hl□, hz+
, calculate the trace value T of the matrix H and the determinant (iD) from the hz□O value.
そして、上記の計算を画像上の全ての画素について行い
、その計算結果である行列Hのトレースの値Tと行列式
の値りの分布を表すと、第3図に示すように、背景領域
3及び文字領域4に対応する同一カテゴリーのサンプル
が互いにクラスタを作って、D=T” /4で表される
曲線lで分離される安定らせん焦点領域に含まれる領域
8及び安定節焦点領域に含まれる領域9の2つのテクス
チャパターンに分類される。Then, the above calculation is performed for all pixels on the image, and the distribution of the value T of the trace of the matrix H and the value of the determinant, which is the calculation result, is as shown in FIG. Samples of the same category corresponding to character region 4 form a cluster with each other, and are included in region 8 included in the stable helical focal region and in the stable nodal focal region separated by the curve l represented by D=T”/4. The area 9 is classified into two texture patterns.
したがって、第3図から背景領域と文字領域とを分離す
るための適切な境界10を求めることにより、この境界
10を使用して文字領域に該当する画素を正確に抽出す
ることができる。Therefore, by finding an appropriate boundary 10 for separating the background area and character area from FIG. 3, pixels corresponding to the character area can be accurately extracted using this boundary 10.
そして、抽出された文字領域について文字認識を行うこ
とにより、正確な文字読取を行うことが可能となる。Then, by performing character recognition on the extracted character area, it becomes possible to perform accurate character reading.
なお、上記実施例においては、画像上の各画素について
フラクタル行列のトレース後と行列式の値を計算する場
合について説明したが、これに限定されるものではなく
、画像の解像度や文字のサイズによっては、例えば数画
素置きにサンプリングして大局的な文字領域を先ず抽出
した後、その領域内をさらに細かく計算するようにする
こともでき、この場合には文字読取処理速度を速めるこ
とができる利点がある。In addition, in the above embodiment, the case was explained in which the fractal matrix is traced and the determinant value is calculated for each pixel on the image, but this is not limited to this, and depending on the resolution of the image and the size of the characters, For example, it is possible to first extract a global character area by sampling every few pixels, and then perform more detailed calculations within that area, which has the advantage of speeding up the character reading process. There is.
以上説明したように、この発明によれば、多値画像の2
次元フラクタル行列を用いて、テクスチャパターンを分
類し、背景領域と文字領域とを分離して文字領域を抽出
するようにしているので、背景領域が多少複雑な濃度値
の分布を示すような画像であっても、従来例のように画
像の濃度値のみにより文字領域を抽出する場合に比較し
て、正確に文字領域を抽出することができ、正確な文字
読取り行うことができる効果が得られる。As explained above, according to the present invention, two
A dimensional fractal matrix is used to classify the texture pattern, separate the background area from the text area, and extract the text area. Even if there is a problem, the character area can be extracted more accurately and the characters can be read more accurately than in the case of extracting the character area only based on the density value of the image as in the conventional example.
第1図はこの発明の一実施例を示すブロック図、第2図
(a)及び(ハ)は夫々この発明の文字抽出方法の説明
に供する説明図、第3図はトレース値と行列式の値との
関係を示すグラフ、第4図(a)〜(d)並びに第5図
(a)及び(b)は夫々従来例の文字抽出方法の説明に
供する説明図である。
図中、1は撮像装置、2は文字読取装置、3は背景領域
、4は文字領域、5は多値画像、6は画像面、7x、7
yは切り口曲線、10は境界である。
i炭質
濃度値FIG. 1 is a block diagram showing an embodiment of the present invention, FIGS. 2(a) and (c) are explanatory diagrams for explaining the character extraction method of the present invention, and FIG. 3 is a diagram showing trace values and determinants. Graphs showing the relationship with values, FIGS. 4(a) to 4(d) and FIGS. 5(a) and 5(b) are explanatory diagrams for explaining the conventional character extraction method, respectively. In the figure, 1 is an imaging device, 2 is a character reading device, 3 is a background area, 4 is a character area, 5 is a multivalued image, 6 is an image plane, 7x, 7
y is the cut curve and 10 is the boundary. i Carbon concentration value
Claims (1)
抽出して文字を読取る文字読取装置において、前記多値
画像の各点に対して、画像の濃度値によってできる画像
面上の直交する切口曲線の長さを単位長さで計測したと
きの曲線被覆数から求めた2次元フラクタル行列のトレ
ースの値と行列式の値を演算し、両演算値によって前記
背景領域と文字領域とを分離して文字領域を抽出するよ
うにしたことを特徴とする文字読取装置における文字領
域抽出方法。In a character reading device that extracts a character area from a multi-valued image consisting of a background area and a character area and reads the characters, an orthogonal cut on the image surface made by the density value of the image is made for each point of the multi-valued image. The trace value and the determinant value of the two-dimensional fractal matrix obtained from the curve coverage number when the length of the curve is measured in unit length are calculated, and the background area and the character area are separated using both calculated values. 1. A method for extracting a character area in a character reading device, characterized in that a character area is extracted using a character reading device.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2069965A JPH03269782A (en) | 1990-03-20 | 1990-03-20 | Character area extracting method for character reader |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2069965A JPH03269782A (en) | 1990-03-20 | 1990-03-20 | Character area extracting method for character reader |
Publications (1)
Publication Number | Publication Date |
---|---|
JPH03269782A true JPH03269782A (en) | 1991-12-02 |
Family
ID=13417875
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
JP2069965A Pending JPH03269782A (en) | 1990-03-20 | 1990-03-20 | Character area extracting method for character reader |
Country Status (1)
Country | Link |
---|---|
JP (1) | JPH03269782A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6055335A (en) * | 1994-09-14 | 2000-04-25 | Kabushiki Kaisha Toshiba | Method and apparatus for image representation and/or reorientation |
-
1990
- 1990-03-20 JP JP2069965A patent/JPH03269782A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6055335A (en) * | 1994-09-14 | 2000-04-25 | Kabushiki Kaisha Toshiba | Method and apparatus for image representation and/or reorientation |
US6275615B1 (en) | 1994-09-14 | 2001-08-14 | Kabushiki Kaisha Toshiba | Method and apparatus for image representation and/or reorientation |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP2940936B2 (en) | Tablespace identification method | |
US4905296A (en) | System for shape recognition | |
CN104751142B (en) | A kind of natural scene Method for text detection based on stroke feature | |
Liang et al. | Multi-spectral fusion based approach for arbitrarily oriented scene text detection in video images | |
CN106228528B (en) | A kind of multi-focus image fusing method based on decision diagram and rarefaction representation | |
CN114529459B (en) | Method, system and medium for enhancing image edge | |
CN107871319B (en) | Method and device for detecting beam limiter area, X-ray system and storage medium | |
CN106934806A (en) | It is a kind of based on text structure without with reference to figure fuzzy region dividing method out of focus | |
KR20190095606A (en) | System and Method for recognizing art objects in video combining deep learning technology and image feature extraction technology | |
Joseph | Processing of engineering line drawings for automatic input to CAD | |
Islam et al. | An approach to extract text regions from scene image | |
CN110120013A (en) | A kind of cloud method and device | |
CN110363192A (en) | Object image identification system and object image discrimination method | |
JPH03269782A (en) | Character area extracting method for character reader | |
CN114820987A (en) | Three-dimensional reconstruction method and system based on multi-view image sequence | |
CN115115535A (en) | Depth map denoising method, device, medium and equipment | |
JP2011170554A (en) | Object recognition device, object recognition method, and object recognition program | |
Amarnath et al. | Automatic localization and extraction of tables from handheld mobile-camera captured handwritten document images | |
Tan et al. | Character recognition based on corner detection and convolution neural network | |
JP2008152611A (en) | Image recognition device, electronic equipment, image recognition method, and image recognition program | |
CN116777905B (en) | Intelligent industrial rotation detection method and system based on long tail distribution data | |
JP2004145592A (en) | Motion vector extraction device, method and program, and its recording medium | |
JPS60189084A (en) | Character area extracting circuit | |
CN113674197B (en) | Method for dividing back electrode of solar cell | |
CN114581448B (en) | Image detection method, device, terminal equipment and storage medium |