JPS592192A - Character recognizing system - Google Patents

Character recognizing system

Info

Publication number
JPS592192A
JPS592192A JP57112215A JP11221582A JPS592192A JP S592192 A JPS592192 A JP S592192A JP 57112215 A JP57112215 A JP 57112215A JP 11221582 A JP11221582 A JP 11221582A JP S592192 A JPS592192 A JP S592192A
Authority
JP
Japan
Prior art keywords
character
signal
contraction
standard
expansion
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
JP57112215A
Other languages
Japanese (ja)
Inventor
Atsushi Tsukumo
津雲 淳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
NEC Corp
Original Assignee
NEC Corp
Nippon Electric Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by NEC Corp, Nippon Electric Co Ltd filed Critical NEC Corp
Priority to JP57112215A priority Critical patent/JPS592192A/en
Publication of JPS592192A publication Critical patent/JPS592192A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/751Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
    • G06V10/7515Shifting the patterns to accommodate for positional errors

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Theoretical Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Character Discrimination (AREA)

Abstract

PURPOSE:To enable character recognition by matching that absorbs two-dimensional expansion and contraction, by performing firstly horizontal one-directional normalization of expansion and contraction from projecting information on the horizontal axis of a character pattern, and then performing expansion and contraction matching in the vertical direction. CONSTITUTION:An input pattern is read from an input character pattern memory 1 as a signal 101, projecting information histogram on the horizontal axis is obtained and outputted as a projecting information signal 102. A mapping function generating device 3 obtains a mapping function of the projecting information signal 102 and a standard projecting information signal 104 and outputs as a mapping functon signal 103. A two- dimensional expansion and contraction matching device 5 reads the signal 101 and performs one-directional expansion and contraction normalization by using the number of mapping for each read character read as a signal 103, and reads the standard character pattern signal 106 of read character corresponding to above-mentioned mapping function stored in a standard character function memory 6, performs expansion and contraction matching in the vertical direction to find out difference, and outputs the difference between input character pattern and each read character as a signal 105. A discriminating device 7 reads it as a signal 105, and outputs, for instance, the result of output of character which is leaset in difference as a signal 107.

Description

【発明の詳細な説明】 本発明は、漢字、平仮名9炸仮名、英数字等のような多
くのストロークによって構成されている文字の認識方式
に関する。
DETAILED DESCRIPTION OF THE INVENTION The present invention relates to a method for recognizing characters composed of many strokes, such as kanji, hiragana, alphanumeric characters, etc.

近年光学式文字認識技術の発展は目覚ましいものがあシ
、英数字を認識対象とするものは手書き文字、印刷文字
のいずれも製品化され、実用に供している。また漢字、
平仮名を含む日本語用の文字を認識対象とするものは、
印刷文字単一フォントに限れば試作機の開発等が既に発
表されている。
The development of optical character recognition technology has been remarkable in recent years, and products that recognize alphanumeric characters, both handwritten and printed characters, have been commercialized and put into practical use. Also kanji,
For those that recognize Japanese characters including hiragana,
As far as single fonts for printed characters are concerned, the development of prototype machines has already been announced.

しかし漢字、平仮名1斥仮名、英数字等の手書き文字を
認識するために、手書き英数字の認識方式を拡張して手
書き漢字まで認識することや、印刷漢字認識からのアプ
ローチ等がとられているが、いまだに効果的なものが得
られていない。その理由の一つとして、漢字は英数字に
比べ複雑な形状をしているために、特徴情報の選択が難
しく、また他の理由として、変形が多いことから、安定
した特徴情報を得るのが困難であること等が挙げられる
However, in order to recognize handwritten characters such as kanji, hiragana, hiragana, and alphanumeric characters, approaches such as extending the recognition method for handwritten alphanumeric characters to recognize handwritten kanji, and approaches from printed kanji recognition are being taken. However, nothing effective has yet been achieved. One reason for this is that kanji have more complex shapes than alphanumeric characters, making it difficult to select feature information.Another reason is that kanji have many deformations, making it difficult to obtain stable feature information. For example, it is difficult.

一方、はけの効果とテンプレートマツチング法とを組み
合わせることにより、少数のデータに対して、実験が試
みられているもののあまシ良好な結果は得られていない
のが現状である。
On the other hand, although experiments have been attempted on small amounts of data by combining the brush effect and the template matching method, good results have not yet been obtained.

さて手書き文字の変動の原因として考えられるものは、 (1)位置ずれ(伸縮) (2)回転 の2点であシ、手書き文字を構成する各ストロークがそ
れぞれ独立に(1)と(2)の変動が起こるために、文
字パタン全体として歪みが生じるものである。
Now, there are two possible causes of fluctuations in handwritten characters: (1) positional deviation (expansion and contraction), and (2) rotation. This causes distortion in the entire character pattern.

このうち(1)の位置ずれに関しては、文字バタンか二
次元情報であるために、二方向に位置ずれが起こること
が、基本的な原因となっている。これに比べ音声認識処
理に注目すると、基本的に時間軸方向の一次元情報であ
ることから、DPマツチング法を用いて時間軸方向への
伸縮整合を行ない、位置ずれの問題を解決している。(
例えば特公昭49−45943号公報「バタン類似度計
算装置」)。
Of these, the basic cause of the positional deviation (1) is that the positional deviation occurs in two directions because it is a character slam or two-dimensional information. In contrast, if we focus on speech recognition processing, since it is basically one-dimensional information in the time axis direction, the DP matching method is used to perform expansion/contraction matching in the time axis direction to solve the problem of positional deviation. . (
For example, Japanese Patent Publication No. 49-45943 ``Batan Similarity Calculation Device'').

本発明は擬似的に二次元上を相異なる二方向についての
伸縮整合を実現し、従来困難とされた二次元的な位置ず
れを吸収した高精度の文字認識方式を提供するものであ
る。
The present invention provides a high-precision character recognition method that achieves stretch matching in two different directions on a two-dimensional plane in a pseudo manner, and absorbs two-dimensional positional deviations that have been considered difficult in the past.

以下図を用いて本発明について詳細な説明を行なうが、
相異なる二方向として水平方向と垂直方向を例にとる。
The present invention will be explained in detail below using the figures.
Let us take the horizontal direction and the vertical direction as two different directions.

その理由は、説明するうえでのわかシやすさと、二次元
バタンを扱うときに採用される頻度が多いためであシ、
他の相異なる二方向を採用しても同じ効果を得ることが
できる。
The reason for this is that it is easy to explain and is often used when dealing with two-dimensional batons.
The same effect can be obtained by using two different directions.

第1図は、二次元的な伸縮整合を直観的に説明するため
の図であシ、(atは標準文字バタン、(b)は入力文
字バタン、(elは入力文字バタン(blの水平軸上へ
の投影情報と標準文字バタン(a)水平軸上への投影情
報との伸縮整合が最適となるように入力文字バタンを水
平方向に伸縮正規化した一方向正規化文字バタン、そし
て(dlは前記一方向正規化文字バタン(e)と標準文
字バタン(a)との垂直方向の伸縮整合が最適になるよ
うに垂直方向に伸縮正規化した二方向正規化文字バタン
を示しておシ、本発明は標準文字バタン(atと入力文
字バタン(b)との整合を行なうときに、あたかも標準
文字バタン(a)と二方向正規化文字バタン(d)との
整合を行なうことを実現するものであり、この結果、二
次元的なストロークの位置ずれを吸収して文字を認識す
ることができる。
Figure 1 is a diagram for intuitively explaining two-dimensional expansion/contraction matching. (a) A one-way normalized character button in which the input character button is stretched and normalized in the horizontal direction so that the expansion and contraction matching between the upward projection information and the standard character button (a) is optimally matched with the projection information on the horizontal axis, and (dl indicates a two-way normalized character button that has been vertically expanded and contracted to optimize vertical expansion/contraction matching between the one-way normalized character button (e) and the standard character button (a); The present invention realizes, when matching the standard character BUTTON (at) and the input character BUTTON (b), as if matching the standard character BUTTON (a) and the two-way normalized character BUTTON (d). As a result, characters can be recognized while absorbing two-dimensional stroke positional deviations.

第2図(a)〜11)は、二次元的な伸縮整合の実現手
段とその効果を説明するための図であシ、同図(a)の
21は標準文字バタン、同図(b)の22は入力文字バ
タン、(C)の23は前記標準文字バタン21の水平軸
上への標準投影情報、同図(dlの24は前記入力文字
バタン22の水平軸上への入力投影情報、同図telは
前記標準投影情報23と前記入力投影情報24との水平
方向への伸縮整合を行ない、写像関数25を求めている
ことを示す図、同図(flは写像関数25を用いて、前
記入力文字バタン22の水平方向の一方向伸縮正規化バ
タン26を求めていることを示す図であり、そして同図
(g)は前記一方向伸縮正規化バタン26と前記標準文
字バタン21との垂直方向への伸縮整合を行ない、その
ときの写像関数はCとな゛ることを示すだめの図である
Figures 2 (a) to 11) are diagrams for explaining means for realizing two-dimensional expansion/contraction matching and their effects; 21 in Figure 2 (a) is a standard character button; 22 is an input character button, 23 in (C) is standard projection information on the horizontal axis of the standard character button 21, and in the same figure (24 in dl is input projection information on the horizontal axis of the input character button 22, tel in the figure shows that the standard projection information 23 and the input projection information 24 are expanded and contracted in the horizontal direction to obtain the mapping function 25; This is a diagram showing that a horizontal one-way stretch/contract normalization button 26 of the input character button 22 is obtained, and (g) of the same figure shows the difference between the one-way stretch/contraction normalization button 26 and the standard character button 21. This is a schematic diagram showing that the mapping function at that time is equal to C when expansion/contraction matching is performed in the vertical direction.

上述の説明の中で、伸縮整合で用いられる文字バタンは
、MxNのマトリクスから成っていて、Mが水平方向の
画素数、Nが垂直方向の画素数とすると、M次元ベクト
ルのN個の系列として記述されているものとみなし、水
平軸上への投影情報は一次元ベクトル、すなわちスカラ
ー量のM蘭の系列として記述されているものとする。ま
た(flで示している入力文字パタ/(b)の水平方向
の一方向伸縮正規化処理は、N個のM次元ベクトルをそ
れぞれ順次伸縮正規化するものである。
In the above explanation, the character button used in stretch matching consists of an MxN matrix, where M is the number of pixels in the horizontal direction and N is the number of pixels in the vertical direction. It is assumed that the projection information on the horizontal axis is described as a one-dimensional vector, that is, a series of M-rans of scalar quantities. Further, the horizontal one-way expansion/contraction normalization process of the input character pattern (indicated by fl/(b)) is to sequentially expand/contract and normalize each of the N M-dimensional vectors.

以上の説明で示す過少、本発明は、一方向伸縮整合を2
度行なうことによって、二次元的な伸縮整合を実現しよ
うとするものである。
In contrast to the understatement shown in the above explanation, the present invention can achieve two-way expansion/contraction alignment.
By repeating this process multiple times, the aim is to achieve two-dimensional expansion/contraction matching.

一方第2図(h)は水平方向の一方向伸縮正規化処理ρ
効果を示すための図であシ、図中21と22の黒の部分
は垂直方向へ伸縮して整合がとれた部分で、白ヌキの部
分は垂直方向へ伸縮しても整合されない部分を示してい
る。第2図(h)と前出第2図(glとを比較すること
によυ、垂直方向の伸縮整合処理の前に入力文字バタン
22に対して水平方向の一方向伸縮正規化処理を行なっ
た効果が示される。
On the other hand, Fig. 2 (h) shows the horizontal one-way expansion/contraction normalization process ρ.
This is a diagram to show the effect. In the figure, the black parts 21 and 22 are the parts that are aligned by vertical expansion and contraction, and the white blank parts are the parts that are not aligned even if they are expanded and contracted in the vertical direction. ing. By comparing Figure 2 (h) and Figure 2 (gl) above, it is found that unidirectional horizontal expansion/contraction normalization processing is performed on the input character button 22 before vertical expansion/contraction matching processing. The effect is shown.

第2図(1)は水平方向の一方向伸縮正規化処理に投影
情報ではなく二次元バタン情報そのものを使った場合を
説明するだめの図であシ、図中21と22の黒の部分は
水平方向へ伸縮して整合がとれた部分で、白ヌキの部分
は水平方向に伸縮しても整合されない部分を示している
が、ストロークの位置ずれに対して非常に不安定な整合
であることがわかる。本水平方向の一方向伸縮正規化処
理は次の垂直方向の伸縮整合処理の精度を大きく左右す
るものであシ、写像関数を求めるために安定な整合が必
要であシ、そのために文字バタンとしての情報が欠けて
もストロークの位置ずれを吸収している投影情報を用い
ることが必要となる。
Figure 2 (1) is a temporary diagram to explain the case where two-dimensional baton information itself is used instead of projection information for the horizontal one-way stretching/contraction normalization process, and the black parts 21 and 22 in the figure are This is the part that has been aligned by expanding and contracting in the horizontal direction.The white blank area shows the part that is not aligned even if it is expanded and contracted in the horizontal direction, but the alignment is extremely unstable due to positional displacement of the stroke. I understand. This one-way horizontal stretching/contraction normalization process greatly affects the accuracy of the next vertical stretching/contraction matching process, and stable matching is required to obtain the mapping function. Even if this information is missing, it is necessary to use projection information that absorbs the positional deviation of the stroke.

第3図(a) 、 (b) 、 (e) 、 (d)は
一方向伸縮整合処理として、音声認識で用いられている
DPマツチング法の一例を説明するだめの図である。
FIGS. 3(a), (b), (e), and (d) are diagrams for explaining an example of the DP matching method used in speech recognition as one-way expansion/contraction matching processing.

標準バタンAoがM次元ベクトルAo e Ao +・
・・Ao の系列から成シ、入力パタ/AがM次元ベク
トルA 、 A2. ・・・、ANの系列から成ってい
るとする。また、標準バタンの任意のベクトルA oj
と、人力バタンの任意のベクトルA: との距離をd(
l。
The standard baton Ao is an M-dimensional vector Ao e Ao +・
...Ao sequence, the input pattern /A is an M-dimensional vector A, A2. ..., is made up of a series of ANs. Also, any vector of standard batons A oj
and any vector A of the human-powered slam: Let the distance be d(
l.

j)とする。単純な整合をとると、入力バタンAと標準
バタンAoとの相違度D(A、Ao)は、例えば下式で
求めることになる。
j). When simple matching is performed, the degree of difference D (A, Ao) between the input baton A and the standard baton Ao can be obtained, for example, by the following formula.

この式は第3図(alの写像関数j=i 上で、J AとAoとを対応させて、両バタンの相違度を求めてい
るが、同図の写像関数j=ψ(1)上で、A1とAoj
とを対応させることができれば、両パタ/の相違度を求
めるのに、入力バタンAを部分的に伸縮して標準バタン
Aoと整合をとることができる。
This formula is shown in Figure 3 (on the mapping function j = i of al, J A and Ao are made to correspond and the degree of difference between the two batons is determined. So, A1 and Aoj
If it is possible to make these patterns correspond to each other, it is possible to partially expand or contract the input batten A to match it with the standard batten Ao in order to find the degree of difference between the two putters.

DPマツチング法は、入力バタ/を部分的に伸縮して整
合をとるための手法であシ、例えば第3図1b)では下
記の初期値及び漸化式から、f (N、N)を求めるこ
とによシ、写像関数j=ψ(1)上でAjとAojとを
対応させて整合をとることができる。
The DP matching method is a method for matching the input data by partially expanding or contracting it. For example, in Fig. 3 (1b), f (N, N) is found from the following initial values and recurrence formula. In particular, matching can be achieved by making Aj and Aoj correspond on the mapping function j=ψ(1).

g(1,1)=d(1,1) す(’ yD”d(1−j)+*x(l’ (i 1 
、 j) −9(11−j 1)+g(l −1tj−
2) ) ただし、d(1,j)=cx:+  (1≦O1たはj
≦0)である。
g(1,1)=d(1,1)('yD"d(1-j)+*x(l' (i 1
, j) −9(11−j 1)+g(l −1tj−
2) ) However, d(1,j)=cx:+ (1≦O1 or j
≦0).

第3図(c)は上記漸化式を求めるDPマツチング法の
一例を示すための図であル、入力バタンは5個の一次元
ベクトル、すなわちスカラー量の系列(1,2,4,5
,5)であシ、標準バタンは同じく5個の系列(1,2
,3,4,5)であシ、(1,j)が(1,1)、(2
,2)、(3,4)、(4,5)、(5,5)となる写
像関数上の伸縮整合を行なっている。
FIG. 3(c) is a diagram showing an example of the DP matching method for obtaining the above recurrence formula.
, 5), the standard baton has the same 5 series (1, 2
, 3, 4, 5) and (1, j) is (1, 1), (2
, 2), (3, 4), (4, 5), and (5, 5).

第3図(d)は上記漸化式計算の計算量を減少させるた
めに I−△≦j≦ 1十Δ つ範囲内で、漸化式計算を行なうことを示しておシ、一
般にDPマツチング法では、この範囲を整合窓と呼び、
実際に計算量の効率化を図っている。
Figure 3(d) shows that the recurrence formula calculation is performed within the range of I-△≦j≦10∆ in order to reduce the amount of calculation in the above recurrence formula calculation. In the law, this range is called the matching window.
We are actually trying to make the amount of calculation more efficient.

前記漸化式は単に相違度を求めるためだけのものである
が、 1n(f(1−1,D、f(1−1,j−1)、g(1
−1,j−2))=9(11,j(1−1ρ (ただし、j(1−1)はj、j−1,j−2のいずれ
がである)のとき、 h(l 、j)=j(1−1) として、関数h(i、J)を求めておくことにより、相
違度が求められた後にh(1,1)の値をh(N、N)
から順次J1(、t、t)まで求めることにょシ写像関
数を求めることができる。例えば第3図(clの例では
h(5,5) h(5,5)=5 、 h(4,5)=4 、 h(3
,4)=2 。
The above recurrence formula is only for calculating the degree of dissimilarity, but 1n(f(1-1, D, f(1-1, j-1), g(1
-1,j-2))=9(11,j(1-1ρ (however, j(1-1) is j, j-1, j-2), then h(l, By calculating the function h(i, J) as j)=j(1-1), the value of h(1, 1) can be changed to h(N, N) after the degree of dissimilarity is calculated.
The mapping function can be obtained by sequentially obtaining J1(, t, t) from J1(, t, t). For example, in Figure 3 (cl example) h(5,5) h(5,5)=5 , h(4,5)=4 , h(3
,4)=2.

h(2,2)=1 であるから、写像関数(1,j)が (1,1) 、 (2,,2) 、 (3,4) 、 
(4,5)、(5,5)と求まる。
Since h(2,2)=1, the mapping function (1,j) is (1,1), (2,,2), (3,4),
(4,5), (5,5) are found.

第4図は伸縮正規化処理の一例を示すだめの図であり、
)]1)(1≦i≦16)は入カッ(タン、Y (j)
(1≦j≦16)は伸縮正規化ノ(タンで、j=ψ(1
)は伸縮正規化のための写像関数である。この例ではY
(jlは次の規則によって定まる。
FIG. 4 is a diagram showing an example of expansion/contraction normalization processing,
)] 1) (1≦i≦16) is input (tan, Y (j)
(1≦j≦16) is the expansion/contraction normalization (tan), and j=ψ(1
) is a mapping function for stretch normalization. In this example, Y
(jl is determined by the following rule.

(1)  j=ψ(i)〉ψ(i−1)かつψ(i)〈
ψ(++1)のときY(j)=X(1) (2)  j=ψ(1)=ψ(i−1)+2のとき Y
 (j−1) =X(1)(3)  j=ψ(1)=ψ
(t−i)<ψ(i+1)のときY(j)=X(1)第
5図は本発明方式を実現するだめの装置の一実施例を示
すブロック図である。100は入力文字バタン信号であ
り、1は前記入力文字ノくタンを格納する入力文字ノく
タン記憶部である。2は投影情報抽出手段でアシ、入力
文字ノくタン記憶部2力≧ら入力バタンを信号101と
して読み込み水平軸上の投影情報ヒストグラムを求め、
投影情報信号102として出力する。3は写像関数生成
手段であシ、前記投影情報信号102と、標準投影情報
記憶部4に前記投影情報信号102と同一形式で格納さ
れている各被読取シ字種ごとの標準投影情報信号104
との写像関数を求め、写像関数信号103として出力す
る。5は二次元伸縮整合手段で、入力文字バタン信号1
01を読込み、信号103として読込まれる各被読取ル
字種ごとの写像関数を用いて一方向伸縮正規化処理を行
ない、標準文字バタン記憶部6に格納されている前記写
像関数に対応する被読取シ字種の標準文字バタン信号1
06を読込み、垂直方向の伸縮整合を行なって相違度を
求め、入力文字バタンと各被読取シ字種との相違度を信
号105として出力する。識別手段7では前記各被読取
シ字種との相違度を信号105として読込み、例えば単
に相違度の最も小さい字種を出力結果としたル、或いは
最も小さい相違度と、2番目に小さい相違度の差がある
値以上のときに最も相違度の小さい字種を出力結果とし
、他の場合にはりジェクトを出力結果とする等の文字認
識における通常の方法により認識結果を信号107とし
て出力する。
(1) j=ψ(i)〉ψ(i-1) and ψ(i)〈
When ψ(++1), Y(j)=X(1) (2) When j=ψ(1)=ψ(i-1)+2, Y
(j-1) =X(1)(3) j=ψ(1)=ψ
When (ti)<ψ(i+1), Y(j)=X(1) FIG. 5 is a block diagram showing an embodiment of a device for realizing the method of the present invention. 100 is an input character button signal, and 1 is an input character button storage section that stores the input character button. 2 is a projection information extracting means that reads the input button as a signal 101 from the input character noktan storage unit 2 and obtains a projection information histogram on the horizontal axis;
It is output as a projection information signal 102. 3 is a mapping function generating means which generates the projection information signal 102 and a standard projection information signal 104 for each character type to be read, which is stored in the standard projection information storage section 4 in the same format as the projection information signal 102.
A mapping function is obtained and output as a mapping function signal 103. 5 is a two-dimensional expansion/contraction matching means, which receives input character slam signal 1;
01 is read, one-way expansion/contraction normalization processing is performed using the mapping function for each character type read as the signal 103, and the mapping function corresponding to the mapping function stored in the standard character button storage section 6 is read. Standard character slam signal 1 for reading type
06 is read in, vertical expansion/contraction matching is performed to determine the degree of difference, and the degree of difference between the input character BUTTON and each type of character to be read is output as a signal 105. The identification means 7 reads the degree of dissimilarity from each character type to be read as a signal 105, and outputs, for example, simply the character type with the smallest degree of dissimilarity, or the smallest degree of dissimilarity and the second smallest degree of dissimilarity. When the difference is greater than a certain value, the character type with the smallest degree of difference is outputted as the output result, and in other cases, the recognition result is outputted as a signal 107 using a normal method in character recognition, such as outputting a blemish.

上記説明において、入力バタン記憶部1と投影情報抽出
手段2とは、一般にバタン処理で用いられているもので
よい。
In the above description, the input button storage section 1 and the projection information extraction means 2 may be those commonly used in the button processing.

第6図は写像関数生成手段3の構成の一例を示すブロッ
ク図である。ここでの処理は前記DPマツチング法の説
明の中の、漸化式f(1,j)の計算と、漸化式計算の
結果得られる軌跡h(i、j)を求め、h(1,j)か
ら写像関数を求めるものである。
FIG. 6 is a block diagram showing an example of the configuration of the mapping function generating means 3. The process here is to calculate the recurrence formula f(1, j) in the explanation of the DP matching method, find the trajectory h(i, j) obtained as a result of the recurrence formula calculation, and calculate h(1, j) to find the mapping function.

102は前記投影情報信号で、スカラー量の系列A、・
・・A 、・・・、Aに対応し、104は前記標準投影
情報信号で、咎被読取シ字種毎のスカラー1の系列Ao
 e 、Ao +・・・、 Ao に対応し、31は距
離演算部で上記2信号を入力とし、d(i、j)を計算
し、信号311として出力する。32は前出の漸化式 %式%) )) を計算する漸化式演算部で、d(1,j)を信号311
゜mg*(f(1−1,j)、f(量−1,j−1)、
f(1−1,j−2) )を信号341として入力し、
演算結果の9(1,j)を信号321として、累積値記
憶部33に出力する。
102 is the projection information signal, which is a series of scalar quantities A, .
...A, ..., A, and 104 is the standard projection information signal, which is a series Ao of scalar 1 for each character type to be read.
31 corresponds to e, Ao + . 32 is a recurrence formula calculation unit that calculates the above-mentioned recurrence formula %)).
゜mg*(f(1-1,j), f(amount-1,j-1),
f(1-1,j-2)) as the signal 341,
The calculation result 9(1,j) is output to the cumulative value storage section 33 as a signal 321.

34は最小値選択部で、累積値記憶部33から9(1−
1,j)、f(1−1,j−1)そしてq(s −1、
j−2)を信号331、信号332そして信号333と
して読込み、m(f(1−1,j)、f(1−1,j−
1)、f(i−1,j−2))を信号341、そしてh
(1,j)を信号342として写像軌跡記憶部35に出
力する。漸化式演算が終了すると前記写像軌跡記憶部3
5から写像関数を信号103として出力する。
34 is a minimum value selection unit, and cumulative value storage unit 33 to 9 (1-
1,j), f(1-1,j-1) and q(s-1,
j-2) as the signal 331, signal 332 and signal 333, m(f(1-1,j), f(1-1,j-
1), f(i-1, j-2)) as the signal 341, and h
(1, j) is output to the mapping locus storage unit 35 as a signal 342. When the recurrence formula calculation is completed, the mapping locus storage unit 3
5 outputs the mapping function as a signal 103.

第7図は、二次元伸縮整合手段5の構成の一例を示すブ
ロック図である。51は一方向伸縮正規化手段で、入力
文字バタン信号101と、各被読取シ字種に対応する写
像関数信号103とから、各被読取シ字種に対応する一
方向伸縮正規化手段バタン信号510を出力する。52
は文字バタン伸縮整合手段で、各被読取シ字種に対応す
る、一方向伸縮正規化文字バタン信号511と標準文字
バタン信号106とから、各被読取シ字種に対応する相
違度を信号105として出力する。
FIG. 7 is a block diagram showing an example of the configuration of the two-dimensional expansion/contraction matching means 5. As shown in FIG. Reference numeral 51 denotes a one-way expansion/contraction normalization means, which generates a one-way expansion/contraction normalization means bang signal corresponding to each type of character to be read from the input character bang signal 101 and the mapping function signal 103 corresponding to each type of character to be read. 510 is output. 52
is a character slam expansion/contraction matching means that calculates the degree of difference corresponding to each type of characters to be read from the one-way expansion/contraction normalized character slam signal 511 and the standard character slam signal 106 corresponding to each type of characters to be read. Output as .

伸縮正規化手段51は、入力文字バタン101をベクト
ルA 、 A 、−・・、A の系列として読込み、各
ベクトルについて信号103で決められた写像関数を用
いて第4図で説明した規則に従って、ベクトルA’、A
2.・・・1Mの系列を信号510として出力するが、
これは一方向伸縮正規化文字バタンとなっている。この
一方向伸縮正規化文字バタンを各被読取シ字種に対して
求める。すなわち各被読取り字種に対応する写像関数に
対して、一方向伸縮正規化文字バタンを信号510とし
て順次出力する。
The expansion/contraction normalization means 51 reads the input character button 101 as a series of vectors A , A , ..., A , and uses the mapping function determined by the signal 103 for each vector according to the rules explained in FIG. 4. Vector A', A
2. ...outputs a 1M sequence as signal 510,
This is a one-way stretch normalized character bang. This one-way expansion/contraction normalized character pattern is obtained for each type of character to be read. That is, the one-way expansion/contraction normalized character button is sequentially outputted as a signal 510 to the mapping function corresponding to each character type to be read.

第8図は文字バタン伸縮整合手段の構成の一例を示すブ
ロック図である。521はベクトル距離演算部で、各被
読取シ字種に対応する一方向伸縮正規化処理バタンと標
準文字バタンを、それぞれベクトルの系列の信号510
と信号106として読込んで、DPマツチング法のd(
1,j)の距離演算を行ない、信号5211として出力
する。522は漸化式演算部で、写像関数生成手段3の
漸化式演算部32と同一のものでよく、漸化式9式%) )) を計算するもので、d(i、j)を信号5211として
、そして月(f(1−1,j)、f(i−1,j−1)
、9(1−1,j−2))を信号5241として読込み
、g(t、j)を信号5221として相違度累積値記憶
部523に出力する。
FIG. 8 is a block diagram illustrating an example of the configuration of the character punch expansion/contraction matching means. Reference numeral 521 denotes a vector distance calculation unit, which converts the one-way expansion/contraction normalization process button and standard character button corresponding to each character type to be read into vector series signals 510.
is read as the signal 106, and the DP matching method d(
1, j) is performed and output as a signal 5211. Reference numeral 522 denotes a recurrence formula calculation unit, which may be the same as the recurrence formula calculation unit 32 of the mapping function generation means 3, and is used to calculate the recurrence formula 9). as the signal 5211, and the moon (f(1-1,j), f(i-1,j-1)
, 9(1-1, j-2)) as a signal 5241, and output g(t, j) as a signal 5221 to the difference cumulative value storage unit 523.

524は相違度最小値選択部で、相違度累積値記憶部5
23からf(i−x+j)−g(t−i、j−1)そし
てg(i −1、j−2)を信号5231 、信号52
32、そして信号5233として胱込み、m(g(i−
1,j)、q(i−1゜j−1)、g(1−1,j−2
)) を信号5241として出力する。漸化式演算が終
了すると、相違度累積値記憶部523は、相違度g(N
、N)を信号105として出力する。上記の処理によシ
、各被読取シ字種に対応する相違度を信号105として
順次出力する。
524 is a difference minimum value selection unit, and a difference degree cumulative value storage unit 5
23 to f(i-x+j)-g(t-i, j-1) and g(i-1, j-2) as signals 5231 and 52
32, and the signal 5233 includes the bladder, m(g(i-
1,j), q(i-1゜j-1), g(1-1,j-2
)) is output as a signal 5241. When the recurrence formula calculation is completed, the cumulative difference value storage unit 523 stores the difference g(N
, N) as a signal 105. Through the above processing, the degree of difference corresponding to each character type to be read is sequentially outputted as a signal 105.

第9図は投影情報として用いることのできる別の情報の
例を示す図で、文字バタンを垂直方向に走査して、文字
部と交差する回数を投影情報として採用するもので、取
シ扱いは、先に説明した投影情報と同様である。
Figure 9 is a diagram showing another example of information that can be used as projection information.The number of times the character stamp crosses the character part by scanning it in the vertical direction is used as the projection information. , is similar to the projection information described above.

以上の説明により、本発明によれば、文字パタンの水平
軸上の投影情報から、まず水平方向への一方向伸縮正規
化処理を行ない、次に垂直方向に伸縮整合を行なうこと
により、二次元的な伸縮を吸収する整合による文字認識
を実現することができる。
As described above, according to the present invention, from the projection information of the character pattern on the horizontal axis, first a one-way stretching/contracting normalization process in the horizontal direction is performed, and then a stretching/contracting matching process is performed in the vertical direction. It is possible to realize character recognition using alignment that absorbs natural expansion and contraction.

上記処理とは反対に、文字パタンの垂直軸上の投影情報
から垂直方向への一方向伸縮正規化処理を行ない、次に
水平方向に伸縮整合を行なう仁とにより、同様に二次元
的な伸縮を吸収する整合による文字認識を実現すること
ができる。また相異なる二方向としては上記の垂直方向
と水平方向に限るものではない。
Contrary to the above process, by performing unidirectional stretching/contraction normalization processing in the vertical direction from the projection information on the vertical axis of the character pattern, and then performing stretching/contraction matching in the horizontal direction, similarly two-dimensional stretching/contraction is performed. It is possible to realize character recognition by matching that absorbs Further, the two different directions are not limited to the above-mentioned vertical direction and horizontal direction.

文字認識方式では一般に位置や大きさの正規化、文字パ
タンの平滑化やぼけ処理等を行なって、認識方式の効果
を出そうとするものが多いが、本発明による文字認識方
式も、入力文字パタンに対して前処理を行なうことによ
っても他の方式と同様の効果を得ることができる。
In general, many character recognition methods try to achieve the effectiveness of the recognition method by normalizing the position and size, smoothing character patterns, blurring, etc., but the character recognition method according to the present invention also Effects similar to those of other methods can also be obtained by performing preprocessing on the pattern.

またDPマツチング法もこれまでに様々な方法が発表さ
れておシ、本明細書で説明した方式に限るものでは々い
Furthermore, various DP matching methods have been published so far, and are not limited to the method described in this specification.

【図面の簡単な説明】[Brief explanation of the drawing]

第1図(a) 、 (bl 、 (c) 、 (dlは
二次元的な伸縮整合を直観的に説明するための図、第2
図(a)〜(i)は二次元的な伸縮整合の実現手段とそ
の効果と説明するだめの図、第3図(at 、 (bl
 、 (cl 、 (dlはDPマツチング法の一例を
示すための図、第4図は伸縮正規化処理の一例を示すた
めの図、第5図は本発明方式を実現するための装置の一
実施例を示すブロック図、第6図は写像関数生成手段の
構成の一例を示すブロック図、第7図は二次元伸縮整合
手段の構成の一例を示すブロック図、第8図は文字バタ
ン伸縮整合手段の構成の一例を示すブロック図、第9図
は投影情報として用いることのできる別の情報の例を示
す図である。図中1は入力文字バタン記憶部、2は投影
情報抽出手段、3は写像関数生成子段、4は標準投影情
報記憶部、5は二次元伸縮整合手段、6は標準文字バタ
ン記憶部、7は識別手段、31は距離演算部、32は漸
化式演算部、33は累積値記憶部、34は最小値選択部
、35は写像軌跡記憶部、51は一方向伸縮正規化手段
、52は文字バタン伸縮整合子n21はベクトル距離演
算部522は漸化式演算部、523は相違度累積値記憶
部を示している。 /+ l ロ (Ql                (b)(C) (d) 牙 2 図 ((ス、)                    
      (bン(C)             
(d)(e) オ 2 μs (4) (S) 号 2 ロ (h) 芳 3 μ5 (cL) (b) 才 3 図 (C) 計 4 図 / / / 000 / / / l 000 / /
 / (x (L)牙 5 ロ 牙 6 μs オ 7図 号 8 圓 〉 (
Figure 1 (a), (bl, (c), (dl) is a diagram for intuitively explaining two-dimensional expansion/contraction matching, Figure 2
Figures (a) to (i) are diagrams for explaining the means for realizing two-dimensional expansion/contraction matching and their effects, and Figure 3 (at, (bl)
, (cl, (dl is a diagram showing an example of the DP matching method, FIG. 4 is a diagram showing an example of expansion/contraction normalization processing, and FIG. 5 is an example of an implementation of a device for realizing the method of the present invention. A block diagram showing an example, FIG. 6 is a block diagram showing an example of the configuration of the mapping function generation means, FIG. 7 is a block diagram showing an example of the configuration of the two-dimensional expansion/contraction matching means, and FIG. 8 is a block diagram showing an example of the configuration of the two-dimensional expansion/contraction matching means. FIG. 9 is a block diagram showing an example of the configuration of , and FIG. 9 is a diagram showing another example of information that can be used as projection information. Mapping function generator stage, 4 is a standard projection information storage section, 5 is a two-dimensional expansion/contraction matching means, 6 is a standard character button storage section, 7 is an identification means, 31 is a distance calculation section, 32 is a recurrence formula calculation section, 33 34 is a minimum value selection unit, 35 is a mapping locus storage unit, 51 is a one-way expansion/contraction normalization means, 52 is a character slam expansion/contraction matcher n21 is a vector distance calculation unit, 522 is a recurrence formula calculation unit, 523 indicates the cumulative difference value storage unit.
(bn(C)
(d) (e) O 2 μs (4) (S) No. 2 B (h) Yoshi 3 μ5 (cL) (b) Year 3 Figure (C) Total 4 Figure / / / 000 / / / l 000 / /
/ (x (L) Fang 5 Ro Fang 6 μs O 7 Symbol 8 En> (

Claims (1)

【特許請求の範囲】[Claims] 二次元メツシー状の情報として表わされる入力文字バタ
ンを認識する方式について、前記入力文字バタンを格納
する入力文字バタン記憶手段と、前記入力文字バタンに
対してあらかじめ定められた相異なる二方向のうち、一
方向に対する一次元情報の系列となる投影情報を抽出す
る投影情報抽出手段と、前記入力文字バタンの投影情報
と同一形式で、あらかじめ字種ごとに作成された標準投
影情報を格納している標準投影情報記憶手段と、前記入
力文字バタンの投影情報と標準投影情報とを入力とし、
両者を伸縮整合し一致の尺度が最大となるような写像関
数を求める写像関数生成手段と、あらかじめ字種ごとに
作成された二次元メツシュ状の標準文字バタンを格納す
る標準文字バタン記憶手段と、前記入力文字バタンと前
記写像関数と、前記標準文字パタンとを用いて、前記写
像関数によシ前記入力文字パタンを前記一方向への伸縮
整合した後、前記一方向とは異なる方向に伸縮整合を行
なう二次元伸縮整合手段と、前記伸縮整合の結果得られ
る入力文字バタンと各被読取字種の標準文字パタンとの
相違度から、認識結果を出力する識別手段とを有するこ
とにょシ、二次元的な伸縮変動を吸収して整合を行なう
ことができることを特徴とする文字認識方式。
Regarding the method of recognizing input character bangs expressed as two-dimensional mesh-like information, an input character bang storage means for storing the input character bangs, and two different predetermined directions for the input character bangs are provided. A projection information extraction means for extracting projection information that is a series of one-dimensional information for one direction, and a standard that stores standard projection information created in advance for each character type in the same format as the projection information of the input character slam. a projection information storage means, inputting the projection information of the input character button and the standard projection information;
a mapping function generating means for calculating a mapping function that stretches and contracts the two and maximizes the degree of matching; a standard character button storage means for storing two-dimensional mesh-like standard character buttons created in advance for each character type; Using the input character button, the mapping function, and the standard character pattern, the input character pattern is stretched and matched in the one direction by the mapping function, and then stretched and matched in a direction different from the one direction. and an identification means that outputs a recognition result based on the degree of difference between the input character stamp obtained as a result of the expansion and contraction matching and the standard character pattern of each type of character to be read. A character recognition method that is characterized by being able to perform alignment by absorbing dimensional expansion/contraction fluctuations.
JP57112215A 1982-06-29 1982-06-29 Character recognizing system Pending JPS592192A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP57112215A JPS592192A (en) 1982-06-29 1982-06-29 Character recognizing system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP57112215A JPS592192A (en) 1982-06-29 1982-06-29 Character recognizing system

Publications (1)

Publication Number Publication Date
JPS592192A true JPS592192A (en) 1984-01-07

Family

ID=14581132

Family Applications (1)

Application Number Title Priority Date Filing Date
JP57112215A Pending JPS592192A (en) 1982-06-29 1982-06-29 Character recognizing system

Country Status (1)

Country Link
JP (1) JPS592192A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100383017B1 (en) * 1999-08-06 2003-05-09 가부시끼가이샤 도시바 Pattern string matching device and pattern string matching method

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100383017B1 (en) * 1999-08-06 2003-05-09 가부시끼가이샤 도시바 Pattern string matching device and pattern string matching method

Similar Documents

Publication Publication Date Title
Li et al. Trocr: Transformer-based optical character recognition with pre-trained models
US5343537A (en) Statistical mixture approach to automatic handwriting recognition
El-Yacoubi et al. An HMM-based approach for off-line unconstrained handwritten word modeling and recognition
US7437001B2 (en) Method and device for recognition of a handwritten pattern
Chen et al. A new off-line signature verification method based on graph
Kovalevsky Image pattern recognition
JP4941791B2 (en) Two-dimensional pattern matching method, feature extraction method, apparatus and program used therefor
Fornés et al. Rotation invariant hand-drawn symbol recognition based on a dynamic time warping model
US5392367A (en) Automatic planar point pattern matching device and the matching method thereof
Tomai et al. Discriminatory power of handwritten words for writer recognition
CN107016319A (en) A kind of key point localization method and device
Rajalakshmi et al. Pattern recognition-recognition of handwritten document using convolutional neural networks
US6023529A (en) Handwritten pattern recognizer with selective feature weighting
Fornés et al. A keyword spotting approach using blurred shape model-based descriptors
Schenk et al. Selecting features in on-line handwritten whiteboard note recognition: SFS or SFFS?
Lamghari et al. Template matching for recognition of handwritten Arabic characters using structural characteristics and Freeman code
JPS592192A (en) Character recognizing system
JPS59116885A (en) Character recognition system
JPS5969875A (en) Method for recognizing character
Ye et al. Off-line handwritten signature verification with inflections feature
Srisuk et al. A new shape matching measure for nonlinear distorted object recognition
JP2000137766A (en) Device and method for word recognition
Touj et al. Global feature extraction of off-line arabic handwriting
Williams et al. Word spotting in bitmapped fax documents
Chatwiriya Off-line Thai handwriting recognition in legal amount