JPH1069524A - Character segmentation method - Google Patents

Character segmentation method

Info

Publication number
JPH1069524A
JPH1069524A JP8227181A JP22718196A JPH1069524A JP H1069524 A JPH1069524 A JP H1069524A JP 8227181 A JP8227181 A JP 8227181A JP 22718196 A JP22718196 A JP 22718196A JP H1069524 A JPH1069524 A JP H1069524A
Authority
JP
Japan
Prior art keywords
character
size
characters
image
character string
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
JP8227181A
Other languages
Japanese (ja)
Inventor
Yuji Adachi
祐二 安達
Kazuki Hashimoto
和樹 橋本
Kohei Otsuki
幸平 大槻
Motoi Hashizume
基 橋詰
Noboru Takahashi
暢 高橋
Hiroyuki Hasegawa
裕之 長谷川
Hitoshi Goto
仁 後藤
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
JFE Steel Corp
Hitachi Engineering Co Ltd
Kobe Steel Ltd
JFE Engineering Corp
Nippon Steel Corp
Original Assignee
Hitachi Engineering Co Ltd
Kobe Steel Ltd
Sumitomo Metal Industries Ltd
Kawasaki Steel Corp
NKK Corp
Nippon Kokan Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hitachi Engineering Co Ltd, Kobe Steel Ltd, Sumitomo Metal Industries Ltd, Kawasaki Steel Corp, NKK Corp, Nippon Kokan Ltd filed Critical Hitachi Engineering Co Ltd
Priority to JP8227181A priority Critical patent/JPH1069524A/en
Publication of JPH1069524A publication Critical patent/JPH1069524A/en
Pending legal-status Critical Current

Links

Landscapes

  • Character Input (AREA)

Abstract

PROBLEM TO BE SOLVED: To provide a character segmentation method capable of improving a character recognition rate and character recognition accuracy by segmenting characters so as to be the character frame of a size corresponding to the size even in the case that the sizes of the characters inside images are different. SOLUTION: A character segmentation part 44 projects a binary image in a character string direction, calculates a first reference point, specifies the area of a character string with the first reference point as a center, performs projection in a direction orthogonal to the character string direction in the specified area, calculates plural second reference points and obtains a character interval. An arithmetic expression for computing the size of the character frame based on the character interval is set to the character segmentation part 44 beforehand and the character segmentation part 44 calculates the size (column size × row size) of the character frame by substituting the character interval to the arithmetic expression, segments the character string so as to be the calculated column size with the first reference point as the center and segments the respective characters included in the character string so as to be the calculated row size with the second reference point as the center.

Description

【発明の詳細な説明】DETAILED DESCRIPTION OF THE INVENTION

【0001】[0001]

【発明の属する技術分野】本発明は、物品に付けた数字
及び英字等の文字を撮像し、得られた画像中の文字を認
識する場合において、画像中の文字を一文字ずつ切り出
す方法に関する。
BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to a method for picking up characters such as numbers and letters attached to an article, and cutting out the characters in the image one by one when recognizing the characters in the obtained image.

【0002】[0002]

【従来の技術】製造業にあっては、中間品及び製品にロ
ット番号,製造番号を直接印字し、又はそれらを印字し
たシールを貼着し、文字認識装置が備える撮像装置を製
造ライン又は入出庫ライン等に設置しておき、該撮像装
置によって中間品又は製品に付した文字を読み取り、読
み取った文字からロット番号又は製造番号を認識し、認
識したロット番号又は製造番号を管理用のコンピュータ
に与え、該コンピュータによって製造工程及び在庫の集
中管理が行われている。ロット番号,製造番号等の文字
は、中間品又は製品に直接印字されることが多いが、か
すれ,にじみ及び汚れ等によって印字品質は低い。ま
た、照明度が低く撮像環境が悪い。一方、製造業におけ
る文字認識は、高速・高率・高精度で文字を認識するこ
とが要求されている。そこで、これらの要求に対応すべ
く、次のような文字認識方法が採用されている。
2. Description of the Related Art In the manufacturing industry, lot numbers and serial numbers are directly printed on intermediate products and products, or stickers on which these numbers are printed are attached, and an image pickup device provided with a character recognition device is manufactured or installed on a manufacturing line. It is installed in a retrieval line or the like, reads characters attached to intermediate products or products by the imaging device, recognizes a lot number or production number from the read characters, and sends the recognized lot number or production number to a management computer. The computer performs centralized control of the manufacturing process and inventory. Characters such as lot numbers and serial numbers are often printed directly on intermediate products or products, but the printing quality is low due to blurring, bleeding, dirt, and the like. In addition, the illuminance is low and the imaging environment is poor. On the other hand, the character recognition in the manufacturing industry is required to recognize characters at high speed, at a high rate, and with high accuracy. Therefore, the following character recognition method has been adopted to meet these requirements.

【0003】被撮像物に付した認識対象文字を真上に配
した撮像装置で撮像し、撮像装置が出力したアナログ信
号を所定階調のディジタル信号に変換し、それを画像メ
モリに記憶させる。画像メモリに記憶した文字画像を2
値化して2値化画像を生成し、この2値化画像から予め
設定したサイズの文字枠で一文字ずつ文字を切り出す。
切り出したそれぞれの文字について、辞書に予め登録し
た複数の基準文字パターンとのパターンマッチングを行
って、各基準文字パターンと切り出した文字との一致度
を次の(1)式に基づいてそれぞれ算出し、一致度が高
いものから順番に第1候補文字,第2候補文字とし、第
1候補文字が次の(2)式及び(3)式のいずれをも満
足する場合、切り出した文字は第1候補文字であると判
断し、両式のいずれか一方のみ満足する場合又はいずれ
も満足しない場合、切り出した文字は不読であると判断
する。 一致度=(一致した画素の総数)/(基準文字パターン を構成する画素の総数)…(1) 第1候補文字の一致度≧第1閾値 …(2) 第1候補文字の一致度と 第2候補文字の一致度との差≧第2閾値 …(3)
A recognition target character attached to an object to be imaged is imaged by an imaging device arranged directly above, and an analog signal output by the imaging device is converted into a digital signal of a predetermined gradation, which is stored in an image memory. The character image stored in the image memory is 2
The binarized image is generated by binarization, and characters are cut out from the binarized image one by one in a character frame of a preset size.
The extracted characters are subjected to pattern matching with a plurality of reference character patterns registered in advance in the dictionary, and the degree of coincidence between each reference character pattern and the extracted characters is calculated based on the following equation (1). If the first candidate character satisfies both of the following expressions (2) and (3), the extracted character is the first candidate character and the second candidate character in order from the one with the highest matching degree. If it is determined that the character is a candidate character and only one of the two expressions is satisfied, or if neither is satisfied, the extracted character is determined to be unreadable. Matching degree = (total number of matched pixels) / (total number of pixels forming reference character pattern) (1) Matching degree of first candidate character ≧ first threshold (2) Matching degree of first candidate character Difference from matching degree of two candidate characters ≧ second threshold value (3)

【0004】このような文字認識にあっては、文字を切
り出す際に、切り出した文字と文字枠とのバランスがよ
いことが重要である。
In such character recognition, when cutting out characters, it is important that the cut-out characters and the character frame be well-balanced.

【0005】図7及び図8は従来の文字の切り出し方法
を説明する説明図である。図7(a)の如く、2値化画
像に“1,2,3,4,5”及び“6,7,8,9,
0”の文字列がある場合、2値化画像を構成する各2値
化値を文字列方向に累計する操作を繰り返して文字列方
向の投影を生成すると、図7(b)の如く、各文字列に
応じて2つのピークが得られる。そして、各ピークの中
心を求めて切り出しのための第1基準点とし、各第1基
準点を中心に予め設定された列サイズになるように両文
字列を切り出す。
FIGS. 7 and 8 are explanatory views for explaining a conventional character cutting method. As shown in FIG. 7A, “1, 2, 3, 4, 5” and “6, 7, 8, 9, 9,
When there is a character string of "0", the operation of accumulating the respective binarized values forming the binarized image in the character string direction is repeated to generate a projection in the character string direction, as shown in FIG. 7B. Two peaks are obtained according to the character string, and the center of each peak is obtained and used as a first reference point for cutting out, and both peaks are set so as to have a preset column size around each first reference point. Cut out a character string.

【0006】切り出した文字列の内、例えば“1,2,
3,4,5”について、文字列方向と直交する方向に2
値化値を累計する操作を繰り返して投影を生成すると、
図8(b)の如く、各文字に応じた5つのピークが得ら
れる。そして、各ピークの中心を求めて第2基準点と
し、各第2基準点を中心に予め設定された行サイズにな
るように各文字を切り出すことによって、図8(a)の
如く、“1”,“2”,“3”,“4”,“5”の各文
字を切り出していた。また、“6,7,8,9,0”の
文字列についても前同様の処理を行って“6”,
“7”,“8”,“9”,“0”の各文字を切り出して
いた。
For example, "1, 2, 2, 3"
For 3,4,5 ", 2 in the direction orthogonal to the character string direction
By repeating the operation of accumulating quantified values to generate projections,
As shown in FIG. 8B, five peaks corresponding to each character are obtained. Then, the center of each peak is obtained and set as a second reference point, and each character is cut out so as to have a predetermined line size centered on each second reference point, thereby obtaining "1" as shown in FIG. The characters ",""2,""3,""4," and "5" were cut out. In addition, the same processing as described above is performed on the character string of “6, 7, 8, 9, 0” to obtain “6”,
The characters "7", "8", "9", and "0" were cut out.

【0007】[0007]

【発明が解決しようとする課題】しかしながら、従来の
方法では被撮像物と撮像装置との間の距離の相違によっ
て次のような問題が生じていた。
However, in the conventional method, the following problem has arisen due to the difference in the distance between the object to be imaged and the imaging device.

【0008】図9は、被撮像物を撮像している状態を示
す模式図であり、図中S1 ,S2 ,S3 は被撮像物であ
る。被撮像物S1 ,S2 ,S3 の高さはそれぞれ異なっ
ており、図9にあっては、被撮像物S2 の高さ<被撮像
物S3 の高さ<被撮像物S1,高さである。また、各被
撮像物S1 ,S2 ,S3 の上面の所定位置には同じサイ
ズの認識対象文字が印字してある。被撮像物S1 ,S
2 ,S3 は搬送装置1によって矢符方向に搬送されるよ
うになっており、搬送方向の下流には、搬送装置1の真
上に認識対象文字を撮像する撮像装置21が図示しない支
持部材によって支持固定されている。そして、移動装置
1によって被撮像物S1 ,S2 ,S3 の認識対象文字が
撮像装置21の真下まで搬送されたとき、撮像装置21は各
被撮像物S 1 ,S2 ,S3 の認識対象文字を撮像するよ
うになっている。
FIG. 9 shows a state in which an object is imaged.
FIG.1 , STwo , SThree Is the object to be imaged
You. Object S1 , STwo , SThree Of different heights
In FIG. 9, the object STwo Height <imaged
Thing SThree Height <Object S1, Height. In addition, each
Object S1 , STwo , SThree The same size
The characters to be recognized are printed. Object S1 , S
Two , SThree Is transported in the direction of the arrow by the transport device 1.
The transport device 1 is located downstream of the transport direction.
An image pickup device 21 for picking up a recognition target character is provided with a support (not shown).
It is supported and fixed by a holding member. And mobile devices
1, the object S to be imaged1 , STwo , SThree Characters to be recognized
When transported to a position directly below the imaging device 21, the imaging device 21
Object S 1 , STwo , SThree I will capture the characters to be recognized
Swelling.

【0009】図10は図9に示した撮像装置21によって被
撮像物S1 ,S2 ,S3 の認識対象文字を撮像して得た
画像から文字を切り出した結果を説明する説明図であ
り、(a)は図9に示した被撮像物S1 を撮像した場合
を、(b)は図9に示した被撮像物S2 を撮像した場合
を、(c)は図9に示した被撮像物S3 を撮像した場合
をそれぞれ示している。図中、“1”,“2”,…は画
像から切り出した文字を、また、各文字を囲んでいる実
線は文字枠をそれぞれ示している。
FIG. 10 is an explanatory diagram for explaining a result obtained by extracting characters from an image obtained by imaging the characters to be recognized of the objects S 1 , S 2 , and S 3 by the imaging device 21 shown in FIG. the case of the captured (a) is imaged object S 1 shown in FIG. 9, the case of the captured (b) is imaged object S 2 shown in FIG. 9, illustrating (c) in FIG. 9 shows the case of the captured object to be imaged S 3, respectively. In the figure, “1”, “2”,... Indicate characters cut out from the image, and solid lines surrounding each character indicate character frames.

【0010】高さが異なる被撮像物を一定の高さから撮
像すると、画像内の文字のサイズは、撮像装置から被撮
像物までの距離が短いほど、即ち、被撮像物の高さが高
いほど大きい。そのため、一定サイズの文字枠で各文字
を切り出すと、図10(a)のように、文字枠に文字が重
なる場合、又は図10(b)のように、文字枠のサイズに
対して文字のサイズが小さい場合が生じる。なお、図10
(c)の場合は、文字枠のサイズに対して文字のサイズ
が適当である。一方、文字認識の基準とする基準文字パ
ターンは文字枠のサイズに対して適当なサイズで登録し
てあるため、図10(a)及び(b)の場合にあっては、
基準文字パターンとの一致度が低下し、文字認識精度及
び文字認識率が低下する。
When images of objects having different heights are picked up from a fixed height, the size of characters in the image is such that the shorter the distance from the image pickup device to the object, that is, the higher the height of the object. As large. Therefore, when each character is cut out in a character frame of a fixed size, the character is overlapped with the character frame as shown in FIG. 10A, or as shown in FIG. Sometimes the size is small. Note that FIG.
In the case (c), the size of the character is appropriate for the size of the character frame. On the other hand, since the reference character pattern used as the reference for character recognition is registered at an appropriate size for the size of the character frame, in the case of FIGS. 10 (a) and (b),
The degree of matching with the reference character pattern decreases, and the character recognition accuracy and the character recognition rate decrease.

【0011】本発明はかかる事情に鑑みてなされたもの
であって、その目的とするところは画像内の各文字列
を、文字列の方向と直交する方向に投影して複数のピー
クを得、各ピーク間の距離を算出し、算出した距離から
文字枠のサイズを定め、定めたサイズの文字枠になるよ
うに各文字パターンを切り出すことによって、画像内の
文字のサイズが異なる場合でもそのサイズに対応するサ
イズの文字枠で文字を切り出して、文字認識率及び文字
認識精度を向上させることができる文字切り出し方法を
提供することにある。
The present invention has been made in view of the above circumstances, and has as its object to project a plurality of peaks by projecting each character string in an image in a direction orthogonal to the direction of the character string. Calculate the distance between each peak, determine the size of the character frame from the calculated distance, and cut out each character pattern so that it becomes a character frame of the determined size, even if the size of characters in the image is different, that size An object of the present invention is to provide a character cutout method that can cut out a character in a character frame having a size corresponding to the character size and improve the character recognition rate and the character recognition accuracy.

【0012】[0012]

【課題を解決するための手段】本発明に係る文字切り出
し方法は、1又は複数列の対象文字を撮像して画像を
得、該画像内の1又は複数の文字列に含まれる複数の文
字パターンを、文字列方向及びそれと直交する方向に投
影して文字パターンの形状に応じたピークを得、得られ
たピークに基づいて、適宜サイズの文字枠になるように
各文字パターンを切り出す方法において、各文字列を、
文字列の方向と直交する方向に投影して複数のピークを
得、各ピーク間の距離を算出し、算出した距離から前記
文字枠のサイズを定め、定めたサイズの文字枠になるよ
うに各文字パターンを切り出すことを特徴とする。
According to the character extracting method of the present invention, an image is obtained by capturing one or more columns of target characters, and a plurality of character patterns included in one or more character strings in the image are obtained. In the method of projecting in the character string direction and the direction orthogonal to the character string to obtain a peak according to the shape of the character pattern, and based on the obtained peak, cut out each character pattern so that a character frame of an appropriate size is obtained. Each string is
A plurality of peaks are obtained by projecting in the direction orthogonal to the direction of the character string, the distance between the peaks is calculated, the size of the character frame is determined from the calculated distance, and each character frame is determined to have a predetermined size. Character patterns are cut out.

【0013】図3〜図6は、本発明に係る文字切り出し
方法を説明する説明図である。いま、高さが異なる3つ
の被撮像物に同じサイズの認識対象文字を印字してそれ
を撮像し、図3(a),(b),(c)のような3つの
画像が得られたとする。なお、各画像は複数の画素から
構成されている。図3の如く、各画像には“1,2,
3,4,5”の文字列が表示されており、文字列を構成
する各文字の相対的な大きさは、(b)<(c)<
(a)である。
FIGS. 3 to 6 are explanatory diagrams for explaining a character segmentation method according to the present invention. Now, it is assumed that a recognition target character having the same size is printed on three objects having different heights and is imaged, and three images as shown in FIGS. 3A, 3B, and 3C are obtained. I do. Each image is composed of a plurality of pixels. As shown in FIG. 3, each image has "1, 2,
3, 4, 5 "are displayed, and the relative size of each character constituting the character string is (b) <(c) <
(A).

【0014】図4(a)(b)(c)のように、各文字
列を含む所定領域において、各文字列を文字列の方向と
直交する方向に投影し、各文字に対応するピーク複数を
得る。そして、各ピークの中央位置をそれぞれ求め、相
隣る中央位置の間の画素数の平均値を算出して文字間隔
を得る。例えば、図5(a)では文字間隔は44画素で
あり、(b)では文字間隔は40画素であり、(c)で
は文字間隔は42画素である。
As shown in FIGS. 4 (a), 4 (b) and 4 (c), in a predetermined area including each character string, each character string is projected in a direction orthogonal to the direction of the character string, and a plurality of peaks corresponding to each character are projected. Get. Then, the central position of each peak is obtained, and the average value of the number of pixels between adjacent central positions is calculated to obtain the character spacing. For example, in FIG. 5A, the character interval is 44 pixels, in FIG. 5B, the character interval is 40 pixels, and in FIG. 5C, the character interval is 42 pixels.

【0015】このようにして算出した文字間隔は、撮像
装置と被撮像物との間の距離が異っている場合における
画像内の文字の大きさに相関している。そこで、サンプ
ル材を用いて、画像内の文字間隔と、その文字を適正に
切り出せる文字枠のサイズとの関係式を予め求めてお
き、その関係式に前述した如く求めた各文字間隔をそれ
ぞれ代入して、適正な文字枠のサイズを算出する。前述
した関係式は撮像装置の配置等によって異なるが、例え
ば次の(1)式及び(2)式であった場合、図5(a)
(b)(c)の各文字を適正に切り出すことができる文
字枠のサイズ(列サイズ×行サイズ)は、図5(a)に
あっては列サイズ×行サイズ=44画素×33画素であ
り、図5(b)にあっては列サイズ×行サイズ=40画
素×30画素であり、図5(c)にあっては列サイズ×
行サイズ=42画素×32画素である。 列サイズ=文字間隔 …(1) 行ザイズ=3/4×文字間隔 …(2)
The character spacing calculated in this way is correlated with the size of characters in an image when the distance between the imaging device and the object to be imaged is different. Therefore, using a sample material, a relational expression between the character spacing in the image and the size of the character frame that can appropriately cut out the character is determined in advance, and the character spacing determined as described above is calculated using the relational expression. Substituting to calculate an appropriate character frame size. Although the above-mentioned relational expression varies depending on the arrangement of the imaging device and the like, for example, when the following expressions (1) and (2) are used, FIG.
The character frame size (column size × row size) from which each character of (b) and (c) can be properly cut out is column size × row size = 44 pixels × 33 pixels in FIG. In FIG. 5B, column size × row size = 40 pixels × 30 pixels, and in FIG. 5C, column size × row size
Row size = 42 pixels × 32 pixels. Column size = character spacing ... (1) Row size = 3/4 x character spacing ... (2)

【0016】これらの文字枠によって各文字列から文字
をそれぞれ切り出すと、図6(a)(b)(c)のよう
に、画像内の文字の大きさに対応した文字枠で各文字を
切り出すことができる。これによって、被撮像物の高さ
の相違に拘らず、画像内の文字を適正に切り出すことが
でき、文字認識精度及び文字認識率が低下が防止され
る。
When characters are cut out from each character string by using these character frames, each character is cut out in a character frame corresponding to the size of the character in the image as shown in FIGS. 6 (a), 6 (b) and 6 (c). be able to. Thereby, regardless of the difference in the height of the object to be imaged, characters in the image can be appropriately cut out, and a reduction in character recognition accuracy and a character recognition rate can be prevented.

【0017】[0017]

【発明の実施の形態】以下、本発明の実施の形態を図面
に基づいて具体的に説明する。図1は本発明の実施に使
用する文字認識装置の構成を示すブロック図であり、図
中Sは鋼板,鋼管等の被撮像物である。被撮像物Sは搬
送装置1によって矢符方向に搬送されるようになってお
り、該被撮像物Sの所定位置にはロット番号又は製品番
号等の認識対象文字が所定のサイズで印字してある。被
撮像物Sの搬送領域には、被撮像物Sの認識対象文字が
印字される部分と対向して撮像装置2が図示しない支持
部材によって支持されており、該撮像装置2によって被
撮像物Sの認識対象文字が撮像される。撮像装置2が撮
像した文字画像は、アナログ/ディジタル(A/D)変
換器3に与えられ、A/D変換器3によって所定階調の
濃淡画像に変換されて、コンピュータ4に備えられた中
央演算装置(CPU)41によって画像メモリ42に与えら
れ、そこに記憶される。
Embodiments of the present invention will be specifically described below with reference to the drawings. FIG. 1 is a block diagram showing the configuration of a character recognition device used in the embodiment of the present invention. In FIG. 1, S denotes an object to be imaged such as a steel plate or a steel pipe. The object S is conveyed in the direction of the arrow by the conveyance device 1, and characters to be recognized such as a lot number or a product number are printed in a predetermined size on a predetermined position of the object S. is there. In the transport area of the object S, the imaging device 2 is supported by a support member (not shown) so as to face a portion of the object S on which the recognition target character is printed. Is recognized. The character image picked up by the image pickup device 2 is supplied to an analog / digital (A / D) converter 3, which converts the character image into a gray-scale image having a predetermined gradation by the A / D converter 3. An arithmetic unit (CPU) 41 gives the image data to the image memory 42 and stores it there.

【0018】画像メモリ42に記憶された濃淡画像は2値
化部43によって2値化処理される。得られた2値化画像
に基づいて、文字切り出し部44は後述するように求めた
文字枠で一文字ずつ文字を切り出し、切り出した各文字
を前処理部45に与える。コンピュータ4は、パターンマ
ッチングに用いる複数の基準文字パターンが登録してあ
る辞書46を備えており、前処理部45は与えられた文字と
辞書46内の基準文字パターンとを比較できるように、切
り出した文字の方向,サイズ,線幅等を正規化し、それ
をパターンマッチング部47に与える。パターンマッチン
グ部47は、正規化された文字と辞書46内の各基準文字パ
ターンとの一致度を次の(1)式によってそれぞれ算出
し、その算出結果を判断部48に与える。 一致度=(一致した画素の総数)/(基準文字パターン を構成する画素の総数)…(1)
The grayscale image stored in the image memory 42 is binarized by a binarization unit 43. Based on the obtained binarized image, the character cutout unit 44 cuts out characters one by one in a character frame obtained as described later, and supplies each cutout character to the preprocessing unit 45. The computer 4 has a dictionary 46 in which a plurality of reference character patterns used for pattern matching are registered. The preprocessing unit 45 cuts out a given character and compares it with a reference character pattern in the dictionary 46. The direction, size, line width, and the like of the character are normalized, and are provided to the pattern matching unit 47. The pattern matching unit 47 calculates the degree of coincidence between the normalized character and each reference character pattern in the dictionary 46 according to the following equation (1), and gives the calculation result to the determination unit 48. Matching degree = (total number of matched pixels) / (total number of pixels forming reference character pattern) (1)

【0019】判断部48は辞書46の各基準文字パターンの
内、一致度が高いものから順番に第1候補文字,第2候
補文字,…として、第1候補文字の一致度M1 と第2候
補文字の一致度M2 との差を算出し、第1候補文字の一
致度M1 と予め設定された第1閾値D1 とを、また第1
候補文字の一致度と第2候補文字の一致度との差(M 1
−M2 )と予め設定された第2閾値D2 とをそれぞれ比
較する。
The judgment unit 48 determines the reference character pattern of the dictionary 46
Among them, the first candidate character and the second candidate character in descending order of matching degree
As the complement character,..., The matching degree M of the first candidate character1And the second weather
Complementary character matching degree MTwoIs calculated, and one of the first candidate characters is calculated.
Severity M1And a preset first threshold D1And the first
The difference between the matching degree of the candidate character and the matching degree of the second candidate character (M 1
-MTwo) And a preset second threshold value DTwoAnd the respective
Compare.

【0020】判断部48は、比較した結果、次の(2)式
かつ(3)式を満足するか否かを判断し、満足すると判
断した場合、切り出した文字は第1候補文字であると決
定し、満足しないと判断した場合、不読であると判断す
る。そして、出力部49はそれらの結果を出力する。 第1候補文字の一致度≧第1閾値 …(2) 第1候補文字の一致度と 第2候補文字の一致度との差≧第2閾値 …(3)
The judging section 48 judges whether the following expressions (2) and (3) are satisfied as a result of the comparison. If it is determined that the expression is satisfied, the extracted character is determined to be the first candidate character. If it is determined and it is determined that the user is not satisfied, it is determined that the content is not read. Then, the output unit 49 outputs those results. Matching degree of first candidate character ≧ first threshold (2) Difference between matching degree of first candidate character and matching degree of second candidate character ≧ second threshold (3)

【0021】図2は、図1に示した文字切り出し部44に
よる文字の切り出し手順を示すフローチャートである。
文字切り出し部44は、2値化画像を文字列方向に投影し
(ステップS1)、得られたピークの中央位置を算出し
(ステップS2)、その位置を第1基準点とする。第1
基準点を算出すると、文字切り出し部44は、算出した第
1基準点を中心予め定められた列サイズとなるように文
字列の領域を特定し(ステップS3)、特定した領域に
おいて、文字列方向と直交する方向に投影して(ステッ
プS4)、その文字列に含まれる複数の文字に対応した
ピークをそれぞれ得る。文字切り出し部44は、各ピーク
の中央位置を算出して(ステップS5)、その位置をそ
れぞれ第2基準点とし、相隣る2基準点の間の距離の平
均を算出して文字間隔を求める(ステップS6)。
FIG. 2 is a flowchart showing a procedure for extracting characters by the character extracting section 44 shown in FIG.
The character cutout unit 44 projects the binarized image in the character string direction (step S1), calculates the center position of the obtained peak (step S2), and sets that position as the first reference point. First
When the reference point is calculated, the character cutout unit 44 specifies a character string area so that the calculated first reference point is centered at a predetermined column size (step S3). (Step S4), and peaks corresponding to a plurality of characters included in the character string are obtained. The character cutout unit 44 calculates the center position of each peak (step S5), sets the position as the second reference point, calculates the average of the distance between two adjacent reference points, and obtains the character spacing. (Step S6).

【0022】文字切り出し部44には、文字間隔に基づい
て文字枠のサイズを演算する演算式が予め設定してあ
り、文字切り出し部44はその演算式にステップS6にて
求めた文字間隔を代入して文字枠のサイズ(列サイズ×
行サイズ)を算出する(ステップS7)。そして、文字
切り出し部44は、ステップS2にて求めた第1基準点を
中心に、ステップS7にて求めた列サイズになるように
文字列を切り出し(ステップS8)、切り出した文字列
について、ステップS5にて求めた各第2基準点を中心
にステップS7にて求めた行サイズになるように、その
文字列に含まれる各文字をそれぞれ切り出す(ステップ
S9)。これによって、被撮像物の高さの相違に拘ら
ず、画像内の文字を適正に切り出すことができ、文字認
識精度及び文字認識率の低下が防止される。
An arithmetic expression for calculating the size of the character frame based on the character interval is set in the character cutout unit 44 in advance, and the character cutout unit 44 substitutes the character interval obtained in step S6 into the arithmetic expression. And the size of the character frame (column size x
(Line size) is calculated (step S7). Then, the character cutout unit 44 cuts out the character string around the first reference point obtained in step S2 so as to have the column size obtained in step S7 (step S8). Each character included in the character string is cut out so that the line size obtained in step S7 is centered on each second reference point obtained in S5 (step S9). Thereby, regardless of the difference in the height of the object to be imaged, characters in the image can be properly cut out, and a reduction in character recognition accuracy and a character recognition rate can be prevented.

【0023】[0023]

【発明の効果】本発明に係る文字切り出し方法にあって
は、被撮像物の高さの相違に拘らず、画像内の文字を適
正な文字枠で一字ずつ切り出すことができるため、文字
認識精度及び文字認識率の低下が防止される等、本発明
は優れた効果を奏する。
According to the character extracting method of the present invention, characters in an image can be extracted one by one in an appropriate character frame regardless of the height of the object to be imaged. The present invention has excellent effects, such as preventing a decrease in accuracy and character recognition rate.

【図面の簡単な説明】[Brief description of the drawings]

【図1】本発明の実施に使用する文字認識装置の構成を
示すブロック図である。
FIG. 1 is a block diagram showing a configuration of a character recognition device used for implementing the present invention.

【図2】図1に示した文字切り出し部による文字の切り
出し手順を示すフローチャートである。
FIG. 2 is a flowchart illustrating a character extracting procedure performed by the character extracting unit illustrated in FIG. 1;

【図3】本発明に係る文字切り出し方法を説明する説明
図である。
FIG. 3 is an explanatory diagram illustrating a character segmentation method according to the present invention.

【図4】本発明に係る文字切り出し方法を説明する説明
図である。
FIG. 4 is an explanatory diagram illustrating a character segmentation method according to the present invention.

【図5】本発明に係る文字切り出し方法を説明する説明
図である。
FIG. 5 is an explanatory diagram illustrating a character segmentation method according to the present invention.

【図6】本発明に係る文字切り出し方法を説明する説明
図である。
FIG. 6 is an explanatory diagram illustrating a character segmentation method according to the present invention.

【図7】従来の文字切り出し方法を説明する説明図であ
る。
FIG. 7 is an explanatory diagram illustrating a conventional character segmentation method.

【図8】従来の文字切り出し方法を説明する説明図であ
る。
FIG. 8 is an explanatory diagram illustrating a conventional character segmentation method.

【図9】被撮像物を撮像している状態を示す模式図であ
る。
FIG. 9 is a schematic diagram illustrating a state in which an object is imaged.

【図10】図9に示した撮像装置によって被撮像物の認
識対象文字を撮像して得た画像から文字を切り出した結
果を説明する説明図である。
10 is an explanatory diagram illustrating a result obtained by extracting characters from an image obtained by capturing a recognition target character of an object to be captured by the imaging device illustrated in FIG. 9;

【符号の説明】[Explanation of symbols]

2 撮像装置 4 コンピュータ 42 画像メモリ 44 文字切り出し部 47 パターンマッチング部 48 判断部 2 Imaging device 4 Computer 42 Image memory 44 Character cutout unit 47 Pattern matching unit 48 Judgment unit

───────────────────────────────────────────────────── フロントページの続き (71)出願人 000001199 株式会社神戸製鋼所 兵庫県神戸市中央区脇浜町1丁目3番18号 (71)出願人 390023928 日立エンジニアリング株式会社 茨城県日立市幸町3丁目2番1号 (72)発明者 安達 祐二 大阪府大阪市中央区北浜4丁目5番33号 住友金属工業株式会社内 (72)発明者 橋本 和樹 大阪府大阪市中央区北浜4丁目5番33号 住友金属工業株式会社内 (72)発明者 大槻 幸平 大阪府大阪市中央区北浜4丁目5番33号 住友金属工業株式会社内 (72)発明者 橋詰 基 東京都千代田区丸の内一丁目1番2号 日 本鋼管株式会社内 (72)発明者 高橋 暢 岡山県倉敷市水島川崎通1丁目(番地な し) 川崎製鉄株式会社水島製鉄所内 (72)発明者 長谷川 裕之 兵庫県加古川市金沢町1番地 株式会社神 戸製鋼所加古川製鉄所内 (72)発明者 後藤 仁 茨城県日立市幸町三丁目2番1号 日立エ ンジニアリング株式会社内 ──────────────────────────────────────────────────の Continued from the front page (71) Applicant 000001199 Kobe Steel, Ltd. 1-3-18, Wakihama-cho, Chuo-ku, Kobe-shi, Hyogo (71) Applicant 390023928 Hitachi Engineering Co., Ltd. 3-chome, Sachicho, Hitachi-shi, Ibaraki 2-1 (72) Inventor Yuji Adachi 4-5-33 Kitahama, Chuo-ku, Osaka City, Osaka Prefecture Inside Sumitomo Metal Industries, Ltd. (72) Inventor Kazuki Hashimoto 4-5-33 Kitahama, Chuo-ku, Osaka City, Osaka Prefecture Within Sumitomo Metal Industries, Ltd. (72) Kohei Otsuki, Inventor 4-53, Kitahama, Chuo-ku, Osaka City, Osaka Prefecture Inside Sumitomo Metal Industries, Ltd. Inside Nippon Kokan Co., Ltd. (72) Inventor Nobu Takahashi 1-chome, Kawasaki-dori, Mizushima, Kurashiki-shi, Okayama Pref. 2) Inventor Hiroyuki Hasegawa 1 Kanazawacho, Kakogawa City, Hyogo Prefecture Inside the Kobe Steel Works Kakogawa Works (72) Inventor Jin Goto 3-2-1 Sachimachi, Hitachi City, Hitachi City, Ibaraki Prefecture Within Hitachi Engineering Co., Ltd.

Claims (1)

【特許請求の範囲】[Claims] 【請求項1】 1又は複数列の対象文字を撮像して画像
を得、該画像内の1又は複数の文字列に含まれる複数の
文字パターンを、文字列方向及びそれと直交する方向に
投影して文字パターンの形状に応じたピークを得、得ら
れたピークに基づいて、適宜サイズの文字枠になるよう
に各文字パターンを切り出す方法において、 各文字列を、文字列の方向と直交する方向に投影して複
数のピークを得、各ピーク間の距離を算出し、算出した
距離から前記文字枠のサイズを定め、定めたサイズの文
字枠になるように各文字パターンを切り出すことを特徴
とする文字切り出し方法。
An image is obtained by capturing one or more rows of target characters, and projecting a plurality of character patterns included in one or more character strings in the image in a character string direction and a direction orthogonal thereto. In the method of obtaining a peak corresponding to the shape of the character pattern and cutting out each character pattern so as to form a character frame of an appropriate size based on the obtained peak, each character string is set in a direction orthogonal to the direction of the character string. To obtain a plurality of peaks, calculate the distance between each peak, determine the size of the character frame from the calculated distance, cut out each character pattern so that the character frame of the determined size. Character extraction method.
JP8227181A 1996-08-28 1996-08-28 Character segmentation method Pending JPH1069524A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP8227181A JPH1069524A (en) 1996-08-28 1996-08-28 Character segmentation method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP8227181A JPH1069524A (en) 1996-08-28 1996-08-28 Character segmentation method

Publications (1)

Publication Number Publication Date
JPH1069524A true JPH1069524A (en) 1998-03-10

Family

ID=16856765

Family Applications (1)

Application Number Title Priority Date Filing Date
JP8227181A Pending JPH1069524A (en) 1996-08-28 1996-08-28 Character segmentation method

Country Status (1)

Country Link
JP (1) JPH1069524A (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPWO2004015619A1 (en) * 2002-08-07 2005-12-02 松下電器産業株式会社 Character recognition processing device, character recognition processing method, and portable terminal device
JP2008204203A (en) * 2007-02-20 2008-09-04 Kobe Steel Ltd Image processing method of character string image, and program and device therefor
US8213748B2 (en) 2008-02-26 2012-07-03 Fuji Xerox Co., Ltd. Generating an electronic document with reference to allocated font corresponding to character identifier from an image
US8411955B2 (en) 2007-02-21 2013-04-02 Fuji Xerox Co., Ltd. Image processing apparatus, image processing method and computer-readable medium
WO2013132709A1 (en) 2012-03-05 2013-09-12 オムロン株式会社 Image processing method for character recognition, and character recognition device and program using this method

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPWO2004015619A1 (en) * 2002-08-07 2005-12-02 松下電器産業株式会社 Character recognition processing device, character recognition processing method, and portable terminal device
JP4566740B2 (en) * 2002-08-07 2010-10-20 パナソニック株式会社 Mobile terminal device
US7903875B2 (en) 2002-08-07 2011-03-08 Panasonic Corporation Character recognition processing device, character recognition processing method, and mobile terminal device
JP2008204203A (en) * 2007-02-20 2008-09-04 Kobe Steel Ltd Image processing method of character string image, and program and device therefor
US8411955B2 (en) 2007-02-21 2013-04-02 Fuji Xerox Co., Ltd. Image processing apparatus, image processing method and computer-readable medium
US8213748B2 (en) 2008-02-26 2012-07-03 Fuji Xerox Co., Ltd. Generating an electronic document with reference to allocated font corresponding to character identifier from an image
WO2013132709A1 (en) 2012-03-05 2013-09-12 オムロン株式会社 Image processing method for character recognition, and character recognition device and program using this method
US9704052B2 (en) 2012-03-05 2017-07-11 Omron Corporation Image processing method for character recognition, character recognition apparatus using this method, and program

Similar Documents

Publication Publication Date Title
US5276315A (en) Method and apparatus for processing low resolution images of degraded bar code symbols
US5487115A (en) Method and apparatus for determining the fine angular orientation of bar code symbols in two-dimensional CCD images
JP2010123090A (en) Method and device for recognizing character string
US5748804A (en) Method and apparatus for processing images with symbols with dense edges
CN112508913A (en) Cable section edge detection method based on image detection
JP4062987B2 (en) Image area dividing method, image area dividing apparatus, and image area dividing program
US5438636A (en) Apparatus for simultaneously convolving multiple digital binary images using a single convolver with a binary mask to determine pixel densities
JPH1069524A (en) Character segmentation method
JPH02306386A (en) Character recognizing device
KR100295360B1 (en) Image Processing Method Using Shading Algorithm
KR102457712B1 (en) System and Method for Recognizing Double Loading of Baggage
JP4492258B2 (en) Character and figure recognition and inspection methods
JPH0256688A (en) Character segmenting device
JP3066448B2 (en) Character extraction method
JP2005250786A (en) Image recognition method
US7203363B2 (en) String extraction method for image based on multiple stroke width patterns matching
JP3170622B2 (en) Character recognition method and device
JP2972516B2 (en) Print pattern inspection device for inkjet printer
JP3705216B2 (en) Character entry frame detection method, character entry frame detection device and program
JP4089807B2 (en) Bar code recognition method, apparatus, and program
EP0797804B1 (en) Method and apparatus for simultaneously convolving multiple digital binary images using a single convolver with a binary mask to determine pixel densities
JP2008107946A (en) Method and apparatus for determining character string area of image
JP4231476B2 (en) An image processing method for collating an input image inputted with a printed matter or a sticker label of a container with a master image registered in advance
JPH04268987A (en) Character recognizing device
JPH06282684A (en) Character recognizing device