WO2010150383A1 - Recognition device - Google Patents

Recognition device Download PDF

Info

Publication number
WO2010150383A1
WO2010150383A1 PCT/JP2009/061615 JP2009061615W WO2010150383A1 WO 2010150383 A1 WO2010150383 A1 WO 2010150383A1 JP 2009061615 W JP2009061615 W JP 2009061615W WO 2010150383 A1 WO2010150383 A1 WO 2010150383A1
Authority
WO
WIPO (PCT)
Prior art keywords
vector
feature vector
unit
stroke
identification
Prior art date
Application number
PCT/JP2009/061615
Other languages
French (fr)
Japanese (ja)
Inventor
洋次郎 登内
Original Assignee
株式会社東芝
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 株式会社東芝 filed Critical 株式会社東芝
Priority to PCT/JP2009/061615 priority Critical patent/WO2010150383A1/en
Publication of WO2010150383A1 publication Critical patent/WO2010150383A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/32Digital ink

Definitions

  • the present invention relates to a technique for recognizing characters from input strokes.
  • character recognition is performed for one stroke obtained by combining an actual stroke and a virtual stroke.
  • a real weight is given a large weight (thickening the handwriting), and a virtual stroke is given a small weight (thinning the handwriting).
  • a technique for recognizing a feature vector is disclosed.
  • the feature vector is extracted and recognized without distinguishing the actual stroke and the virtual stroke. Therefore, there is a possibility of being misrecognized.
  • the present invention has been made in view of the above circumstances, and an object of the present invention is to provide a recognition device that improves recognition performance by clearly distinguishing and recognizing a real stroke and a virtual stroke.
  • the recognition apparatus includes an acquisition unit that acquires coordinate data of handwriting input by the user from the input unit in chronological order, and before pen-up and before pen-up is performed.
  • a creation unit that creates a virtual stroke in which the actual stroke as a handwriting is created, and the coordinate data of the end point of the actual stroke and the coordinate data of the start point of the next actual stroke of the actual stroke are virtually connected by a line segment;
  • the feature vector of the actual stroke is extracted for each divided region obtained by dividing the region where the handwriting exists, and a first feature vector composed of each extracted feature vector is obtained, and the feature vector of the virtual stroke is obtained for each divided region.
  • An extraction unit for extracting and obtaining a second feature vector composed of the extracted feature vectors, and corresponding to the first feature vector and the identification code;
  • the handwriting is recognized using a distance between a vector corresponding to the actual stroke among vectors and a distance between a vector corresponding to the virtual stroke among vectors corresponding to the second feature vector and the identification code.
  • a recognizing unit for extracting and obtaining a second feature vector composed of the extracted feature vectors, and corresponding to the first feature vector and the identification code
  • the recognition performance can be improved by clearly distinguishing the actual stroke from the virtual stroke.
  • the figure which shows the recognition apparatus of 1st Embodiment The figure which shows the example of a real stroke and a virtual stroke. The figure which shows the example of a 1st feature vector. The figure which shows the example of a 2nd feature vector. The figure which shows the example of the dictionary memorize
  • the first feature vector of the actual stroke and the second feature vector of the virtual stroke are extracted separately.
  • recognition processing is performed using feature vectors that are combined while leaving the features of each extracted feature vector.
  • FIG. 1 is a block diagram showing an example of the configuration of the recognition apparatus 1 of the present embodiment.
  • the recognition apparatus 1 includes an input unit 10, a display unit 20, a storage unit 30, an acquisition unit 40, a creation unit 50, an extraction unit 60, a combination unit 70, and a recognition unit 80. And a display control unit 90.
  • the input unit 10 allows a user to input handwriting from a predetermined device such as a pen or a finger.
  • a predetermined device such as a pen or a finger.
  • it can be realized by an existing coordinate input device such as a touch pad, a touch panel, or a tablet.
  • the display unit 20 displays a recognition result of the recognition unit 80 described later and the like according to an instruction from the display control unit 90 described later. It can be realized by an existing display device such as a liquid crystal display, a plasma display, an organic EL display, or a touch panel display.
  • the storage unit 30 stores information used for various processes performed by the recognition device 1. This can be realized by an existing storage medium that can be stored magnetically, electrically, or optically.
  • the storage unit 30 includes a dictionary storage unit 32. Details of the dictionary storage unit 32 will be described later.
  • the acquisition unit 40 acquires the coordinate data of the handwriting input to the input unit 10 in chronological order.
  • the acquisition unit 40 acquires handwriting coordinate data at regular time intervals while a pen or the like is in contact with the input surface of the input unit 10.
  • the creation unit 50 creates an actual stroke that is a handwriting input from the input unit 10 and before the pen-up is performed. At the same time, a virtual stroke is created in which the coordinate data of the end point of the actual stroke and the coordinate data of the start point of the next actual stroke of the actual stroke are virtually connected by a line segment.
  • the actual stroke is a handwriting for one stroke, and refers to a handwriting from when the pen or the like touches the input surface of the input unit 10 until it leaves (from before pen-down to before pen-up).
  • the creation unit 50 creates an actual stroke by connecting the coordinate data acquired from when the pen or the like comes into contact with the input surface of the input unit 10 until the pen leaves the line, in a chronological order.
  • the coordinate data from the start point to the end point of the i-th actual stroke is (X i [1], Y i [1]), (X i [2], Y i [2]), ..., (X i [N i ], Y i [N i ]).
  • the creation unit 50 creates an actual stroke by connecting these coordinate data with line segments in order of time series.
  • i is a natural number.
  • Ni is a natural number of 1 or more, and indicates the number of coordinate data of the i-th actual stroke.
  • the creation unit 50 creates a virtual stroke in which the coordinate data of the end point of the actual stroke and the coordinate data of the start point of the next actual stroke after the actual stroke are virtually connected by a line segment.
  • the coordinate data of the end point of the i-th actual stroke is represented as (X i [N i ], Y i [N i ])
  • the coordinate data of the start point of the i + 1-th actual stroke is (X i + 1 [1], Y i + 1 [1]).
  • the creating unit 50 virtually connects the two coordinate data with a line segment to create an i-th virtual stroke (a virtual stroke between the i-th actual stroke and the i + 1-th actual stroke).
  • the creation unit 50 described an example in which a virtual stroke is created by virtually connecting two end points of an actual stroke and the start point of the next actual stroke with a line segment.
  • a virtual stroke may be created by interpolating between two points and virtually connecting three or more points with line segments.
  • the interpolation may be linear interpolation or interpolation considering the continuity of the direction change at the start point and the end point.
  • FIG. 2 is a diagram illustrating an example of an actual stroke and a virtual stroke created by the creation unit 50.
  • dots indicate coordinate data
  • solid lines connecting the dots indicate actual strokes
  • broken arrows indicate virtual strokes.
  • the extraction unit 60 extracts the feature vector of the actual stroke for each divided region obtained by dividing the region where the handwriting input to the input unit 10 exists, and obtains the first feature vector composed of the extracted feature vectors. obtain. Further, a feature vector of the virtual stroke is extracted for each divided region, and a second feature vector composed of each extracted feature vector is obtained.
  • the extraction unit 60 extracts the appearance frequency distribution in the direction between the coordinate data of the actual stroke for each divided area.
  • a first feature vector is obtained by arranging the extracted appearance frequency distributions in one column. Further, the appearance frequency distribution in the direction between the coordinate data of the virtual strokes is extracted for each divided region.
  • a second feature vector is obtained by arranging the extracted appearance frequency distributions in one column.
  • the distribution of direction component density features will be described as an example of the appearance frequency distribution.
  • the direction component density feature is represented by a three-dimensional array F [d] [x] [y] indicating the x direction, the y direction, and the direction between the coordinate data (direction of actual stroke or virtual stroke).
  • the array F [d] [x] [y] is (F [ 1] [1] [1], F [1] [1] [2],..., F [1] [1] [Ny], F [1] [2] [1],. [Nx] [Ny-1], F [D] [Nx] [Ny]) Nx * Ny * D-dimensional vectors.
  • Nx, Ny, and D are all natural numbers of 2 or more.
  • the extraction unit 60 individually calculates the direction component density features of the real stroke and the virtual stroke to obtain the first feature vector and the second feature vector, respectively.
  • direction component density feature for example, "" Feature selection type character recognition by minimum classification error learning ", IEICE Transactions. D-II, Information / System, II-Information Processing, Vol. 12, December 1998, p.2749-2756 ”can be used.
  • FIG. 3 is a diagram illustrating an example of the first feature vector extracted by the extraction unit 60. Specifically, the directional component density characteristics of the actual stroke shown in FIG. 2 are shown.
  • FIG. 4 is a diagram showing an example of the second feature vector extracted by the extraction unit 60, and more specifically shows the direction component density feature of the virtual stroke shown in FIG.
  • the array F is It is represented by a 128-dimensional vector.
  • the divided areas may be determined in advance, or may be divided into a predetermined number (specifically, values of Nx and Ny) by the extraction unit 60 according to the area where the handwriting exists. Good.
  • the combining unit 70 combines the first feature vector and the second feature vector extracted by the extracting unit 60 to obtain a combined vector.
  • the combining unit 70 obtains a combined vector by arranging the first feature vector and the second feature vector in one column.
  • an array F indicating the first feature vector is an L-dimensional vector (a [1], a [2],..., A [L]).
  • An array F indicating the second feature vector is an M-dimensional vector (b [1], b [2],..., B [M]).
  • L and M are both natural numbers of 8 or more.
  • the combining unit 70 arranges the first feature vector and the second feature vector in one column, and (a [1], a [2],..., A [L], b [1], b [ 2],..., B [M]) to obtain an L + M-dimensional coupling vector.
  • the combining unit 70 may weight each vector element when combining the first feature vector and the second feature vector.
  • the dictionary storage unit 32 stores, for each identification code, an identification vector obtained by combining the first feature vector and the second feature vector of the identification code in association with each other.
  • FIG. 5 shows an example of a dictionary stored in the dictionary storage unit 32.
  • a character code SJIS
  • SJIS character code
  • the identification vectors are combined by a method similar to the method of combining the combination vectors.
  • the recognition unit 80 recognizes the handwriting input to the input unit 10 using the first feature vector and the second feature vector extracted by the extraction unit 60.
  • the recognition unit 80 calculates the distance between the combined vector combined by the combining unit 70 and the identification vector stored in the dictionary storage unit 32.
  • the identification code associated with the identification vector that minimizes the calculated distance is recognized as the handwriting input to the input unit 10. That is, the distance between the first feature vector and the vector corresponding to the actual stroke of the identification vector stored in the dictionary storage unit 32 is calculated. Further, the distance between the second feature vector and the vector corresponding to the virtual stroke of the identification vector stored in the dictionary storage unit 32 is calculated. A comprehensive judgment is made on the two calculated distances to identify the character.
  • the coupling vector is an N-dimensional vector (s [1], s [2],..., S [N]), and the identification vector is (t [1], t [2],..., T [N]. ),
  • the recognition unit 80 calculates the distance between the two vectors using the Euclidean distance R shown in Equation (1). Note that N is a natural number of 16 or more.
  • the display control unit 90 displays the recognition result of the recognition unit 80 on the display unit 20. Specifically, the display control unit 90 causes the recognition unit 80 to display characters or the like indicated by the identification code recognized as the handwriting input to the input unit 10.
  • the acquisition unit 40, the creation unit 50, the extraction unit 60, the combination unit 70, the recognition unit 80, and the display control unit 90 can be realized by an existing arithmetic device such as a CPU.
  • FIG. 6 is a flowchart illustrating an example of a flow of a recognition process performed by the recognition apparatus 1 according to the present embodiment.
  • step S100 handwriting is input to the input unit 10 from a pen or the like.
  • step S102 the acquisition unit 40 acquires the coordinate data of the handwriting input to the input unit 10 in chronological order.
  • step S104 the creation unit 50 creates an actual stroke that is a stroke constituting the input handwriting.
  • step S106 the creation unit 50 creates a virtual stroke in which the coordinate data of the end point of the actual stroke and the coordinate data of the start point of the next actual stroke are virtually connected by a line segment.
  • step S108 the extraction unit 60 extracts a feature vector of the actual stroke for each divided region obtained by dividing the region where the input handwriting exists, and obtains a first feature vector composed of each extracted feature vector.
  • step S110 the extraction unit 60 extracts a feature vector of the virtual stroke for each divided region, and obtains a second feature vector composed of the extracted feature vectors.
  • step S112 the combining unit 70 combines the first feature vector and the second feature vector extracted by the extracting unit 60 to obtain a combined vector.
  • step S114 the recognition unit 80 calculates the distance between the combined vector combined by the combining unit 70 and the identification vector stored in the dictionary storage unit 32.
  • the identification code associated with the identification vector that minimizes the calculated distance is recognized as the handwriting input to the input unit 10.
  • step S116 the display control unit 90 causes the display unit 20 to display the recognition result of the recognition unit 80.
  • FIG. 7 is a block diagram illustrating an example of the configuration of the recognition apparatus 101 according to the present embodiment.
  • the recognition apparatus 101 illustrated in FIG. 7 does not include the dictionary data stored in the dictionary storage unit 132 of the storage unit 130, the processing content of the recognition unit 180, and the combining unit 70. Is different.
  • the dictionary storage unit 132 stores, for each identification code, the first identification vector that is the first feature vector of the identification code and the second identification vector that is the second feature vector of the identification code in association with each other.
  • FIG. 8 shows an example of a dictionary stored in the dictionary storage unit 132.
  • a character code SJIS
  • a first identification vector of the character code and a second identification vector of the character code are associated with each other.
  • P11, P21,... Indicate the first identification vector of the character code, and P12, P22,.
  • the recognizing unit 180 determines the distance between the first feature vector extracted by the extracting unit 60 and the first identification vector stored in the dictionary storage unit 132, and the second feature vector extracted by the extracting unit 60. Each distance from the second identification vector stored in the dictionary storage unit 132 is calculated and integrated by a predetermined method. The identification codes associated with the first identification vector and the second identification vector that minimize the integrated distance are recognized as handwriting input to the input unit 10.
  • the recognizing unit 180 calculates the distance R1 between the first feature vector and the first identification vector, the distance R2 between the second feature vector and the second identification vector, and an integrated function f (R1, R2). Use to integrate. Then, an integrated distance R that is an integrated distance is obtained.
  • the recognition unit 180 may integrate the distance R1 between the first feature vector and the first identification vector in preference to the distance R2 between the second feature vector and the second identification vector. For example, the recognition unit 180 may integrate the distance R2 as 0 when the distance R1 is smaller than the threshold value. That is, the recognition unit 180 may set the distance R1 as the integrated distance R when the distance R1 is smaller than the threshold value.
  • the threshold value may always be the same value, or may be changed for each identification code.
  • FIG. 9 is a flowchart showing an example of the operation of the recognition apparatus 101 of the present embodiment.
  • step S200 to step S210 is the same as the processing from step S100 to S110 in FIG.
  • step S212 the recognition unit 180 determines the distance between the first feature vector extracted by the extraction unit 60 and the first identification vector stored in the dictionary storage unit 132, the second feature vector extracted by the extraction unit 60, and the dictionary. Each distance from the second identification vector stored in the storage unit 132 is calculated and integrated by a predetermined method. The identification codes associated with the first identification vector and the second identification vector that minimize the integrated distance are recognized as handwriting input to the input unit 10.
  • step S214 the display control unit 90 causes the display unit 20 to display the recognition result of the recognition unit 180.
  • the first feature vector of the actual stroke and the second feature vector of the virtual stroke are extracted separately, and recognition processing is performed for each extracted feature vector. For this reason, also in this embodiment, a real stroke and a virtual stroke can be clearly distinguished and recognized, and recognition accuracy can be improved.
  • either the first feature vector or the second feature vector can be preferentially recognized in accordance with the handwriting mode (for example, the first feature vector is mainly used and the second feature vector is used. Vectors can be used supplementarily), and recognition accuracy can be further increased.
  • the recognition apparatus includes a control device such as a CPU, a storage device such as a ROM and a RAM, an external storage device such as an HDD and a removable drive device, a display device such as a display, a keyboard and a mouse. Etc., and has a hardware configuration using a normal computer.
  • the first feature vector and the second feature vector may be dimensionally compressed using a principal component analysis method (PCA) or the like.
  • PCA principal component analysis method
  • each feature vector extracted by the extraction unit 60 may be blurred.
  • a feature vector of a predetermined divided area (x, y) is converted into peripheral divided areas (x-1, y), (x + 1, y), (x, y-1), (x , Y + 1).
  • the habit of the handwriting input person can be blurred, and the recognition accuracy can be prevented from deteriorating.
  • F [d] [x] [y] (4 * F [d] [x] [y] + F [d] [x ⁇ 1] [y] + F [d] [x + 1] [y] + F [d] [X] [y-1] + F [d] [x] [y + 1]) / 8 (3)
  • d is any value from 1 to D.
  • the character recognition is described as an example, but the same method can be applied to gesture recognition.
  • the functions of the recognition devices of the first and second embodiments may be realized by executing a recognition program.
  • the recognition program executed by the recognition apparatuses of the first and second embodiments is stored in a computer-readable storage medium in an installable format or an executable file format and provided as a computer program product.
  • the recognition program executed by the recognition apparatus according to the first or second embodiment may be provided by being incorporated in advance in a ROM or the like.
  • the recognition program executed by the recognition apparatuses of the first and second embodiments has a module configuration for realizing the above-described units on a computer.
  • the CPU reads out a recognition program from the HDD or the like on the RAM and executes it, so that the above-described units are realized on the computer.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Character Discrimination (AREA)

Abstract

An acquiring section acquires coordinate data about an inputted stroke in order of time series.  A creating section creates a real stroke which is the stroke from a pen-down to a pen-up and creates a virtual stroke virtually connecting the coordinate data about the end point of the real stroke and the coordinate data about the start point of the next real stroke with a line segment.  An extracting section extracts feature vectors of the real stroke for each of the divisions defined by dividing the region where the stroke exists and obtains a first feature vector composed of the extracted feature vectors.  The extracting section furthermore extracts feature vectors of the virtual stroke for each of the divisions and obtains a second feature vector composed of the extracted feature vectors.  A recognizing section recognizes the stroke by using the distance between the first feature vector and the vector corresponding to the real stroke out of the vectors corresponding to the identification code and the distance between the second feature vector and the vector corresponding to the virtual stroke out of the vectors corresponding to the identification code.

Description

認識装置Recognition device
 本発明は、入力されたストロークから文字を認識する技術に関する。 The present invention relates to a technique for recognizing characters from input strokes.
 実際に入力されたストロークである実ストロークと、実ストローク間を接続したストロークである仮想ストロークとを用いて文字認識を行う技術がある。 There is a technology that performs character recognition using an actual stroke that is an actually input stroke and a virtual stroke that is a stroke connecting the actual strokes.
 このような文字認識技術では、実ストロークと仮想ストロークとを結合した1つのストロークに対して文字認識を行う。例えば特許文献1、2には、誤認識を防ぐために実ストロークには大きい重み付けを行い(筆跡を太くする)、仮想ストロークには小さい重み付を行った(筆跡を細くする)ストロークの、1つの特徴ベクトルに対して認識を行う技術が開示されている。 In such character recognition technology, character recognition is performed for one stroke obtained by combining an actual stroke and a virtual stroke. For example, in Patent Documents 1 and 2, in order to prevent misrecognition, a real weight is given a large weight (thickening the handwriting), and a virtual stroke is given a small weight (thinning the handwriting). A technique for recognizing a feature vector is disclosed.
特許第4099248号公報Japanese Patent No. 4099248 特許第4048716号公報Japanese Patent No. 4048716
 しかしながら、上述したような従来技術では、実ストローク及び仮想ストロークを区別せずに特徴ベクトルを抽出して認識を行っている。そのため、誤認識されてしまう可能性がある。 However, in the conventional technology as described above, the feature vector is extracted and recognized without distinguishing the actual stroke and the virtual stroke. Therefore, there is a possibility of being misrecognized.
 本発明は、上記事情に鑑みてなされたものであり、実ストロークと仮想ストロークとを明確に区別して認識することで、認識性能を向上する認識装置を提供することを目的とする。 The present invention has been made in view of the above circumstances, and an object of the present invention is to provide a recognition device that improves recognition performance by clearly distinguishing and recognizing a real stroke and a virtual stroke.
 上述した課題を解決するために、本発明の認識装置は、ユーザーが入力部から入力する筆跡の座標データを時系列の順に取得する取得部と、ペンダウンからペンアップがなされる前までの間の筆跡である実ストロークを作成するとともに、実ストロークの終点の座標データと当該実ストロークの次の実ストロークの始点の座標データとを仮想的に線分で接続した仮想ストロークを作成する作成部と、前記筆跡が存在する領域を分割した分割領域毎に前記実ストロークの特徴ベクトルを抽出し、抽出した各特徴ベクトルから成る第1特徴ベクトルを得るとともに、前記分割領域毎に前記仮想ストロークの特徴ベクトルを抽出し、抽出した各特徴ベクトルから成る第2特徴ベクトルを得る抽出部と、前記第1特徴ベクトルと識別コードに対応するベクトルのうち前記実ストロークに対応するベクトルとの距離と、前記第2特徴ベクトルと前記識別コードに対応するベクトルのうち前記仮想ストロークに対応するベクトルとの距離と、を用いて、前記筆跡を認識する認識部と、を備えることを特徴とする。 In order to solve the above-described problem, the recognition apparatus according to the present invention includes an acquisition unit that acquires coordinate data of handwriting input by the user from the input unit in chronological order, and before pen-up and before pen-up is performed. A creation unit that creates a virtual stroke in which the actual stroke as a handwriting is created, and the coordinate data of the end point of the actual stroke and the coordinate data of the start point of the next actual stroke of the actual stroke are virtually connected by a line segment; The feature vector of the actual stroke is extracted for each divided region obtained by dividing the region where the handwriting exists, and a first feature vector composed of each extracted feature vector is obtained, and the feature vector of the virtual stroke is obtained for each divided region. An extraction unit for extracting and obtaining a second feature vector composed of the extracted feature vectors, and corresponding to the first feature vector and the identification code; The handwriting is recognized using a distance between a vector corresponding to the actual stroke among vectors and a distance between a vector corresponding to the virtual stroke among vectors corresponding to the second feature vector and the identification code. And a recognizing unit.
 本発明によれば、実ストロークと仮想ストロークとを明確に区別して認識することで認識性能を向上できる。 According to the present invention, the recognition performance can be improved by clearly distinguishing the actual stroke from the virtual stroke.
第1実施形態の認識装置を示す図。The figure which shows the recognition apparatus of 1st Embodiment. 実ストローク及び仮想ストロークの例を示す図。The figure which shows the example of a real stroke and a virtual stroke. 第1特徴ベクトルの例を示す図。The figure which shows the example of a 1st feature vector. 第2特徴ベクトルの例を示す図。The figure which shows the example of a 2nd feature vector. 第1実施形態の辞書記憶部に記憶されている辞書の例を示す図。The figure which shows the example of the dictionary memorize | stored in the dictionary memory | storage part of 1st Embodiment. 第1実施形態の動作を示すフローチャート。The flowchart which shows operation | movement of 1st Embodiment. 第2実施形態の認識装置を示す図。The figure which shows the recognition apparatus of 2nd Embodiment. 第2実施形態の辞書記憶部に記憶されている辞書の例を示す図。The figure which shows the example of the dictionary memorize | stored in the dictionary memory | storage part of 2nd Embodiment. 第2実施形態の動作を示すフローチャート。The flowchart which shows operation | movement of 2nd Embodiment.
 以下、図面を参照しながら、本発明の認識装置の実施形態を説明する。 Hereinafter, embodiments of the recognition apparatus of the present invention will be described with reference to the drawings.
(第1の実施形態)
 本実施形態では、実ストロークの第1特徴ベクトルと仮想ストロークの第2特徴ベクトルとを別々に抽出する。抽出した各特徴ベクトルの特徴を残したまま結合した特徴ベクトルで、認識処理を行う例について説明する。
(First embodiment)
In the present embodiment, the first feature vector of the actual stroke and the second feature vector of the virtual stroke are extracted separately. An example in which recognition processing is performed using feature vectors that are combined while leaving the features of each extracted feature vector will be described.
 図1は、本実施形態の認識装置1の構成の一例を示すブロック図である。図1に示すように、認識装置1は、入力部10と、表示部20と、記憶部30と、取得部40と、作成部50と、抽出部60と、結合部70と、認識部80と、表示制御部90とを備える。 FIG. 1 is a block diagram showing an example of the configuration of the recognition apparatus 1 of the present embodiment. As illustrated in FIG. 1, the recognition apparatus 1 includes an input unit 10, a display unit 20, a storage unit 30, an acquisition unit 40, a creation unit 50, an extraction unit 60, a combination unit 70, and a recognition unit 80. And a display control unit 90.
 入力部10は、ユーザーがペン等の所定のデバイスや指などから筆跡を入力する。例えば、タッチパッド、タッチパネル、タブレットなどの既存の座標入力装置により実現できる。 The input unit 10 allows a user to input handwriting from a predetermined device such as a pen or a finger. For example, it can be realized by an existing coordinate input device such as a touch pad, a touch panel, or a tablet.
 表示部20は、後述する表示制御部90の指示により、後述する認識部80の認識結果などを表示する。液晶ディスプレイ、プラズマディスプレイ、有機ELディスプレイ、又はタッチパネル式ディスプレイなどの既存の表示装置により実現できる。 The display unit 20 displays a recognition result of the recognition unit 80 described later and the like according to an instruction from the display control unit 90 described later. It can be realized by an existing display device such as a liquid crystal display, a plasma display, an organic EL display, or a touch panel display.
 記憶部30は、認識装置1で行われる各種処理に使用される情報を記憶する。磁気的、電気的、又は光学的に記憶可能な既存の記憶媒体により実現できる。そして記憶部30は、辞書記憶部32を含む。なお、辞書記憶部32の詳細については後述する。 The storage unit 30 stores information used for various processes performed by the recognition device 1. This can be realized by an existing storage medium that can be stored magnetically, electrically, or optically. The storage unit 30 includes a dictionary storage unit 32. Details of the dictionary storage unit 32 will be described later.
 取得部40は、入力部10に入力された筆跡の座標データを時系列の順に取得する。取得部40は、ペンなどが入力部10の入力面に接触している間、一定時間間隔で筆跡の座標データを取得する。 The acquisition unit 40 acquires the coordinate data of the handwriting input to the input unit 10 in chronological order. The acquisition unit 40 acquires handwriting coordinate data at regular time intervals while a pen or the like is in contact with the input surface of the input unit 10.
 作成部50は、入力部10に入力された筆跡であって、ペンダウンからペンアップがなされる前までの間の筆跡である実ストロークを作成する。また、それと共に、実ストロークの終点の座標データと当該実ストロークの次の実ストロークの始点の座標データとを仮想的に線分で接続した仮想ストロークを作成する。なお、実ストロークは、一画分の筆跡であり、ペンなどが入力部10の入力面に接触してから離れるまで(ペンダウンからペンアップがなされる前まで)の筆跡をいう。 The creation unit 50 creates an actual stroke that is a handwriting input from the input unit 10 and before the pen-up is performed. At the same time, a virtual stroke is created in which the coordinate data of the end point of the actual stroke and the coordinate data of the start point of the next actual stroke of the actual stroke are virtually connected by a line segment. The actual stroke is a handwriting for one stroke, and refers to a handwriting from when the pen or the like touches the input surface of the input unit 10 until it leaves (from before pen-down to before pen-up).
 作成部50は、ペンなどが入力部10の入力面に接触してから離れるまでの間に取得された座標データを時系列の順に線分で接続して実ストロークを作成する。例えば、i番目の実ストロークの始点から終点までの座標データは、(X[1],Y[1]),(X[2],Y[2]),…,(X[N],Y[N])という2次元座標の時系列で表すことができる。作成部50は、これらの座標データを時系列の順に線分で接続して実ストロークを作成する。なお、iは自然数である。また、Nは1以上の自然数であり、i番目の実ストロークの座標データ数を示す。 The creation unit 50 creates an actual stroke by connecting the coordinate data acquired from when the pen or the like comes into contact with the input surface of the input unit 10 until the pen leaves the line, in a chronological order. For example, the coordinate data from the start point to the end point of the i-th actual stroke is (X i [1], Y i [1]), (X i [2], Y i [2]), ..., (X i [N i ], Y i [N i ]). The creation unit 50 creates an actual stroke by connecting these coordinate data with line segments in order of time series. Note that i is a natural number. Ni is a natural number of 1 or more, and indicates the number of coordinate data of the i-th actual stroke.
 また、作成部50は、実ストロークの終点の座標データと当該実ストロークの次の実ストロークの始点の座標データとを仮想的に線分で接続した仮想ストロークを作成する。例えば、i番目の実ストロークの終点の座標データは(X[N],Y[N])と表され、i+1番目の実ストロークの始点の座標データは(Xi+1[1],Yi+1[1])と表される。作成部50は、両座標データを仮想的に線分で接続してi番目の仮想ストローク(i番目の実ストロークとi+1番目の実ストロークの間の仮想ストローク)を作成する。 In addition, the creation unit 50 creates a virtual stroke in which the coordinate data of the end point of the actual stroke and the coordinate data of the start point of the next actual stroke after the actual stroke are virtually connected by a line segment. For example, the coordinate data of the end point of the i-th actual stroke is represented as (X i [N i ], Y i [N i ]), and the coordinate data of the start point of the i + 1-th actual stroke is (X i + 1 [1], Y i + 1 [1]). The creating unit 50 virtually connects the two coordinate data with a line segment to create an i-th virtual stroke (a virtual stroke between the i-th actual stroke and the i + 1-th actual stroke).
 作成部50は、実ストロークの終点と次の実ストロークの始点の2点を仮想的に線分で接続して仮想ストロークを作成する例について述べた。また、2点間を補間して、3点以上を仮想的に線分で接続して仮想ストロークを作成してもよい。補間は、直線補間でもよいし、始点と終点での方向変化の連続性を考慮した補間でもよい。 The creation unit 50 described an example in which a virtual stroke is created by virtually connecting two end points of an actual stroke and the start point of the next actual stroke with a line segment. Alternatively, a virtual stroke may be created by interpolating between two points and virtually connecting three or more points with line segments. The interpolation may be linear interpolation or interpolation considering the continuity of the direction change at the start point and the end point.
 図2は、作成部50により作成される実ストローク及び仮想ストロークの一例を示す図である。図2に示す例では、ドットが座標データを示し、ドットを接続する実線が実ストロークを示し、破線の矢印が仮想ストロークを示している。 FIG. 2 is a diagram illustrating an example of an actual stroke and a virtual stroke created by the creation unit 50. In the example shown in FIG. 2, dots indicate coordinate data, solid lines connecting the dots indicate actual strokes, and broken arrows indicate virtual strokes.
 図1に戻り、抽出部60は、入力部10に入力された筆跡が存在する領域を分割した分割領域毎に実ストロークの特徴ベクトルを抽出し、抽出した各特徴ベクトルから成る第1特徴ベクトルを得る。また、分割領域毎に仮想ストロークの特徴ベクトルを抽出し、抽出した各特徴ベクトルから成る第2特徴ベクトルを得る。 Returning to FIG. 1, the extraction unit 60 extracts the feature vector of the actual stroke for each divided region obtained by dividing the region where the handwriting input to the input unit 10 exists, and obtains the first feature vector composed of the extracted feature vectors. obtain. Further, a feature vector of the virtual stroke is extracted for each divided region, and a second feature vector composed of each extracted feature vector is obtained.
 抽出部60は、分割領域毎に実ストロークの座標データ間の方向の出現頻度分布を抽出する。抽出した出現頻度分布を1列に配列することにより第1特徴ベクトルを得る。また、分割領域毎に仮想ストロークの座標データ間の方向の出現頻度分布を抽出する。抽出した出現頻度分布を1列に配列することにより第2特徴ベクトルを得る。なお、本実施形態では、出現頻度分布として方向成分密度特徴の分布を例にとり説明する。 The extraction unit 60 extracts the appearance frequency distribution in the direction between the coordinate data of the actual stroke for each divided area. A first feature vector is obtained by arranging the extracted appearance frequency distributions in one column. Further, the appearance frequency distribution in the direction between the coordinate data of the virtual strokes is extracted for each divided region. A second feature vector is obtained by arranging the extracted appearance frequency distributions in one column. In the present embodiment, the distribution of direction component density features will be described as an example of the appearance frequency distribution.
 方向成分密度特徴は、x方向、y方向、及び座標データ間の方向(実ストローク又は仮想ストロークの方向)を示す3次元の配列F[d][x][y]で表される。例えば、x方向の量子化数をNx、y方向の量子化数をNy、座標データ間の方向の量子化数をDとすると、配列F[d][x][y]は、(F[1][1][1],F[1][1][2],…,F[1][1][Ny],F[1][2][1],…,F[D][Nx][Ny-1],F[D][Nx][Ny])というNx*Ny*D次元のベクトルを示す。なお、Nx、Ny、及びDは、いずれも2以上の自然数である。 The direction component density feature is represented by a three-dimensional array F [d] [x] [y] indicating the x direction, the y direction, and the direction between the coordinate data (direction of actual stroke or virtual stroke). For example, if the quantization number in the x direction is Nx, the quantization number in the y direction is Ny, and the quantization number in the direction between coordinate data is D, the array F [d] [x] [y] is (F [ 1] [1] [1], F [1] [1] [2],..., F [1] [1] [Ny], F [1] [2] [1],. [Nx] [Ny-1], F [D] [Nx] [Ny]) Nx * Ny * D-dimensional vectors. Nx, Ny, and D are all natural numbers of 2 or more.
 そして、抽出部60は、実ストローク、仮想ストロークの方向成分密度特徴を個別に算出して、それぞれ第1特徴ベクトル、第2特徴ベクトルを得る。方向成分密度特徴の算出には、例えば、「“最小分類誤り学習による特徴選択型文字認識”、電子情報通信学会論文誌.D-II,情報・システム,II-情報処理.Vol.81 No.12,1998年12月,p.2749-2756」に開示された手法を用いることができる。 Then, the extraction unit 60 individually calculates the direction component density features of the real stroke and the virtual stroke to obtain the first feature vector and the second feature vector, respectively. For calculating the direction component density feature, for example, "" Feature selection type character recognition by minimum classification error learning ", IEICE Transactions. D-II, Information / System, II-Information Processing, Vol. 12, December 1998, p.2749-2756 ”can be used.
 図3は、抽出部60が抽出した第1特徴ベクトルの一例を示す図である。詳細には、図2に示す実ストロークの方向成分密度特徴を示している。図4は、抽出部60により抽出された第2特徴ベクトルの一例を示す図であり、詳細には、図2に示す仮想ストロークの方向成分密度特徴を示している。なお、図3及び図4に示す例では、x方向の量子化数を8、y方向の量子化数を8、座標データ間の方向の量子化数を4とされているため、配列Fは128次元のベクトルで表される。 FIG. 3 is a diagram illustrating an example of the first feature vector extracted by the extraction unit 60. Specifically, the directional component density characteristics of the actual stroke shown in FIG. 2 are shown. FIG. 4 is a diagram showing an example of the second feature vector extracted by the extraction unit 60, and more specifically shows the direction component density feature of the virtual stroke shown in FIG. In the example shown in FIGS. 3 and 4, since the quantization number in the x direction is 8, the quantization number in the y direction is 8, and the quantization number in the direction between coordinate data is 4, the array F is It is represented by a 128-dimensional vector.
 なお、分割領域は、予め定められていてもよいし、筆跡が存在する領域に応じて抽出部60が予め定められた数(詳細には、Nx及びNyの値)に分割するようにしてもよい。 The divided areas may be determined in advance, or may be divided into a predetermined number (specifically, values of Nx and Ny) by the extraction unit 60 according to the area where the handwriting exists. Good.
 図1に戻り、結合部70は、抽出部60により抽出された第1特徴ベクトル及び第2特徴ベクトルを結合して、結合ベクトルを得る。結合部70は、第1特徴ベクトル及び第2特徴ベクトルを1列に配列することにより、結合ベクトルを得る。 1, the combining unit 70 combines the first feature vector and the second feature vector extracted by the extracting unit 60 to obtain a combined vector. The combining unit 70 obtains a combined vector by arranging the first feature vector and the second feature vector in one column.
 例えば、第1特徴ベクトルを示す配列Fを(a[1],a[2],…,a[L])というL次元のベクトルとする。第2特徴ベクトルを示す配列Fを(b[1],b[2],…,b[M])というM次元のベクトルとする。なお、L及びMは、いずれも8以上の自然数である。この場合、結合部70は、第1特徴ベクトル及び第2特徴ベクトルを1列に配列して、(a[1],a[2],…,a[L],b[1],b[2],…,b[M])というL+M次元の結合ベクトルを得る。 For example, an array F indicating the first feature vector is an L-dimensional vector (a [1], a [2],..., A [L]). An array F indicating the second feature vector is an M-dimensional vector (b [1], b [2],..., B [M]). Note that L and M are both natural numbers of 8 or more. In this case, the combining unit 70 arranges the first feature vector and the second feature vector in one column, and (a [1], a [2],..., A [L], b [1], b [ 2],..., B [M]) to obtain an L + M-dimensional coupling vector.
 なお、結合部70は、第1特徴ベクトル及び第2特徴ベクトルを結合する際に、それぞれのベクトルの要素に重み付けをしてもよい。 The combining unit 70 may weight each vector element when combining the first feature vector and the second feature vector.
 ここで、辞書記憶部32について説明する。辞書記憶部32は、識別コード毎に、当該識別コードの第1特徴ベクトル及び第2特徴ベクトルを結合した識別ベクトルを対応付けて記憶する。 Here, the dictionary storage unit 32 will be described. The dictionary storage unit 32 stores, for each identification code, an identification vector obtained by combining the first feature vector and the second feature vector of the identification code in association with each other.
 図5は、辞書記憶部32に記憶されている辞書の一例を示すである。図5に示す例では、識別コードである文字コード(SJIS)と、当該文字コードの第1特徴ベクトル及び第2特徴ベクトルを結合した識別ベクトルとが、対応付けられている。識別ベクトルは、結合ベクトルを結合した方法と同様の方法で結合されている。なお、P1、P2、…が文字コードの識別ベクトルを示し、P11、P21、…が文字コードの第1特徴ベクトルを示し、P12、P22、…が文字コードの第2特徴ベクトルを示す。 FIG. 5 shows an example of a dictionary stored in the dictionary storage unit 32. In the example shown in FIG. 5, a character code (SJIS) that is an identification code is associated with an identification vector obtained by combining the first feature vector and the second feature vector of the character code. The identification vectors are combined by a method similar to the method of combining the combination vectors. P1, P2,... Indicate character code identification vectors, P11, P21,... Indicate character code first feature vectors, and P12, P22,.
 図1に戻り、認識部80は、抽出部60により抽出された第1特徴ベクトル及び第2特徴ベクトルを用いて、入力部10に入力された筆跡を認識する。認識部80は、結合部70により結合された結合ベクトルと辞書記憶部32に記憶された識別ベクトルとの距離を計算する。計算した距離が最小となる識別ベクトルに対応付けられた識別コードを入力部10に入力された筆跡として認識する。つまり、第1特徴ベクトルと辞書記憶部32に記憶された識別ベクトルの実ストロークに対応するベクトルとの距離を算出する。また、第2特徴ベクトルと、辞書記憶部32に記憶された識別ベクトルの仮想ストロークに対応するベクトルとの距離を算出する。算出された2つの距離を総合判断し、文字を識別する。 1, the recognition unit 80 recognizes the handwriting input to the input unit 10 using the first feature vector and the second feature vector extracted by the extraction unit 60. The recognition unit 80 calculates the distance between the combined vector combined by the combining unit 70 and the identification vector stored in the dictionary storage unit 32. The identification code associated with the identification vector that minimizes the calculated distance is recognized as the handwriting input to the input unit 10. That is, the distance between the first feature vector and the vector corresponding to the actual stroke of the identification vector stored in the dictionary storage unit 32 is calculated. Further, the distance between the second feature vector and the vector corresponding to the virtual stroke of the identification vector stored in the dictionary storage unit 32 is calculated. A comprehensive judgment is made on the two calculated distances to identify the character.
 例えば、結合ベクトルを(s[1],s[2],…,s[N])というN次元のベクトルとし、識別ベクトルを(t[1],t[2],…,t[N])というN次元のベクトルとすると、認識部80は、数式(1)に示すユークリッド距離Rを用いて、両ベクトル間の距離を計算する。なお、Nは、いずれも16以上の自然数である。 For example, the coupling vector is an N-dimensional vector (s [1], s [2],..., S [N]), and the identification vector is (t [1], t [2],..., T [N]. ), The recognition unit 80 calculates the distance between the two vectors using the Euclidean distance R shown in Equation (1). Note that N is a natural number of 16 or more.
Figure JPOXMLDOC01-appb-M000001
Figure JPOXMLDOC01-appb-M000001
 表示制御部90は、認識部80の認識結果を表示部20に表示させる。具体的には、表示制御部90は、認識部80により、入力部10に入力された筆跡と認識された識別コードが示す文字などを表示部20に表示させる。 The display control unit 90 displays the recognition result of the recognition unit 80 on the display unit 20. Specifically, the display control unit 90 causes the recognition unit 80 to display characters or the like indicated by the identification code recognized as the handwriting input to the input unit 10.
 なお、取得部40、作成部50、抽出部60、結合部70、認識部80、及び表示制御部90は、例えば、CPUなどの既存の演算装置により実現できる。 In addition, the acquisition unit 40, the creation unit 50, the extraction unit 60, the combination unit 70, the recognition unit 80, and the display control unit 90 can be realized by an existing arithmetic device such as a CPU.
 図6は、本実施形態の認識装置1で行われる認識処理の手順の流れの一例を示すフローチャートである。 FIG. 6 is a flowchart illustrating an example of a flow of a recognition process performed by the recognition apparatus 1 according to the present embodiment.
 ステップS100では、入力部10には、ペンなどから筆跡が入力される。 In step S100, handwriting is input to the input unit 10 from a pen or the like.
 ステップS102では、取得部40は、入力部10に入力された筆跡の座標データを時系列の順に取得する。 In step S102, the acquisition unit 40 acquires the coordinate data of the handwriting input to the input unit 10 in chronological order.
 ステップS104では、作成部50は、入力された筆跡を構成するストロークである実ストロークを作成する。 In step S104, the creation unit 50 creates an actual stroke that is a stroke constituting the input handwriting.
 ステップS106では、作成部50は、実ストロークの終点の座標データと当該実ストロークの次の実ストロークの始点の座標データとを仮想的に線分で接続した仮想ストロークを作成する。 In step S106, the creation unit 50 creates a virtual stroke in which the coordinate data of the end point of the actual stroke and the coordinate data of the start point of the next actual stroke are virtually connected by a line segment.
 ステップS108では、抽出部60は、入力された筆跡が存在する領域を分割した分割領域毎に実ストロークの特徴ベクトルを抽出し、抽出した各特徴ベクトルから成る第1特徴ベクトルを得る。 In step S108, the extraction unit 60 extracts a feature vector of the actual stroke for each divided region obtained by dividing the region where the input handwriting exists, and obtains a first feature vector composed of each extracted feature vector.
 ステップS110では、抽出部60は、分割領域毎に仮想ストロークの特徴ベクトルを抽出し、抽出した各特徴ベクトルから成る第2特徴ベクトルを得る。 In step S110, the extraction unit 60 extracts a feature vector of the virtual stroke for each divided region, and obtains a second feature vector composed of the extracted feature vectors.
 ステップS112では、結合部70は、抽出部60により抽出された第1特徴ベクトル及び第2特徴ベクトルを結合して、結合ベクトルを得る。 In step S112, the combining unit 70 combines the first feature vector and the second feature vector extracted by the extracting unit 60 to obtain a combined vector.
 ステップS114では、認識部80は、結合部70により結合された結合ベクトルと辞書記憶部32に記憶された識別ベクトルとの距離を計算する。計算した距離が最小となる識別ベクトルに対応付けられた識別コードを入力部10に入力された筆跡として認識する。 In step S114, the recognition unit 80 calculates the distance between the combined vector combined by the combining unit 70 and the identification vector stored in the dictionary storage unit 32. The identification code associated with the identification vector that minimizes the calculated distance is recognized as the handwriting input to the input unit 10.
 ステップS116では、表示制御部90は、認識部80の認識結果を表示部20に表示させる。 In step S116, the display control unit 90 causes the display unit 20 to display the recognition result of the recognition unit 80.
 本実施形態では、実ストロークの第1特徴ベクトルと仮想ストロークの第2特徴ベクトルとを別々に抽出し、抽出した各特徴ベクトルの特徴を残したまま結合して、認識処理を行う。このため本実施形態によれば、実ストロークと仮想ストロークとを明確に区別して認識することができ、認識精度を高めることができる。例えば、「=」と「Z」のように、実ストロークと仮想ストロークとを明確に区別できなければ認識が困難な文字であっても、正しく認識することができる。 In this embodiment, the first feature vector of the actual stroke and the second feature vector of the virtual stroke are separately extracted, and the recognition process is performed by combining the extracted feature vectors while leaving the features of the feature vectors. Therefore, according to the present embodiment, the actual stroke and the virtual stroke can be clearly distinguished and recognized, and the recognition accuracy can be improved. For example, characters such as “=” and “Z” that are difficult to recognize unless the actual stroke and the virtual stroke can be clearly distinguished can be recognized correctly.
(第2の実施形態)
 本実施形態では、実ストロークの第1特徴ベクトルと仮想ストロークの第2特徴ベクトルとを別々に抽出し、抽出した特徴ベクトル毎に認識処理を行う例について説明する。以下では、第1の実施形態との相違点の説明を主に行い、同様の機能を有する構成要素については、同様の名称・符号を付し、その説明を省略する。
(Second Embodiment)
In the present embodiment, an example will be described in which a first feature vector of an actual stroke and a second feature vector of a virtual stroke are separately extracted, and recognition processing is performed for each extracted feature vector. In the following description, differences from the first embodiment will be mainly described, and components having similar functions will be denoted by the same names and symbols, and description thereof will be omitted.
 図7は、本実施形態の認識装置101の構成の一例を示すブロック図である。図7に示す認識装置101は、記憶部130の辞書記憶部132が記憶する辞書データ、及び認識部180の処理内容、及び結合部70を含まない点で、第1の実施形態の認識装置1と相違する。 FIG. 7 is a block diagram illustrating an example of the configuration of the recognition apparatus 101 according to the present embodiment. The recognition apparatus 101 illustrated in FIG. 7 does not include the dictionary data stored in the dictionary storage unit 132 of the storage unit 130, the processing content of the recognition unit 180, and the combining unit 70. Is different.
 辞書記憶部132は、識別コード毎に、当該識別コードの第1特徴ベクトルである第1識別ベクトルと当該識別コードの第2特徴ベクトルである第2識別ベクトルを対応付けて記憶する。 The dictionary storage unit 132 stores, for each identification code, the first identification vector that is the first feature vector of the identification code and the second identification vector that is the second feature vector of the identification code in association with each other.
 図8は、辞書記憶部132に記憶されている辞書の一例を示すである。図8に示す例では、識別コードである文字コード(SJIS)と、当該文字コードの第1識別ベクトルと、当該文字コードの第2識別ベクトルとが、対応付けられている。なお、P11、P21、…が文字コードの第1識別ベクトルを示し、P12、P22、…が文字コードの第2識別ベクトルを示す。 FIG. 8 shows an example of a dictionary stored in the dictionary storage unit 132. In the example shown in FIG. 8, a character code (SJIS) that is an identification code, a first identification vector of the character code, and a second identification vector of the character code are associated with each other. P11, P21,... Indicate the first identification vector of the character code, and P12, P22,.
 図7に戻り、認識部180は、抽出部60により抽出された第1特徴ベクトルと辞書記憶部132に記憶された第1識別ベクトルとの距離、抽出部60により抽出された第2特徴ベクトルと辞書記憶部132に記憶された第2識別ベクトルとの距離をそれぞれ計算して所定の手法で統合する。統合した距離が最小となる第1識別ベクトル及び第2識別ベクトルに対応付けられた識別コードを入力部10に入力された筆跡として認識する。 Returning to FIG. 7, the recognizing unit 180 determines the distance between the first feature vector extracted by the extracting unit 60 and the first identification vector stored in the dictionary storage unit 132, and the second feature vector extracted by the extracting unit 60. Each distance from the second identification vector stored in the dictionary storage unit 132 is calculated and integrated by a predetermined method. The identification codes associated with the first identification vector and the second identification vector that minimize the integrated distance are recognized as handwriting input to the input unit 10.
 具体的には、認識部180は、第1特徴ベクトルと第1識別ベクトルとの距離R1と、第2特徴ベクトルと第2識別ベクトルとの距離R2とを、統合関数f(R1,R2)を用いて統合する。そして、統合した距離である統合距離Rを求める。なお、統合関数f(R1,R2)には、各距離の線形和を求める数式(2)などを用いることができる。
f(R1,R2)=(k1*R1+k2*R2)/(k1+k2)   …(2)
そして、認識部180は、統合距離Rが最小となる第1識別ベクトル及び第2識別ベクトルに対応付けられた識別コードを入力部10に入力された筆跡として認識する。
Specifically, the recognizing unit 180 calculates the distance R1 between the first feature vector and the first identification vector, the distance R2 between the second feature vector and the second identification vector, and an integrated function f (R1, R2). Use to integrate. Then, an integrated distance R that is an integrated distance is obtained. For the integration function f (R1, R2), equation (2) for obtaining a linear sum of each distance can be used.
f (R1, R2) = (k1 * R1 + k2 * R2) / (k1 + k2) (2)
Then, the recognizing unit 180 recognizes the identification code associated with the first identification vector and the second identification vector having the minimum integrated distance R as the handwriting input to the input unit 10.
 なお、認識部180は、第1特徴ベクトルと第1識別ベクトルとの距離R1を、第2特徴ベクトルと第2識別ベクトルとの距離R2に優先して統合するようにしてもよい。例えば、認識部180は、距離R1が閾値よりも小さい場合には、距離R2を0として統合してもよい。つまり、認識部180は、距離R1が閾値よりも小さい場合には、距離R1を統合距離Rとしてもよい。閾値は、常に同じ値を使用してもよいし、識別コード毎に値を変えてもよい。 Note that the recognition unit 180 may integrate the distance R1 between the first feature vector and the first identification vector in preference to the distance R2 between the second feature vector and the second identification vector. For example, the recognition unit 180 may integrate the distance R2 as 0 when the distance R1 is smaller than the threshold value. That is, the recognition unit 180 may set the distance R1 as the integrated distance R when the distance R1 is smaller than the threshold value. The threshold value may always be the same value, or may be changed for each identification code.
 図9は、本実施形態の認識装置101の動作の一例を示すフローチャートである。 FIG. 9 is a flowchart showing an example of the operation of the recognition apparatus 101 of the present embodiment.
 ステップS200~ステップS210までの処理は、図6のステップS100~S110までの処理と同様である。 The processing from step S200 to step S210 is the same as the processing from step S100 to S110 in FIG.
 ステップS212では、認識部180は、抽出部60により抽出された第1特徴ベクトルと辞書記憶部132に記憶された第1識別ベクトルとの距離、抽出部60により抽出された第2特徴ベクトルと辞書記憶部132に記憶された第2識別ベクトルとの距離をそれぞれ計算して所定の手法で統合する。統合した距離が最小となる第1識別ベクトル及び第2識別ベクトルに対応付けられた識別コードを入力部10に入力された筆跡として認識する。 In step S212, the recognition unit 180 determines the distance between the first feature vector extracted by the extraction unit 60 and the first identification vector stored in the dictionary storage unit 132, the second feature vector extracted by the extraction unit 60, and the dictionary. Each distance from the second identification vector stored in the storage unit 132 is calculated and integrated by a predetermined method. The identification codes associated with the first identification vector and the second identification vector that minimize the integrated distance are recognized as handwriting input to the input unit 10.
 ステップS214では、表示制御部90は、認識部180の認識結果を表示部20に表示させる。 In step S214, the display control unit 90 causes the display unit 20 to display the recognition result of the recognition unit 180.
 このように本実施形態では、実ストロークの第1特徴ベクトルと仮想ストロークの第2特徴ベクトルとを別々に抽出し、抽出した特徴ベクトル毎に認識処理を行う。このため本実施形態においても、実ストロークと仮想ストロークとを明確に区別して認識することができ、認識精度を高めることができる。特に本実施形態によれば、筆跡の態様に合わせて第1特徴ベクトル又は第2特徴ベクトルのいずれかを優先して認識することができ(例えば、第1特徴ベクトルを主に用い、第2特徴ベクトルを補助的に用いることができ)、認識精度をより高めることができる。 Thus, in this embodiment, the first feature vector of the actual stroke and the second feature vector of the virtual stroke are extracted separately, and recognition processing is performed for each extracted feature vector. For this reason, also in this embodiment, a real stroke and a virtual stroke can be clearly distinguished and recognized, and recognition accuracy can be improved. In particular, according to the present embodiment, either the first feature vector or the second feature vector can be preferentially recognized in accordance with the handwriting mode (for example, the first feature vector is mainly used and the second feature vector is used. Vectors can be used supplementarily), and recognition accuracy can be further increased.
 第1~2の実施形態の認識装置は、CPUなどの制御装置と、ROMやRAMなどの記憶装置と、HDDやリムーバブルドライブ装置などの外部記憶装置と、ディスプレイなどの表示装置と、キーボードやマウスなどの入力装置を備えており、通常のコンピュータを利用したハードウェア構成となっている。 The recognition apparatus according to the first and second embodiments includes a control device such as a CPU, a storage device such as a ROM and a RAM, an external storage device such as an HDD and a removable drive device, a display device such as a display, a keyboard and a mouse. Etc., and has a hardware configuration using a normal computer.
(変形例)
 なお、本発明は、上記実施形態そのままに限定されるものではなく、実施段階ではその要旨を逸脱しない範囲で構成要素を変形して具体化することができる。また、上記実施形態に開示されている複数の構成要素の適宜な組み合わせにより、種々の発明を形成することができる。例えば、実施形態に示される全構成要素からいくつかの構成要素を削除してもよい。さらに、異なる実施形態にわたる構成要素を適宜組み合わせても良い。
(Modification)
Note that the present invention is not limited to the above-described embodiment as it is, and can be embodied by modifying the constituent elements without departing from the scope of the invention in the implementation stage. Moreover, various inventions can be formed by appropriately combining a plurality of constituent elements disclosed in the embodiment. For example, some components may be deleted from all the components shown in the embodiment. Furthermore, the constituent elements over different embodiments may be appropriately combined.
 例えば、上記実施形態において、第1特徴ベクトル及び第2特徴ベクトルを、主成分分析法(PCA)などを用いて次元圧縮するようにしてもよい。このようにすると、第1特徴ベクトル及び第2特徴ベクトルのデータ量を削減することができる。 For example, in the above embodiment, the first feature vector and the second feature vector may be dimensionally compressed using a principal component analysis method (PCA) or the like. In this way, the data amount of the first feature vector and the second feature vector can be reduced.
 また例えば、上記実施形態において、抽出部60により抽出された各特徴ベクトルをぼかすようにしてもよい。例えば、数式(3)により、所定の分割領域(x,y)の特徴ベクトルを、周辺の分割領域(x-1,y)、(x+1,y)、(x,y-1)、(x,y+1)の特徴ベクトルでぼかすことができる。このようにすると、筆跡の入力者のクセ等をぼかすことができ、認識精度の低下を防ぐことができる。
F[d][x][y]=(4*F[d][x][y]+F[d][x-1][y]+F[d][x+1][y]+F[d][x][y-1]+F[d][x][y+1])/8   …(3)
なお、dは、1~Dのいずれかの値である。
Further, for example, in the above embodiment, each feature vector extracted by the extraction unit 60 may be blurred. For example, according to Equation (3), a feature vector of a predetermined divided area (x, y) is converted into peripheral divided areas (x-1, y), (x + 1, y), (x, y-1), (x , Y + 1). In this way, the habit of the handwriting input person can be blurred, and the recognition accuracy can be prevented from deteriorating.
F [d] [x] [y] = (4 * F [d] [x] [y] + F [d] [x−1] [y] + F [d] [x + 1] [y] + F [d] [X] [y-1] + F [d] [x] [y + 1]) / 8 (3)
Here, d is any value from 1 to D.
 また、上記実施形態では、文字認識を例にとり説明したが、ジェスチャー認識などにおいても同様の手法を適用できる。 In the above embodiment, the character recognition is described as an example, but the same method can be applied to gesture recognition.
 また例えば、第1~2の実施形態の認識装置の機能を、認識プログラムを実行することにより実現させるようにしてもよい。 Further, for example, the functions of the recognition devices of the first and second embodiments may be realized by executing a recognition program.
 この場合、第1~2の実施形態の認識装置で実行される認識プログラムは、インストール可能な形式又は実行可能なファイル形式でコンピュータが読み取り可能な記憶媒体に記憶されてコンピュータプログラムプロダクトとして提供される。また、第1~2の実施形態の認識装置で実行される認識プログラムを、ROM等に予め組み込んで提供するようにしてもよい。 In this case, the recognition program executed by the recognition apparatuses of the first and second embodiments is stored in a computer-readable storage medium in an installable format or an executable file format and provided as a computer program product. . The recognition program executed by the recognition apparatus according to the first or second embodiment may be provided by being incorporated in advance in a ROM or the like.
 第1~2の実施形態の認識装置で実行される認識プログラムは、上述した各部をコンピュータ上で実現させるためのモジュール構成となっている。実際のハードウェアとしては、CPUがHDD等から認識プログラムをRAM上に読み出して実行することにより、上記各部がコンピュータ上で実現されるようになっている。 The recognition program executed by the recognition apparatuses of the first and second embodiments has a module configuration for realizing the above-described units on a computer. As actual hardware, the CPU reads out a recognition program from the HDD or the like on the RAM and executes it, so that the above-described units are realized on the computer.
 1、101 認識装置
 10 入力部
 20 表示部
 30、130 記憶部
 32、132 辞書記憶部
 40 取得部
 50 作成部
 60 抽出部
 70 結合部
 80、180 認識部
 90 表示制御部
DESCRIPTION OF SYMBOLS 1,101 Recognition apparatus 10 Input part 20 Display part 30,130 Storage part 32,132 Dictionary storage part 40 Acquisition part 50 Creation part 60 Extraction part 70 Combining part 80,180 Recognition part 90 Display control part

Claims (6)

  1.  ユーザーが入力部から入力する筆跡の座標データを時系列の順に取得する取得部と、
     ペンダウンからペンアップがなされる前までの間の筆跡である実ストロークを作成するとともに、実ストロークの終点の座標データと当該実ストロークの次の実ストロークの始点の座標データとを仮想的に線分で接続した仮想ストロークを作成する作成部と、
     前記筆跡が存在する領域を分割した分割領域毎に前記実ストロークの特徴ベクトルを抽出し、抽出した各特徴ベクトルから成る第1特徴ベクトルを得るとともに、前記分割領域毎に前記仮想ストロークの特徴ベクトルを抽出し、抽出した各特徴ベクトルから成る第2特徴ベクトルを得る抽出部と、
     前記第1特徴ベクトルと識別コードに対応するベクトルのうち前記実ストロークに対応するベクトルとの距離と、前記第2特徴ベクトルと前記識別コードに対応するベクトルのうち前記仮想ストロークに対応するベクトルとの距離と、を用いて、前記筆跡を認識する認識部と、
     を備えることを特徴とする認識装置。
    An acquisition unit that acquires coordinate data of handwriting input by the user from the input unit in chronological order;
    Creates an actual stroke, which is the handwriting from the pen-down to before the pen-up, and creates a virtual line segment between the coordinate data of the end point of the actual stroke and the coordinate data of the start point of the next actual stroke. A creation unit that creates virtual strokes connected with
    The feature vector of the actual stroke is extracted for each divided region obtained by dividing the region where the handwriting exists, and a first feature vector composed of each extracted feature vector is obtained, and the feature vector of the virtual stroke is obtained for each divided region. An extraction unit for extracting and obtaining a second feature vector composed of the extracted feature vectors;
    A distance between the first feature vector and a vector corresponding to the actual stroke among vectors corresponding to the identification code, and a vector corresponding to the virtual stroke among vectors corresponding to the second feature vector and the identification code. A recognition unit for recognizing the handwriting using a distance; and
    A recognition apparatus comprising:
  2.  抽出部により抽出された前記第1特徴ベクトル及び前記第2特徴ベクトルを結合して、結合ベクトルを得る結合部と、
     前記識別コード毎に、当該識別コードの第1特徴ベクトル及び第2特徴ベクトルを、前記結合ベクトルを結合した方法と同様の方法で結合した識別ベクトルを対応付けて記憶する辞書記憶部と、を更に備え、
     前記認識部は、前記結合ベクトルと前記識別ベクトルとの距離を計算し、計算した距離が最小となる前記識別ベクトルに対応付けられた識別コードを前記筆跡として認識することを特徴とする請求項1に記載の認識装置。
    A combining unit that combines the first feature vector and the second feature vector extracted by the extracting unit to obtain a combined vector;
    A dictionary storage unit that stores, for each identification code, the first feature vector and the second feature vector of the identification code in association with the identification vector obtained by combining the combination vectors in the same manner as the combination vector. Prepared,
    The recognition unit calculates a distance between the combination vector and the identification vector, and recognizes an identification code associated with the identification vector that minimizes the calculated distance as the handwriting. The recognition device described in 1.
  3.  前記識別コード毎に、当該識別コードの第1特徴ベクトルである第1識別ベクトルと当該識別コードの第2特徴ベクトルである第2識別ベクトルを対応付けて記憶する辞書記憶部を更に備え、
     前記認識部は、抽出部により抽出された前記第1特徴ベクトルと前記第1識別ベクトルとの距離、抽出部により抽出された前記第2特徴ベクトルと前記第2識別ベクトルとの距離をそれぞれ計算して所定の手法で統合し、統合した距離が最小となる前記第1識別ベクトル及び前記第2識別ベクトルに対応付けられた識別コードを前記筆跡として認識することを特徴とする請求項1に記載の認識装置。
    A dictionary storage unit that stores, for each identification code, a first identification vector that is a first feature vector of the identification code and a second identification vector that is a second feature vector of the identification code in association with each other;
    The recognition unit calculates a distance between the first feature vector and the first identification vector extracted by the extraction unit, and a distance between the second feature vector and the second identification vector extracted by the extraction unit, respectively. The identification code associated with the first identification vector and the second identification vector with which the integrated distance is minimized is recognized as the handwriting. Recognition device.
  4.  前記認識部は、前記第1特徴ベクトルと前記第1識別ベクトルとの距離を、前記第2特徴ベクトルと前記第2識別ベクトルとの距離に優先して統合することを特徴とする請求項3に記載の認識装置。 The recognizing unit integrates the distance between the first feature vector and the first identification vector in preference to the distance between the second feature vector and the second identification vector. The recognition device described.
  5.  前記抽出部は、前記分割領域毎に前記実ストロークの座標データ間の方向の出現頻度分布を抽出し、抽出した前記出現頻度分布を1列に配列することにより前記第1特徴ベクトルを得るとともに、前記分割領域毎に前記仮想ストロークの座標データ間の方向の出現頻度分布を抽出し、抽出した前記出現頻度分布を1列に配列することにより第2特徴ベクトルを得ることを特徴とする請求項1に記載の認識装置。 The extraction unit extracts the appearance frequency distribution in the direction between the coordinate data of the actual stroke for each divided region, obtains the first feature vector by arranging the extracted appearance frequency distributions in one column, 2. The second feature vector is obtained by extracting an appearance frequency distribution in a direction between coordinate data of the virtual stroke for each of the divided regions and arranging the extracted appearance frequency distributions in one column. The recognition device described in 1.
  6.  ユーザーが筆跡を入力する入力部と、
     前記認識部の認識結果を表示する表示部と、
     を更に備えることを特徴とする請求項1に記載の認識装置。
    An input unit for the user to input handwriting,
    A display unit for displaying a recognition result of the recognition unit;
    The recognition device according to claim 1, further comprising:
PCT/JP2009/061615 2009-06-25 2009-06-25 Recognition device WO2010150383A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/JP2009/061615 WO2010150383A1 (en) 2009-06-25 2009-06-25 Recognition device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/JP2009/061615 WO2010150383A1 (en) 2009-06-25 2009-06-25 Recognition device

Publications (1)

Publication Number Publication Date
WO2010150383A1 true WO2010150383A1 (en) 2010-12-29

Family

ID=43386177

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2009/061615 WO2010150383A1 (en) 2009-06-25 2009-06-25 Recognition device

Country Status (1)

Country Link
WO (1) WO2010150383A1 (en)

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH01213772A (en) * 1988-02-22 1989-08-28 Oki Electric Ind Co Ltd On-line character recognizing system
JP2005165662A (en) * 2003-12-02 2005-06-23 Ntt Docomo Inc Image processor and image processing method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH01213772A (en) * 1988-02-22 1989-08-28 Oki Electric Ind Co Ltd On-line character recognizing system
JP2005165662A (en) * 2003-12-02 2005-06-23 Ntt Docomo Inc Image processor and image processing method

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
AKINORI KAWAMURA: "Saisho Bunrui Ayamari Gakushu ni yoru Tokucho Sentakugata Moji Ninshiki", THE TRANSACTIONS OF THE INSTITUTE OF ELECTRO, vol. J81-D-11, no. 12, 25 December 1998 (1998-12-25), pages 2749 - 2756 *

Similar Documents

Publication Publication Date Title
US7013046B2 (en) Apparatus, method, and program for handwriting recognition
EP1564675B1 (en) Apparatus and method for searching for digital ink query
CN102449640B (en) Recognizing handwritten words
JP5071914B2 (en) Recognition graph
US8768062B2 (en) Online script independent recognition of handwritten sub-word units and words
US20080240569A1 (en) Character input apparatus and method and computer readable storage medium
WO1995008158A1 (en) Universal symbolic handwriting recognition system
US10346681B2 (en) Method and computing device for optically recognizing mathematical expressions
JP2009543204A (en) Handwritten symbol recognition method and apparatus
CN108701215B (en) System and method for identifying multi-object structures
JP4817297B2 (en) Character search device
US11393231B2 (en) System and method for text line extraction
JP2017090998A (en) Character recognizing program, and character recognizing device
US9250802B2 (en) Shaping device
JP6081606B2 (en) Electronic apparatus and method
Singh et al. Online handwritten Gurmukhi words recognition: An inclusive study
CN115311674A (en) Handwriting processing method and device, electronic equipment and readable storage medium
WO2010150383A1 (en) Recognition device
CN107912062B (en) System and method for overlaying handwriting
Sreeraj et al. k-NN based On-Line Handwritten Character recognition system
Abuzaraida et al. Online recognition of Arabic handwritten words system based on Alignments matching Algorithm
AU2020103527A4 (en) IPDN- Read Handwriting: Intelligent Process to Read Handwriting Using Deep Learning and Neural Networks
KR101438271B1 (en) Device and methode for handwriting character
JP2023164293A (en) Author determination system and author determination method
JP3221488B2 (en) Online character recognition device with verification function

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 09846515

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 09846515

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: JP