JP2005346391A - Object recognition method - Google Patents

Object recognition method Download PDF

Info

Publication number
JP2005346391A
JP2005346391A JP2004165015A JP2004165015A JP2005346391A JP 2005346391 A JP2005346391 A JP 2005346391A JP 2004165015 A JP2004165015 A JP 2004165015A JP 2004165015 A JP2004165015 A JP 2004165015A JP 2005346391 A JP2005346391 A JP 2005346391A
Authority
JP
Japan
Prior art keywords
edge
edges
closed curve
recognition method
shape pattern
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP2004165015A
Other languages
Japanese (ja)
Other versions
JP4543759B2 (en
Inventor
Akihiro Tsukada
明宏 塚田
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Toyota Motor Corp
Original Assignee
Toyota Motor Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Toyota Motor Corp filed Critical Toyota Motor Corp
Priority to JP2004165015A priority Critical patent/JP4543759B2/en
Publication of JP2005346391A publication Critical patent/JP2005346391A/en
Application granted granted Critical
Publication of JP4543759B2 publication Critical patent/JP4543759B2/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Abstract

<P>PROBLEM TO BE SOLVED: To provide an object recognition method for improving discrimination capability about an object of a simple shape. <P>SOLUTION: From the image of an acquired object (a step S1), an internal edge and an external edge are detected (a step S2), and featured values are detected on the basis of the extracted external/internal outlines (a step S3) (steps S3 to S6), and the calculation of its similarity with a registered pattern (a step S7) and the specification of the pattern based on this is carried out (steps S8, S9). When any pattern can not be specified from the external/internal outlines, closed curves based on the edges are successively added, and the featured value detection and the similarity calculation are carried out (steps S10 to S21), so that the pattern can be specified. <P>COPYRIGHT: (C)2006,JPO&NCIPI

Description

本発明は画像中から抽出した物体を認識する物体認識方法に関する。   The present invention relates to an object recognition method for recognizing an object extracted from an image.

ロボットの視覚を用いた制御技術の一例として、視覚装置により得られた画像情報から物体を認識する技術の開発が勧められている。そして、このような物体の認識手法として予め登録されているパターンとの類似性を基にして物体を認識する知識ベースの物体認識技術が知られている。こうした知識ベースの物体認識技術においては、登録されているパターン中から対象物に類似するパターンを精度良く選別する必要がある。   As an example of a control technique using the vision of a robot, development of a technique for recognizing an object from image information obtained by a visual device is recommended. As a recognition method for such an object, a knowledge-based object recognition technique for recognizing an object based on similarity to a pattern registered in advance is known. In such knowledge-based object recognition technology, it is necessary to accurately select a pattern similar to the target object from registered patterns.

特許文献1に記載されている技術は、このような図形の分類・検索に関する技術である。図形の回転、拡縮、移動に不変な形状特徴と、対応する形状パラメータを求め、同じ形状特徴を持つモデル図形への座標変換パラメータを形状パラメータを用いて計算して、モデル識別子と座標変換パラメータの組にそれぞれ投票を行い、投票数の多いモデルに分類するものである。非特許文献1の技術も同様に輪郭から特徴量を抽出してモデルとの対比を行うものである。   The technique described in Patent Document 1 is a technique related to such graphic classification / retrieval. The shape feature that is invariant to the rotation, scaling, and movement of the figure and the corresponding shape parameter are obtained, the coordinate conversion parameter to the model figure with the same shape feature is calculated using the shape parameter, and the model identifier and coordinate conversion parameter Each group is voted and classified into models with many votes. Similarly, the technique of Non-Patent Document 1 extracts feature amounts from contours and compares them with models.

また、特許文献2の技術は、輪郭を形成する点を追跡することで、物体内部および外部の輪郭線を構成する閉曲線を抽出して、これを基にして物体の判別を行うものである。
特開2000−215315号公報 特開平7−220089号公報 Amit Sethi, et al. "Curve and Surface Duals and the Recognition of Curved 3D Objects from their Silhouette", International Journal of Computer Vision Vol. 58, No.1, 2004年6月,pp73-86
Further, the technique of Patent Document 2 extracts a closed curve that constitutes a contour line inside and outside an object by tracking points that form a contour, and performs object discrimination based on this.
JP 2000-215315 A Japanese Patent Laid-Open No. 7-220089 Amit Sethi, et al. "Curve and Surface Duals and the Recognition of Curved 3D Objects from their Silhouette", International Journal of Computer Vision Vol. 58, No.1, June 2004, pp73-86

このように輪郭を用いて、識別を行う場合、認識対象物の外部輪郭が球、楕円、直方体、円柱、円錐のような単純な形状であると、その差異の判別が難しくなる。例えば、缶とコップのように外形が円柱状の2つの物体を識別することが難しいという問題がある。また、形状が似通っていて表面の模様等の異なる物体間の識別も困難である。   In the case of discriminating using the contour in this way, if the external contour of the recognition object is a simple shape such as a sphere, an ellipse, a rectangular parallelepiped, a cylinder, or a cone, it is difficult to determine the difference. For example, there is a problem that it is difficult to distinguish two objects having a cylindrical outer shape such as a can and a cup. In addition, it is difficult to discriminate between objects having similar shapes and different surface patterns.

そこで本発明は、単純な形状の物体についても識別力を向上させた物体認識方法を提供することを課題とする。   Therefore, an object of the present invention is to provide an object recognition method that improves discrimination power even for an object having a simple shape.

上記課題を解決するため、本発明に係る物体認識方法は、画像のエッジを抽出することで、画像中の対象物のエッジを抽出し、エッジからさらに閉曲線を抽出して、これに基づいて物体の形状パターンを認識する物体認識方法において、抽出した内部エッジと外部エッジとを組み合わせることで閉曲線を生成し、生成した閉曲線に対して特徴量を抽出して形状パターンの認識を行うことを特徴とする。   In order to solve the above problems, an object recognition method according to the present invention extracts an edge of an image, extracts an edge of an object in the image, further extracts a closed curve from the edge, and based on the object In the object recognition method for recognizing the shape pattern, a closed curve is generated by combining the extracted internal edge and external edge, and the feature pattern is extracted from the generated closed curve to recognize the shape pattern. To do.

このように内部エッジと外部エッジを組み合わせて可能な閉曲線の組み合わせを求めることで、物体の外部輪郭、内部輪郭だけでなく、単独では閉曲線を構成しない内部のエッジ境界線等の情報も利用して特徴量を抽出する。これにより、輪郭情報のみを用いた場合より、特徴量の次元が高次になる。この内部エッジと外部エッジを組み合わせた閉曲線抽出は、物体の輪郭線抽出のアルゴリズムを拡張することにより行われる。   In this way, by finding combinations of closed curves that can be created by combining internal edges and external edges, not only external contours and internal contours of objects, but also information such as internal edge boundaries that do not constitute closed curves by themselves are used. Extract features. Thereby, the dimension of the feature amount becomes higher than when only the contour information is used. The closed curve extraction combining the inner edge and the outer edge is performed by extending an object contour extraction algorithm.

この特徴量抽出は複数の閉曲線からなる閉曲線群に対して行うものであり、各エッジを優先順位づけし、優先度の高いエッジの組み合わせにより得られる閉曲線からは形状パターンを所定の精度で特定できない場合に、優先度の劣るエッジを順次追加して閉曲線生成と特徴量算出を行うことで、形状パターンの特定を行うとよい。   This feature amount extraction is performed on a group of closed curves composed of a plurality of closed curves. Each edge is prioritized, and a shape pattern cannot be specified with a predetermined accuracy from a closed curve obtained by combining high-priority edges. In this case, the shape pattern may be specified by sequentially adding edges with inferior priorities to generate a closed curve and calculate a feature amount.

つまり、最初から抽出可能な全閉曲線群について特徴量を抽出するのではなく、認識結果に応じてエッジを追加することで、閉曲線を生成し、特徴量の追加算出を行う。   That is, instead of extracting feature values for all closed curve groups that can be extracted from the beginning, a closed curve is generated by adding an edge according to the recognition result, and additional calculation of feature values is performed.

エッジの優先順位づけは、輪郭線を構成するエッジと、内部エッジのうち信頼度が高いエッジの優先度を高くし、内部エッジのうち信頼度の低いエッジの優先度を低くする。この信頼度とは例えば、エッジごとのノイズ成分の比率により判定すればよい。   In the prioritization of edges, the priority of the edges constituting the contour line and the edge having high reliability among the internal edges is increased, and the priority of the edge having low reliability among the internal edges is decreased. This reliability may be determined by, for example, the ratio of noise components for each edge.

ここで、所定の精度とは、いずれかの形状パターンとの類似度が所定値を超える場合であることを意味するものとする。   Here, the predetermined accuracy means that the degree of similarity with any shape pattern exceeds a predetermined value.

あるいは、本発明に係る物体認識方法は、画像のエッジを抽出することで、画像中の対象物の輪郭線を抽出し、輪郭線に基づく特徴量を抽出して、抽出した特徴量に基づいて物体の形状パターンを認識する物体認識方法において、輪郭線由来の特徴量に基づいて類似度が所定値を超える形状パターンが見出せない場合には、さらに、類似度が所定値を超える形状パターンが見つかるまで、単独または複数の内部エッジまたは外部エッジにより構成される閉曲線中から、閉曲線群に未設定の閉曲線のうち信頼度の最も高い閉曲線を順次閉曲線群に追加し、閉曲線群に基づいて特徴量を抽出し、抽出した特徴量に基づいて物体の形状パターン認識を繰り返すことを特徴とする。ここでは、エッジに代えて閉曲線を追加していく構成をとる。   Alternatively, the object recognition method according to the present invention extracts an outline of an object in an image by extracting an edge of the image, extracts a feature amount based on the contour line, and based on the extracted feature amount In the object recognition method for recognizing the shape pattern of an object, when a shape pattern with a similarity exceeding a predetermined value cannot be found based on a feature amount derived from an outline, a shape pattern with a similarity exceeding a predetermined value is further found. From the closed curve consisting of one or more internal edges or external edges, the closed curve with the highest reliability among the closed curves not set in the closed curve group is sequentially added to the closed curve group, and the feature amount is calculated based on the closed curve group. Extracting and repeating object shape pattern recognition based on the extracted feature quantity. Here, a configuration is adopted in which a closed curve is added instead of an edge.

本発明によれば、外部輪郭線および内部輪郭線のみでは認識が困難な単純な形状の物体についても、内部エッジ等に由来する複数の閉曲線を用いて特徴量を抽出するので、特徴量の次元を高次にして、認識精度を向上させることができる。   According to the present invention, since the feature amount is extracted using a plurality of closed curves derived from the inner edge or the like even for an object having a simple shape that is difficult to recognize only by the outer contour line and the inner contour line, the dimension of the feature amount It is possible to improve the recognition accuracy.

また、最初から認識した全ての閉曲線を用いて特徴量を算出するのではなく、順次エッジまたは閉曲線を追加して特徴量算出と類似度判定を行うことで、実際に処理する閉曲線の数を抑えて、処理の最適化を図ることができるため、計算処理量の増大を抑制し、認識精度の向上と高速化を両立させることができる。   Also, instead of using all closed curves recognized from the beginning, feature quantities are calculated, and by sequentially adding edges or closed curves, feature quantity calculation and similarity determination are performed, the number of closed curves actually processed is reduced. Thus, optimization of the processing can be achieved, so that an increase in the amount of calculation processing can be suppressed, and both improvement in recognition accuracy and high speed can be achieved.

また、信頼度の高いエッジ・閉曲線から用いることで、特徴量に対する信用度を確保し、認識精度を確保することができる。   In addition, by using an edge / closed curve with a high degree of reliability, it is possible to ensure the reliability of the feature amount and to ensure the recognition accuracy.

以下、添付図面を参照して本発明の好適な実施の形態について詳細に説明する。説明の理解を容易にするため、各図面において同一の構成要素に対しては可能な限り同一の参照番号を附し、重複する説明は省略する。   DESCRIPTION OF EXEMPLARY EMBODIMENTS Hereinafter, preferred embodiments of the invention will be described in detail with reference to the accompanying drawings. In order to facilitate the understanding of the description, the same reference numerals are given to the same components in the drawings as much as possible, and duplicate descriptions are omitted.

図1は、本発明に係る物体認識方法を実施する画像処理装置の概略構成図である。この画像処理装置100は、対象物200を撮像するカメラ1と、カメラ1の映像信号を受信して画像処理を行う画像処理部2と、処理結果を表示するモニタ3とからなる。   FIG. 1 is a schematic configuration diagram of an image processing apparatus that implements an object recognition method according to the present invention. The image processing apparatus 100 includes a camera 1 that captures an image of an object 200, an image processing unit 2 that receives a video signal from the camera 1 and performs image processing, and a monitor 3 that displays a processing result.

カメラ1は、例えば、ビデオカメラであって、毎秒30フレームで取得した動画像を画像処理部2へ転送することができる。このときの映像信号は、例えば、1画面が640×480ドットで構成され、各画素は、RGBそれぞれ8ビットで表されるデジタル信号である。なお、カメラ1側でデジタル信号へと変換して画像処理部2へ転送する形態のほか、カメラ1に、アナログ映像信号(例えば、NTSC形式)を出力するタイプのものを使用し、画像処理部2内にA/D変換器を設ける構成や、画像処理部2との間に独立してA/D変換器を設ける構成を採用してもよい。ここでは、カメラ1が単体の場合を説明するが、複数台のカメラを用いて、対象物200の立体視を行ったり、複数のカメラからの情報を1つの画像処理部2により並行して処理するようにしてもよい。   The camera 1 is, for example, a video camera, and can transfer a moving image acquired at 30 frames per second to the image processing unit 2. The video signal at this time is, for example, a digital signal in which one screen is composed of 640 × 480 dots and each pixel is represented by 8 bits for each of RGB. In addition to the form in which the camera 1 converts it into a digital signal and transfers it to the image processing unit 2, an image processing unit that uses an analog video signal (for example, NTSC format) is used for the camera 1. A configuration in which an A / D converter is provided in 2 or a configuration in which an A / D converter is provided independently between the image processing unit 2 may be adopted. Here, a case where the camera 1 is a single unit will be described. However, the object 200 is stereoscopically viewed using a plurality of cameras, and information from a plurality of cameras is processed in parallel by one image processing unit 2. You may make it do.

画像処理部2は、CPU、ROM、RAM等によって構成される画像処理演算部20と、対比パターンのデータを蓄積記憶するための、ハードディスク等の記憶手段からなる情報格納部21とを有している。この画像処理部2は、物体認識処理に特化した専用のハードウェアとして構成されていてもよいし、パーソナルコンピュータやワークステーションといった汎用の計算機を用い、これらの計算機上で作動するソフトウェアによって本発明に係る物体認識方法を実現してもよい。   The image processing unit 2 includes an image processing calculation unit 20 configured by a CPU, a ROM, a RAM, and the like, and an information storage unit 21 including a storage unit such as a hard disk for accumulating and storing comparison pattern data. Yes. The image processing unit 2 may be configured as dedicated hardware specialized for object recognition processing, or uses a general-purpose computer such as a personal computer or a workstation, and software that operates on these computers. You may implement | achieve the object recognition method concerning.

次に、本発明に係る画像処理方法について具体的に説明する。図2は、この処理方法の具体的なフローチャートであり、図3〜図7は、個々の処理の詳細を説明する図である。   Next, the image processing method according to the present invention will be specifically described. FIG. 2 is a specific flowchart of this processing method, and FIGS. 3 to 7 are diagrams for explaining the details of individual processes.

まず、カメラから物体(対象物)の画像を取得し(ステップS1)、その画像からエッジ検出等によって物体の外部/内部のエッジを検出する(ステップS2)。図3は、取っ手付きのマグカップ201の画像に抽出したエッジ201h〜201hを重ね合わせて表示した例であり、図4は、コップ202の画像に抽出したエッジ202h〜202hを重ね合わせて表示した例である。ここで抽出するエッジは、いずれも単独または他のエッジと連結することで閉曲線を形成するエッジのみとする。 First, an image of an object (target object) is acquired from the camera (step S1), and external / internal edges of the object are detected from the image by edge detection or the like (step S2). FIG. 3 is an example in which the extracted edges 201h 1 to 201h 3 are superimposed and displayed on the image of the mug 201 with the handle, and FIG. 4 is an overlay of the extracted edges 202h 1 to 202h 3 on the image of the cup 202. It is an example displayed. The edges extracted here are only edges that form a closed curve by singly or connecting to other edges.

次に、抽出したエッジ201h〜201hまたは202h〜202h中からまず、外部輪郭線と内部輪郭線とを抽出する(ステップS3)。このときの外部輪郭線とは、物体をシルエット投影したときの投影像とその周囲の背景または他の物体との境界線(外側境界線)であり、内部輪郭線とは、投影像内部における投影像と背景または他の物体との境界線(内側境界線)である。したがって、マグカップ201の場合には、エッジ201hが外部輪郭線となり、エッジ201hが内部輪郭線となる。一方、コップ202の場合には、外部輪郭線202h、202hのみが検出される。この外部輪郭線と内部輪郭線はともに閉曲線である。 Next, first, an outer contour line and an inner contour line are extracted from the extracted edges 201h 1 to 201h 3 or 202h 1 to 202h 3 (step S3). The external contour line at this time is a boundary line (outer boundary line) between the projected image when the object is silhouette-projected and the surrounding background or other object (outer boundary line), and the internal contour line is a projection inside the projected image. A boundary line (inner boundary line) between the image and the background or another object. Therefore, when the mug 201, edge 201h 1 is an external contour, the edge 201h 2 is an internal contour. On the other hand, in the case of the cup 202, only the outer contour lines 202h 1 and 202h 2 are detected. Both the outer contour line and the inner contour line are closed curves.

次に、抽出した輪郭(外部輪郭線と内部輪郭線)上で勾配が同一となる点を求める(ステップS4)。ここで、勾配が同一となる点とは、図5に示されるように、輪郭の接線の傾き(勾配)が同一になる点、つまり、輪郭の接線が平行な関係を有する点を意味する(ここでは、マグカップ201の場合を例に示す)。そして、同一勾配点の接線間の距離d1〜dmを算出し(ステップS5)、このうちの最長の直線距離dmaxで各直線距離d1〜dmを除して無次元値に変換した値を特徴量(不変量)に設定する(ステップS6)。ここで、d1〜dmの順序は、対応する同一勾配点の組の順序に応じて設定される。実際には、このステップS4〜S6の計算を勾配の異なる多数の点、接線群に対して求める。このとき、勾配を所定角度ずつずらしながら(例えば、5度ずつ)勾配点、接線、距離を求めるとよい。   Next, a point where the gradient is the same on the extracted contour (external contour line and internal contour line) is obtained (step S4). Here, the point where the gradient is the same means a point where the gradient (gradient) of the tangent of the contour is the same, that is, a point where the tangent of the contour has a parallel relationship (see FIG. 5). Here, the case of the mug 201 is shown as an example). Then, distances d1 to dm between the tangents of the same gradient point are calculated (step S5), and a value obtained by dividing each straight line distance d1 to dm by the longest straight line distance dmax and converting it to a dimensionless value is a feature amount. (Invariant) is set (step S6). Here, the order of d1-dm is set according to the order of the group of the corresponding same gradient point. Actually, the calculations in steps S4 to S6 are obtained for a large number of points and tangent groups having different gradients. At this time, the gradient point, tangent, and distance may be obtained while shifting the gradient by a predetermined angle (for example, by 5 degrees).

このようにして求めた特徴量d1〜dmをm次元空間にプロットすると、図6に示されるように所定の曲線を描く(ここでは、特徴量が3次、つまりm=3の場合を例示している)。この特徴量抽出を視点を異ならせた画像について行い、全ての特徴量を同様にプロットすると、その特徴量は、図7に示されるように曲面形状となる。様々なパターンに対して、この曲面形状データに当たる特徴量データを求めておき、予め情報格納部21に蓄積しておく。そして、情報格納部21に蓄積されている特徴量データと、求めた対象物の特徴量データを比較することにより、類似度を算出する(ステップS7)。これは、図6の曲線が図7の曲面の一部に含まれているか否かを調べることに相当する。   When the feature amounts d1 to dm thus obtained are plotted in the m-dimensional space, a predetermined curve is drawn as shown in FIG. 6 (here, the case where the feature amount is cubic, ie, m = 3 is illustrated). ing). When this feature amount extraction is performed on images with different viewpoints and all the feature amounts are similarly plotted, the feature amount has a curved surface shape as shown in FIG. Feature amount data corresponding to the curved surface shape data is obtained for various patterns and stored in the information storage unit 21 in advance. Then, the similarity is calculated by comparing the feature amount data accumulated in the information storage unit 21 with the obtained feature amount data of the object (step S7). This corresponds to checking whether or not the curve of FIG. 6 is included in a part of the curved surface of FIG.

様々なパターンに対してこの類似度を求め、最大類似度を算出したら、その最大類似度と予め設定されている類似度のしきい値とを比較する(ステップS8)。最大類似度が類似度のしきい値を超えている場合には、物体は最大類似となるパターンを有すると特定し(ステップS9)、処理を終了する。   After obtaining the similarity for various patterns and calculating the maximum similarity, the maximum similarity is compared with a preset similarity threshold (step S8). If the maximum similarity exceeds the similarity threshold, it is specified that the object has a pattern that is the maximum similarity (step S9), and the process ends.

一方、最大類似度がしきい値以下の場合には、さらに、閉曲線群を追加して判定を行う。まず、輪郭以外に閉曲線群を構成しうるエッジが存在しているか否かを判定する(ステップS10)。   On the other hand, when the maximum similarity is less than or equal to the threshold value, a determination is made by adding a closed curve group. First, it is determined whether there is an edge that can form a closed curve group other than the outline (step S10).

輪郭以外に閉曲線群を構成しうるエッジが存在していると判定した場合、エッジの組み合わせで構成される閉曲線について優先付けを行う(ステップS11)。この閉曲線の優先順位は、エッジごとの信頼性の程度を基によればよい。つまり、連続性、太さ等を基にしてノイズの少ないエッジ同士を組み合わせた閉曲線の優先順位を高く設定すればよい。   If it is determined that there is an edge that can form a closed curve group other than the contour, prioritization is performed for the closed curve formed by a combination of edges (step S11). The priority order of the closed curves may be based on the degree of reliability for each edge. In other words, the priority order of closed curves obtained by combining edges with less noise based on continuity, thickness, etc. may be set high.

優先順位を設定したら、組み合わせた追加可能な閉曲線の数を変数nに設定し、追加する閉曲線の優先順位を示す変数iに初期値1を設定する(ステップS12)。そして、ループ処理を開始する。   After the priority order is set, the number of closed curves that can be added in combination is set to a variable n, and an initial value 1 is set to a variable i indicating the priority order of the closed curve to be added (step S12). Then, loop processing is started.

最初に、閉曲線群にi番目の優先順位を有する閉曲線を追加する(ステップS13)。追加した閉曲線を含む閉曲線群について、ステップS4〜S7と同様の手法により特徴量を設定し、類似度の算出を行う(ステップS14〜S17)。様々なパターンに対してこの類似度を求め、最大類似度を算出したら、再度求めた最大類似度と予め設定されている類似度のしきい値とを比較する(ステップS18)。最大類似度が類似度のしきい値を超えている場合には、ステップS9へと移行して、物体は最大類似となるパターンを有すると特定し、処理を終了する。   First, a closed curve having the i-th priority order is added to the closed curve group (step S13). For the closed curve group including the added closed curve, the feature amount is set by the same method as in steps S4 to S7, and the similarity is calculated (steps S14 to S17). When the similarity is obtained for various patterns and the maximum similarity is calculated, the maximum similarity obtained again is compared with a preset similarity threshold (step S18). When the maximum similarity exceeds the similarity threshold, the process proceeds to step S9, where it is specified that the object has a pattern that is the maximum similarity, and the process ends.

一方、最大類似度がしきい値以下の場合には、さらに、iがm以上か否かを判定することで、閉曲線をすべて追加済みか否かを判定する(ステップS20)。iがmに満たない場合には、閉曲線群を構成しうる全ての閉曲線の追加が終了していないことを意味するから、iに1を加算して(ステップS21)、ステップS13へと戻り、閉曲線を追加して特徴量算出と類似度判定を行う。   On the other hand, if the maximum similarity is less than or equal to the threshold value, it is further determined whether or not all closed curves have been added by determining whether i is greater than or equal to m (step S20). If i is less than m, it means that the addition of all the closed curves that can constitute the closed curve group has not been completed, so 1 is added to i (step S21), and the process returns to step S13. A closed curve is added to calculate feature values and determine similarity.

一方、iがmに達していた場合には、閉曲線群を構成しうる全ての閉曲線の追加が終了したことを意味する。この場合には、ステップS10で輪郭以外に閉曲線群を構成しうるエッジが存在しないと判定した場合と同様に、閉曲線群からの特徴量抽出に基づく物体認識ができないと判定して、ステップS22へと移行し、物体は未知の物体であると判定して、処理を終了する。   On the other hand, if i has reached m, it means that the addition of all the closed curves that can form the closed curve group has been completed. In this case, similarly to the case where it is determined in step S10 that there is no edge that can form the closed curve group other than the contour, it is determined that the object recognition based on the feature amount extraction from the closed curve group cannot be performed, and the process proceeds to step S22. And the process determines that the object is an unknown object and ends the process.

本発明によれば、マグカップ201の場合、図5に示されるように、特徴量の次元が4次元に増大する。外輪郭のみを用いる従来の手法では、特徴量は2次または3次となる(図8参照)。このため、従来の手法では、類似度(完全に一致する場合で100となる。)が84として識別が難しかった図9に示されるクローバー図形との対比の場合でも類似度が54となり、正確に識別することができる。このため、識別力が向上する。   According to the present invention, in the case of the mug 201, as shown in FIG. 5, the dimension of the feature amount increases to four dimensions. In the conventional method using only the outer contour, the feature amount is quadratic or cubic (see FIG. 8). Therefore, in the conventional method, the similarity (which is 100 when completely matched) is 84, and the similarity is 54 even when compared with the clover figure shown in FIG. Can be identified. For this reason, discrimination power improves.

さらに、コップ202の場合、図10(a)に示されるような輪郭情報(エッジ202h、202h)のみからでは、同形の缶や円柱との区別がつかない。この場合も、図10(b)に示されるエッジ202h、202hからなる閉曲線や、さらには、図10(c)に示されるエッジ202h、202hからなる閉曲線を順次追加して認識を行うことで、識別が可能となる。特に外形は単純な形状でも内部形状や模様に特徴のある物体の識別に有効となる。マグカップ201の場合も、さらに形状の類似した物体(例えば、表面の形状、模様が異なるマグカップ等)と識別する場合には、外部輪郭、内部輪郭のほかに、内部エッジと輪郭を組み合わせた閉曲線を追加して識別を行うとよい。 Further, in the case of the cup 202, it is impossible to distinguish from the same shape can and cylinder only from the contour information (edges 202h 1 and 202h 2 ) as shown in FIG. Again, closed curve and consisting of edges 202h 1, 202h 3 shown in FIG. 10 (b), further sequentially added to recognize a closed curve consisting of the edge 202h 2, 202h 3 that shown in FIG. 10 (c) By doing so, identification becomes possible. In particular, even if the outer shape is simple, it is effective for identifying an object having a feature in the internal shape or pattern. In the case of the mug 201 as well, when identifying an object having a similar shape (for example, a mug having a different surface shape or pattern), in addition to the outer contour and the inner contour, a closed curve combining the inner edge and the contour is used. It is good to add and identify.

ここでは、閉曲線を順次、追加していく実施形態を説明したが、多数のエッジを有する場合には、閉曲線を追加するのではなく、エッジを順次追加していき、追加されたエッジを含む閉曲線を求めて、それを対象にして特徴量を算出する形式をとってもよい。   Here, the embodiment in which closed curves are sequentially added has been described. However, when there are a large number of edges, instead of adding closed curves, the edges are added sequentially and closed curves including the added edges are included. May be obtained, and a feature amount may be calculated for that.

以上の説明では、特徴量として同一勾配を有する輪郭の接線間の距離を用いたが、特徴量はパターンの画面内の回転と拡縮に対して不変であれば、これに限られるものではない。例えば、曲率、色情報、ヒストグラム等を用いることができる。   In the above description, the distance between the tangent lines of the contour having the same gradient is used as the feature amount. However, the feature amount is not limited to this as long as the feature amount is invariable with respect to the rotation and expansion / contraction in the pattern screen. For example, curvature, color information, a histogram, etc. can be used.

例えば、図11は、対象画像が三つの色の異なる円203〜205を近接させて配置した図形の場合を示している。円203は内部が緑色、円204は内部が黄色、円205は内部が赤色に塗られている。この場合に図12に示されるように、外部輪郭203h〜205hのみを抽出して比較を行うと、色・模様の異なる図形等と識別を行うことができない。そこで、図13に示されるように、上述した距離情報に色情報を加味して特徴量として設定する(ここでは、色に変えて輪郭203h’〜205h’の濃度として示している。)ことで、色・模様の異なる画像についても識別が可能となる。この場合の色情報としては、輪郭位置の色値、色相、輝度等の情報を基にして設定を行うとよい。   For example, FIG. 11 shows a case where the target image is a figure in which three circles 203 to 205 having different colors are arranged close to each other. The circle 203 is painted green, the circle 204 is painted yellow, and the circle 205 is painted red. In this case, as shown in FIG. 12, if only the external contours 203h to 205h are extracted and compared, it is not possible to identify a figure or the like having a different color / pattern. Therefore, as shown in FIG. 13, color information is added to the above-described distance information and set as a feature amount (here, the density is shown as the density of contours 203 h ′ to 205 h ′ instead of color). It is also possible to identify images with different colors and patterns. The color information in this case may be set based on information such as the color value, hue, and luminance of the contour position.

本発明は、例えば、ロボットの視覚認識技術や、監視システム等に対して適用することができる。   The present invention can be applied to, for example, a robot visual recognition technique, a monitoring system, and the like.

本発明に係る物体認識方法を実施する画像処理装置の概略構成図である。It is a schematic block diagram of the image processing apparatus which implements the object recognition method which concerns on this invention. 本発明にかかる物体認識方法の処理を示すフローチャートである。It is a flowchart which shows the process of the object recognition method concerning this invention. 取っ手付きのマグカップの画像に抽出したエッジを重ね合わせて表示した例である。This is an example in which the extracted edge is superimposed and displayed on the image of the mug with the handle. コップの画像に抽出したエッジを重ね合わせて表示した例である。This is an example in which the extracted edges are superimposed on the cup image. マグカップからの特徴量抽出の様子を示す図である。It is a figure which shows the mode of the feature-value extraction from a mug. 画像から抽出した特徴量を3次元空間にプロットした様子を示すグラフである。It is a graph which shows a mode that the feature-value extracted from the image was plotted in three-dimensional space. 格納されている特徴量を3次元空間にプロットした様子を示すグラフである。It is a graph which shows a mode that the stored feature-value was plotted in three-dimensional space. 従来の手法によるマグカップからの特徴量抽出の様子を示す図である。It is a figure which shows the mode of the feature-value extraction from the mug by the conventional method. 比較対象のクローバー図形を示す図である。It is a figure which shows the clover figure of a comparison object. コップにおける閉曲線群の追加の様子を説明する図である。It is a figure explaining the mode of addition of the closed curve group in a cup. 色情報を有する対象画像の例示である。It is an illustration of the target image which has color information. 図11から抽出した輪郭情報を示す図である。It is a figure which shows the outline information extracted from FIG. 図11から抽出した色情報を含む輪郭情報を示す図である。It is a figure which shows the outline information containing the color information extracted from FIG.

符号の説明Explanation of symbols

1…カメラ、2…画像処理部、3…モニタ、20…画像処理演算部、21…情報格納部、100…画像処理装置、200…対象物、201…マグカップ、202…コップ、203〜205…円、201h〜201h、202h〜202h…エッジ。 DESCRIPTION OF SYMBOLS 1 ... Camera, 2 ... Image processing part, 3 ... Monitor, 20 ... Image processing calculating part, 21 ... Information storage part, 100 ... Image processing apparatus, 200 ... Object, 201 ... Mug, 202 ... Cup, 203-205 ... Circle, 201h 1 to 201h 3 , 202h 1 to 202h 3 ... Edge.

Claims (5)

画像のエッジを抽出することで、画像中の対象物のエッジを抽出し、エッジからさらに閉曲線を抽出して、これに基づいて物体の形状パターンを認識する物体認識方法において、
抽出した内部エッジと外部エッジとを組み合わせることで閉曲線を生成し、生成した閉曲線に対して特徴量を抽出して形状パターンの認識を行うことを特徴とする物体認識方法。
In the object recognition method for extracting the edge of the image, extracting the edge of the object in the image, further extracting the closed curve from the edge, and recognizing the shape pattern of the object based on this,
An object recognition method, wherein a closed curve is generated by combining an extracted internal edge and an external edge, and a feature amount is extracted from the generated closed curve to recognize a shape pattern.
前記特徴量抽出は複数の閉曲線からなる閉曲線群に対して行うものであり、各エッジを優先順位づけし、優先度の高いエッジの組み合わせにより得られる閉曲線からは形状パターンを所定の精度で特定できない場合に、優先度の劣るエッジを順次追加して閉曲線生成と特徴量算出を行うことで、形状パターンの特定を行うことを特徴とする請求項1記載の物体認識方法。   The feature amount extraction is performed on a group of closed curves including a plurality of closed curves. Each edge is prioritized, and a shape pattern cannot be specified with a predetermined accuracy from a closed curve obtained by a combination of edges having high priority. The object recognition method according to claim 1, wherein the shape pattern is specified by sequentially adding edges with inferior priorities to generate a closed curve and calculating a feature amount. 前記エッジの優先順位づけは、輪郭線を構成するエッジと、内部エッジのうち信頼度が高いエッジの優先度を高くし、内部エッジのうち信頼度の低いエッジの優先度を低くすることを特徴とする請求項2記載の物体認識方法。   The prioritization of the edges is characterized by increasing the priority of the edges constituting the contour line and the edge having high reliability among the internal edges and decreasing the priority of the edge having low reliability among the internal edges. The object recognition method according to claim 2. 前記所定の精度とは、いずれかの形状パターンとの類似度が所定値を超える場合であることを特徴とする請求項3記載の物体認識方法。   The object recognition method according to claim 3, wherein the predetermined accuracy is a case where the degree of similarity with any shape pattern exceeds a predetermined value. 画像のエッジを抽出することで、画像中の対象物の輪郭線を抽出し、輪郭線に基づく特徴量を抽出して、抽出した特徴量に基づいて物体の形状パターンを認識する物体認識方法において、
輪郭線由来の特徴量に基づいて類似度が所定値を超える形状パターンが見出せない場合には、さらに、類似度が所定値を超える形状パターンが見つかるまで、単独または複数の内部エッジまたは外部エッジにより構成される閉曲線中から、閉曲線群に未設定の閉曲線のうち信頼度の最も高い閉曲線を順次閉曲線群に追加し、閉曲線群に基づいて特徴量を抽出し、抽出した特徴量に基づいて物体の形状パターン認識を繰り返すことを特徴とする物体認識方法。
In an object recognition method for extracting an outline of an object in an image by extracting an edge of the image, extracting a feature amount based on the contour line, and recognizing an object shape pattern based on the extracted feature amount ,
If a shape pattern with a similarity exceeding the predetermined value cannot be found based on the feature value derived from the contour line, it is further possible to use one or more internal edges or external edges until a shape pattern with a similarity exceeding the predetermined value is found. From the configured closed curves, the closed curves with the highest reliability among the closed curves not set in the closed curve group are sequentially added to the closed curve group, and feature quantities are extracted based on the closed curve groups. An object recognition method characterized by repeating shape pattern recognition.
JP2004165015A 2004-06-02 2004-06-02 Object recognition method Expired - Fee Related JP4543759B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2004165015A JP4543759B2 (en) 2004-06-02 2004-06-02 Object recognition method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2004165015A JP4543759B2 (en) 2004-06-02 2004-06-02 Object recognition method

Publications (2)

Publication Number Publication Date
JP2005346391A true JP2005346391A (en) 2005-12-15
JP4543759B2 JP4543759B2 (en) 2010-09-15

Family

ID=35498707

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2004165015A Expired - Fee Related JP4543759B2 (en) 2004-06-02 2004-06-02 Object recognition method

Country Status (1)

Country Link
JP (1) JP4543759B2 (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2014178941A (en) * 2013-03-15 2014-09-25 Nec Corp Image collation system
US9219891B2 (en) 2012-01-13 2015-12-22 Brain Co., Ltd. Object identification apparatus
US9256802B2 (en) 2010-11-26 2016-02-09 Nec Corporation Object or shape information representation method

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04112277A (en) * 1990-08-31 1992-04-14 Toyobo Co Ltd Shape discriminating method
JPH07220089A (en) * 1994-01-31 1995-08-18 Matsushita Electric Ind Co Ltd Position detecting method
JPH11203461A (en) * 1997-10-28 1999-07-30 Ricoh Co Ltd Graph sorting method and system, graph retrieving method and system, graph sorting feature extracting method, graph sorting table preparing method, information recording medium, and method for evaluating similarity or difference between graphs

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH04112277A (en) * 1990-08-31 1992-04-14 Toyobo Co Ltd Shape discriminating method
JPH07220089A (en) * 1994-01-31 1995-08-18 Matsushita Electric Ind Co Ltd Position detecting method
JPH11203461A (en) * 1997-10-28 1999-07-30 Ricoh Co Ltd Graph sorting method and system, graph retrieving method and system, graph sorting feature extracting method, graph sorting table preparing method, information recording medium, and method for evaluating similarity or difference between graphs

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9256802B2 (en) 2010-11-26 2016-02-09 Nec Corporation Object or shape information representation method
US9219891B2 (en) 2012-01-13 2015-12-22 Brain Co., Ltd. Object identification apparatus
JP2014178941A (en) * 2013-03-15 2014-09-25 Nec Corp Image collation system

Also Published As

Publication number Publication date
JP4543759B2 (en) 2010-09-15

Similar Documents

Publication Publication Date Title
EP2063393B1 (en) Color classifying method, color recognizing method, color classifying device, color recognizing device, color recognizing system, computer program, and recording medium
JP4284288B2 (en) Pattern recognition apparatus and method
CN107437060B (en) Object recognition apparatus, object recognition method, and program
US8447114B2 (en) Method and apparatus for calculating pixel features of image data
US20150261803A1 (en) Edge-based recognition, systems and methods
US20060204035A1 (en) Method and apparatus for tracking a movable object
CN106529573A (en) Real-time object detection method based on combination of three-dimensional point cloud segmentation and local feature matching
KR101567792B1 (en) System and method for describing image outlines
US9256802B2 (en) Object or shape information representation method
US8705864B2 (en) Marker generation device, marker generation detection system, marker generation detection device, marker, marker generation method, and program
WO2012046426A1 (en) Object detection device, object detection method, and object detection program
JP2014010633A (en) Image recognition device, image recognition method, and image recognition program
JP2013218605A (en) Image recognition device, image recognition method, and program
JP2010504575A (en) Method and apparatus for recognizing face and face recognition module
JP4543759B2 (en) Object recognition method
JP2002133413A (en) Method and device for identifying three-dimensional object using image processing
KR101521136B1 (en) Method of recognizing face and face recognition apparatus
CN107085725B (en) Method for clustering image areas through LLC based on self-adaptive codebook
JP4311278B2 (en) Object recognition method
JP6717049B2 (en) Image analysis apparatus, image analysis method and program
JP6278757B2 (en) Feature value generation device, feature value generation method, and program
KR101601564B1 (en) Face detection method using circle blocking of face and apparatus thereof
JP2007004709A (en) Object pattern detecting method and device thereof
CN113095147A (en) Skin area detection method, system, image processing terminal and storage medium
KR100977756B1 (en) A integral image generation method for skin region based face detection

Legal Events

Date Code Title Description
A621 Written request for application examination

Free format text: JAPANESE INTERMEDIATE CODE: A621

Effective date: 20070529

A977 Report on retrieval

Free format text: JAPANESE INTERMEDIATE CODE: A971007

Effective date: 20100308

A131 Notification of reasons for refusal

Free format text: JAPANESE INTERMEDIATE CODE: A131

Effective date: 20100316

A521 Request for written amendment filed

Free format text: JAPANESE INTERMEDIATE CODE: A523

Effective date: 20100513

TRDD Decision of grant or rejection written
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20100608

A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20100621

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20130709

Year of fee payment: 3

R151 Written notification of patent or utility model registration

Ref document number: 4543759

Country of ref document: JP

Free format text: JAPANESE INTERMEDIATE CODE: R151

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20130709

Year of fee payment: 3

LAPS Cancellation because of no payment of annual fees