JPH06288712A - Edge detecting method - Google Patents

Edge detecting method

Info

Publication number
JPH06288712A
JPH06288712A JP10028493A JP10028493A JPH06288712A JP H06288712 A JPH06288712 A JP H06288712A JP 10028493 A JP10028493 A JP 10028493A JP 10028493 A JP10028493 A JP 10028493A JP H06288712 A JPH06288712 A JP H06288712A
Authority
JP
Japan
Prior art keywords
edge
difference
sum
predetermined width
pixels
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP10028493A
Other languages
Japanese (ja)
Other versions
JP2880041B2 (en
Inventor
Hideji Watari
秀二 渡
Haruhito Okago
治仁 尾篭
Tsutomu Oki
勉 大木
Fuminori Nakano
文則 中野
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nippon Steel Corp
Original Assignee
Nippon Steel Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nippon Steel Corp filed Critical Nippon Steel Corp
Priority to JP10028493A priority Critical patent/JP2880041B2/en
Publication of JPH06288712A publication Critical patent/JPH06288712A/en
Application granted granted Critical
Publication of JP2880041B2 publication Critical patent/JP2880041B2/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Landscapes

  • Length Measuring Devices By Optical Means (AREA)

Abstract

PURPOSE:To provide an edge detecting method by which the edge position of an object can be specified easily even if the difference of a brightness level between the edge of the object and its surrounding is small. CONSTITUTION:Not only a simple difference between picture elements adjoining to each other but also a sum of brightness signal levels in a predetermined width (n pieces of picture elements) before and after each picture element are acquired and then, a difference is acquired between the sum of the brightness signal level on the front side and the sum of the brightness signal level on the rear side, and by considering a point where the value becomes the largest as the edge position of an object even if a difference in the brightness signal levels between the edge of the subject and its surrounding is small, the edge position of the object can be easily and accurately detected, and reliability on recognized shape and measured dimension of the object is improved.

Description

【発明の詳細な説明】Detailed Description of the Invention

【0001】[0001]

【産業上の利用分野】本発明は、画像入力装置にて撮影
された1次元画像から対象物のエッジを検出するための
方法に関するものである。
BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to a method for detecting an edge of an object from a one-dimensional image taken by an image input device.

【0002】[0002]

【従来の技術】従来、撮影画像から対象物の形状、寸法
等を計測するべく、画像入力装置として1次元CCDカ
メラを用い、これを揺動させて連続的に撮影を行い、画
像処理する装置が例えば特開昭61−209305号公
報に開示されている。
2. Description of the Related Art Conventionally, in order to measure the shape, size, etc. of an object from a photographed image, a one-dimensional CCD camera has been used as an image input device, which is rocked to continuously photograph and process an image. Is disclosed, for example, in Japanese Patent Laid-Open No. 61-209305.

【0003】このような装置にあっては、まず撮影画像
の明るさから対象物のエッジを検出して形状を認識する
と共にエッジ間の距離を演算して各寸法を求めるように
していた。例えば、加熱炉から粗圧延工程に向かう赤熱
状態の鋼材を対象物とした場合、画像処理装置に取り込
まれた赤外線映像データの隣接する画素間の出力(輝度
レベル)の差分を初期座標から最終座標まで演算し、こ
の差分が最大となる点を鋼材のスタートエッジとし、最
小となる点(負の値)をエンドエッジとしていた。
In such an apparatus, first, the edge of the object is detected from the brightness of the photographed image to recognize the shape, and the distance between the edges is calculated to obtain each dimension. For example, when the object is a red hot steel material that goes from the heating furnace to the rough rolling process, the difference in the output (brightness level) between adjacent pixels of the infrared image data captured by the image processing device is calculated from the initial coordinate to the final coordinate. Up to this point, the point at which this difference is the maximum is the start edge of the steel material, and the point at which this difference is the minimum (negative value) is the end edge.

【0004】しかしながら、鋼材端部の温度が低い場
合、鋼材とその周囲との温度差が小さく、エッジ部分の
発光輝度レベルの立ち上がりが緩やかになることからエ
ッジ位置の特定が難しいと云う問題があった。特に、近
年のCCDカメラの発達により取り込み画像の分解能が
高まっていることから、互いに隣接する前後1つづつの
画素の輝度レベルの差を見るだけではエッジ位置の特定
は一層困難であった。
However, when the temperature of the end of the steel material is low, there is a problem that it is difficult to specify the edge position because the temperature difference between the steel material and its surroundings is small and the emission luminance level at the edge portion rises gently. It was In particular, since the resolution of the captured image is increasing due to the recent development of CCD cameras, it is more difficult to specify the edge position only by looking at the difference in the brightness levels of the pixels adjacent to each other.

【0005】[0005]

【発明が解決しようとする課題】本発明はこのような従
来技術の問題点に鑑みなされたものであり、その主な目
的は、対象物のエッジとその周囲との輝度レベルの差が
小さい場合でも対象物のエッジ位置を容易に特定するこ
とが可能なエッジ検出方法を提供することにある。
SUMMARY OF THE INVENTION The present invention has been made in view of the above problems of the prior art, and its main purpose is to reduce the difference in brightness level between the edge of an object and its surroundings. However, it is to provide an edge detection method capable of easily specifying the edge position of an object.

【0006】[0006]

【課題を解決するための手段】このような目的は、本発
明によれば、対象物を含む1次元画像から前記対象物の
エッジを検出する方法であって、走査方向に沿う各点の
前後所定幅の輝度信号レベルの和または積分値を求める
と共に前記前側所定幅の輝度信号レベルの和または積分
値と前記後側所定幅の輝度信号レベルの和または積分値
との差を求め、前記差が最大となる点を前記対象物のエ
ッジ位置とすることを特徴とするエッジ検出方法を提供
することにより達成される。
According to the present invention, such an object is a method for detecting an edge of a target object from a one-dimensional image including the target object, which is before and after each point along the scanning direction. The sum or integrated value of the luminance signal level of a predetermined width is obtained, and the difference between the sum or integrated value of the luminance signal level of the front side predetermined width and the sum or integrated value of the luminance signal level of the rear side predetermined width is obtained, and the difference is obtained. It is achieved by providing an edge detection method characterized in that the point at which is the maximum is the edge position of the object.

【0007】[0007]

【作用】このように、互いに隣接する画素間の単純な差
分ではなく、或る画素の前後の所定幅(n個の画素)の
輝度信号レベルの和を求めた後、前側の輝度信号レベル
の和と後側の輝度信号レベルの和との差を求め、その値
が最大となる点を対象物のエッジ位置とすることでエッ
ジの特定が容易になる。
As described above, after the sum of the luminance signal levels of a predetermined width (n pixels) before and after a certain pixel is obtained instead of the simple difference between adjacent pixels, the luminance signal level of the front side is calculated. The edge can be easily specified by obtaining the difference between the sum and the sum of the luminance signal levels on the rear side and setting the point having the maximum value as the edge position of the object.

【0008】このとき、対象物の輝度レベルとその周囲
の輝度レベルとの差が大きければnは小さくて良く、上
記各輝度レベルの差が大きければその差に応じてnを大
きくすることにより、処理速度が不必要に遅くなること
もない。
At this time, if the difference between the brightness level of the object and the surrounding brightness level is large, n may be small. If the difference between the brightness levels is large, n may be increased according to the difference. The processing speed does not unnecessarily slow down.

【0009】[0009]

【実施例】以下、本発明の好適実施例を添付の図面につ
いて詳しく説明する。
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT A preferred embodiment of the present invention will now be described in detail with reference to the accompanying drawings.

【0010】図1は、本発明が適用された画像入力装置
及び画像処理装置の構成を示す斜視図である。図1に示
される搬送コンベア1上を加熱炉から粗圧延工程に向か
う赤熱状態の厚板スラブ2が搬送されるようになってお
り、この厚板スラブ2の寸法を求めるために、厚板スラ
ブ2を所定位置にて停止し、搬送ロール1の上方にて所
定の高さに配置された画像入力装置としてのカメラ装置
3により撮影し、その厚板スラブ2の画像信号を画像処
理装置4に送り、該画像処理装置4にて画像処理を行う
ようにしている。
FIG. 1 is a perspective view showing the configurations of an image input device and an image processing device to which the present invention is applied. A red-hot thick plate slab 2 is conveyed from the heating furnace to the rough rolling process on the conveyer conveyor 1 shown in FIG. 1. In order to obtain the dimensions of the thick plate slab 2, the thick plate slab 2 is slab. 2 is stopped at a predetermined position, and is photographed by a camera device 3 as an image input device arranged at a predetermined height above the transport roll 1, and the image signal of the thick plate slab 2 is sent to the image processing device 4. The image processing device 4 sends the image and performs image processing.

【0011】カメラ装置3は、厚板スラブ2からの熱の
影響を回避するべく、厚板スラブ2の上方に設けられた
計測室内にて例えば厚板スラブ2から21m離れた高さ
に配置されており、図示されない架台により支持された
駆動モータ5と、その駆動軸に固着された揺動板6に一
体的に取り付けられた光学系部材としてのカメラ本体7
及びレンズユニット8とからなる。カメラ本体7の検出
部には図の実線で示される範囲を計測可能な1次元CC
Dが設けられており、レンズユニット8には、長焦点
(例えば焦点距離170mm)にて初期調節を行う際に
その調節を容易に行い得るように例えばズームレンズが
用いられている。実際には4096画素の1次元CCD
を用いて、分解能が1画素当たり1.0mmとなるよう
に設定されている。
The camera device 3 is arranged, for example, at a height 21 m away from the thick plate slab 2 in a measuring chamber provided above the thick plate slab 2 in order to avoid the influence of heat from the thick plate slab 2. And a camera body 7 as an optical system member integrally attached to a drive motor 5 supported by a mount (not shown) and an oscillating plate 6 fixed to the drive shaft.
And the lens unit 8. The detection unit of the camera body 7 has a one-dimensional CC capable of measuring the range indicated by the solid line in the figure.
D is provided, and for the lens unit 8, for example, a zoom lens is used so that the initial adjustment can be easily performed at a long focus (for example, a focal length of 170 mm). Actually a one-dimensional CCD with 4096 pixels
Is used to set the resolution to 1.0 mm per pixel.

【0012】そして、駆動モータ5により揺動盤6を回
動することにより、カメラ本体7及びレンズユニット8
が図の矢印Aに示される向きに揺動し、その揺動方向に
所定のピッチ毎に厚板スラブ2の所定方向の画像を入力
する。このとき、撮影画界は想像線のようになる。ま
た、図示は省略するが、図1に示すスキャニングの方向
と直交する方向の長さを計測するべく、1次元CCDの
計測方向及び揺動方向をそれぞれ図示したものとは直交
させたカメラ装置も設けられている。
Then, the swing plate 6 is rotated by the drive motor 5, whereby the camera body 7 and the lens unit 8 are rotated.
Oscillates in the direction indicated by arrow A in the figure, and images of the slab 2 in the predetermined direction are input at predetermined pitches in the oscillating direction. At this time, the shooting image world becomes like an imaginary line. Although not shown, there is also a camera device in which the measuring direction and the swinging direction of the one-dimensional CCD are orthogonal to those shown in order to measure the length in the direction orthogonal to the scanning direction shown in FIG. It is provided.

【0013】尚、実際には分解能を確保するために複数
台の1次元CCDを組み合わせて使用し、各CCDにて
上記領域を分割してスキャニングしても良い。
In practice, a plurality of one-dimensional CCDs may be used in combination in order to secure the resolution, and the above areas may be divided and scanned by each CCD.

【0014】次に、厚板スラブ2のエッジ位置の検出処
理の手順を説明する。
Next, the procedure for detecting the edge position of the thick plate slab 2 will be described.

【0015】本実施例では、x方向の画素数を例えば4
096個ずつとし、座標(x,y)を(0,0)−(4
095,499)として説明する(図2)。処理の概要
としては、図1及び図2に於けるy方向座標を固定して
座標(0,Y)から座標(4095,Y)まで、或る座
標(X,Y)について、その座標(X,Y)を含むx方
向前側n個(所定幅)の座標の輝度レベルの和PF
(X,Y)を求め、続いてその座標(X,Y)を含むx
方向後側n個の座標の輝度レベルの和PR(X,Y)を
求め、これらPF(X,Y)とPR(X,Y)との差D
(X,Y)を求める。同様な処理を座標(0,Y)から
座標(4095,Y)まで行い、差D(X,Y)の絶対
値のうちの極大値を各々y座標=Yに於ける厚板スラブ
2のエッジと判断する。尚、実際には座標(0,Y)か
ら座標(4095,Y)まで全て上記処理を行う必要は
ない。
In this embodiment, the number of pixels in the x direction is set to, for example, four.
096 pieces each, and the coordinates (x, y) are (0, 0)-(4
095, 499) (FIG. 2). The outline of the processing is as follows. With the y-direction coordinates in FIGS. 1 and 2 fixed, coordinates (0, Y) to coordinates (4095, Y), for a certain coordinates (X, Y), the coordinates (X , Y), the sum PF of the brightness levels of the n coordinates (predetermined width) on the front side in the x direction including
(X, Y) is obtained, and then x including the coordinates (X, Y)
The sum PR (X, Y) of the brightness levels of the n coordinates on the rear side in the direction is obtained, and the difference D between these PF (X, Y) and PR (X, Y) is calculated.
Find (X, Y). Similar processing is performed from the coordinate (0, Y) to the coordinate (4095, Y), and the maximum value of the absolute values of the difference D (X, Y) is the edge of the thick plate slab 2 at the y coordinate = Y. To judge. In addition, it is not necessary to actually perform the above process from the coordinates (0, Y) to the coordinates (4095, Y).

【0016】以下に、本実施例に於ける処理手順を図3
のフローチャート沿って説明する。まず画像を画像処理
装置4に取り込んだら(ステップ1)、まず厚板スラブ
2の平均輝度レベルに応じて後記する演算画素数nを決
定する(ステップ2)。
The processing procedure in this embodiment will be described below with reference to FIG.
This will be described with reference to the flowchart in FIG. First, when the image is captured in the image processing device 4 (step 1), first, the number of calculated pixels n to be described later is determined according to the average luminance level of the thick plate slab 2 (step 2).

【0017】ここで、本実施例では鋼材の平均輝度レベ
ルが256階長の70以上(800℃以上)である場合
には演算画素数nを4、鋼材の平均輝度レベルが70未
満(800℃未満)である場合にはnを10としてい
る。しかしながら、実際には鋼材の色や環境に応じて演
算画素数nの値を変えて良いことは云うまでもない。ま
た、CCDによるスキャニングを行う場合、その出力用
アンプを複数用いることが多いが、各アンプに個体差が
あることからその出力値が変化することとなる。例えば
アンプを2つ用いる場合、奇数番目の画素と偶数番目の
画素とで使用するアンプを分けており、即ち奇数番目の
画素と偶数番目の画素とで出力値が変化することとな
る。そこで、演算画素数nをアンプの数若しくはその倍
数とすることでアンプによる検出誤差を排除することが
でき、より正確な検出を行うことができる。
Here, in this embodiment, when the average luminance level of the steel material is 70 or more (800 ° C. or more) of 256 floor length, the number of calculation pixels n is 4, and the average luminance level of the steel material is less than 70 (800 ° C.). N is 10. However, it goes without saying that the value of the number of calculated pixels n may actually be changed according to the color and environment of the steel material. Further, when performing scanning by CCD, a plurality of output amplifiers are often used, but the output value changes due to individual differences in each amplifier. For example, when two amplifiers are used, the amplifiers used are divided into odd-numbered pixels and even-numbered pixels, that is, the output value changes between odd-numbered pixels and even-numbered pixels. Therefore, by setting the number of calculation pixels n to the number of amplifiers or a multiple thereof, it is possible to eliminate the detection error due to the amplifiers, and it is possible to perform more accurate detection.

【0018】次に、変数Y=1として(ステップ3)、
座標(0,Y)からXの値をプラス方向にシフトさせ
て、上記PF(X,Y)、PR(X,Y)、D(X,
Y)を行って一方のエッジ(スタートエッジ)を求め
(ステップ4)、次に(4095,Y)からXの値をマ
イナス方向にシフトさせて、上記PF(X,Y)、PR
(X,Y)、D(X,Y)を求める処理を行って他方の
エッジ(エンドエッジ)を求める。このとき、スタート
エッジ、エンドエッジはD(X,Y)が最大となる点で
ある。
Next, the variable Y = 1 is set (step 3),
By shifting the value of X from the coordinates (0, Y) in the positive direction, the above PF (X, Y), PR (X, Y), D (X,
Y) is performed to obtain one edge (start edge) (step 4), then the value of X is shifted from (4095, Y) in the negative direction, and the above PF (X, Y), PR
The other edge (end edge) is obtained by performing processing for obtaining (X, Y) and D (X, Y). At this time, the start edge and the end edge are points where D (X, Y) becomes maximum.

【0019】図4に示すように、最大となるD(X,
Y)からシフト方向の200画素先で演算を打ち切るこ
とで処理速度が向上する。この200画素は変更可能な
設定値である。これを式で示すと、
As shown in FIG. 4, the maximum D (X,
The processing speed is improved by stopping the calculation at 200 pixels in the shift direction from Y). This 200 pixels is a changeable setting value. If this is shown by the formula,

【0020】[0020]

【数1】 [Equation 1]

【0021】の演算を座標(0,Y)から座標(409
5,Y)方向へXの値をプラス方向にシフトさせて行
い、その中のD(X,Y)の最大値から、一方のエッジ
を求め、
From the coordinates (0, Y) to the coordinates (409
The value of X is shifted in the positive direction in the (5, Y) direction, and one edge is obtained from the maximum value of D (X, Y) among them.

【0022】[0022]

【数2】 [Equation 2]

【0023】の演算を座標(4095,Y)から座標
(0,Y)方向へXの値をマイナス方向にシフトさせて
行い、その中のD(X,Y)の最大値から、他方のエッ
ジを求める。次に、y座標を1つシフトさせて(ステッ
プ6)、座標(0,Y+1)から座標(4095,Y+
1)まで同様な処理を行う。このようにして、y座標=
0から499まで上記処理を繰り返し(ステップ4〜ス
テップ7)、y方向に沿う厚板スラブ2のエッジを求め
る。
The calculation of is performed by shifting the value of X from the coordinate (4095, Y) to the coordinate (0, Y) direction in the negative direction, and from the maximum value of D (X, Y) in the value, the other edge is calculated. Ask for. Next, the y coordinate is shifted by 1 (step 6), and the coordinates (0, Y + 1) to the coordinates (4095, Y +) are shifted.
The same processing is performed up to 1). In this way, y coordinate =
The above processing is repeated from 0 to 499 (step 4 to step 7) to obtain the edge of the thick plate slab 2 along the y direction.

【0024】y方向に沿う厚板スラブ2のエッジが全て
求まったらX座標とy座標とを交換して、X方向に沿う
厚板スラブ2のエッジ全て求め(ステップ8〜ステップ
12)、ステップ13、14にて厚板スラブ2のX方向
及びy方向長さを演算し、その結果を表示し(ステップ
15)、記憶して処理を終了する。
When all the edges of the thick plate slab 2 along the y direction have been found, the X coordinate and the y coordinate are exchanged to find all the edges of the thick plate slab 2 along the X direction (steps 8 to 12), step 13 , 14 calculates the lengths of the thick plate slab 2 in the X and y directions, displays the results (step 15), stores the results, and ends the process.

【0025】尚、本実施例では画像入力装置にCCDカ
メラを用い、デジタル処理を行ったが、通常のアナログ
カメラを用い、n個の画素を所定幅と置き換えて輝度信
号を積算するような構成としても良いことは云うまでも
ない。
In this embodiment, a CCD camera is used as the image input device to perform digital processing. However, a normal analog camera is used to replace the n pixels with a predetermined width and integrate the luminance signal. It goes without saying that it is also good.

【0026】[0026]

【発明の効果】以上説明したように本発明によるエッジ
検出方法によれば、互いに隣接する画素間の単純な差分
ではなく、互いに隣接する画素間の単純な差分ではな
く、或る画素の前後の所定幅(n個の画素)の輝度信号
レベルの和を求めた後、前側の輝度信号レベルの和と後
側の輝度信号レベルの和との差を求め、その値が最大と
なる点を対象物のエッジ位置とすることで、対象物のエ
ッジとその周囲との輝度信号レベルの差が小さい場合で
も対象物のエッジ位置を容易に、かつ確実に検出するこ
とができ、対象物の認識形状、各測定寸法などの信頼性
が向上する。
As described above, according to the edge detection method of the present invention, it is not a simple difference between pixels adjacent to each other, but a simple difference between pixels adjacent to each other. After obtaining the sum of the luminance signal levels of a predetermined width (n pixels), the difference between the sum of the front side luminance signal levels and the sum of the rear side luminance signal levels is obtained, and the point where the value is maximum is targeted. By setting the edge position of the object, the edge position of the object can be easily and reliably detected even if the difference in the luminance signal level between the edge of the object and its surroundings is small. , The reliability of each measurement dimension is improved.

【図面の簡単な説明】[Brief description of drawings]

【図1】本発明が適用された画像入力装置及び画像処理
装置の構成を示す斜視図である。
FIG. 1 is a perspective view showing configurations of an image input device and an image processing device to which the present invention is applied.

【図2】図1の厚板スラブの平面図である。2 is a plan view of the plank slab of FIG. 1. FIG.

【図3】図1の厚板スラブのエッジを検出する手順を示
すフローチャートである。
FIG. 3 is a flowchart showing a procedure for detecting an edge of the thick plate slab of FIG.

【図4】演算結果を示す説明図である。FIG. 4 is an explanatory diagram showing calculation results.

【符号の説明】[Explanation of symbols]

1 搬送ロール 2 厚板スラブ 3 カメラ装置 4 画像処理装置 5 駆動モータ 6 揺動板 7 カメラ本体 8 レンズユニット 1 Transport Roll 2 Thick Plate Slab 3 Camera Device 4 Image Processing Device 5 Drive Motor 6 Oscillating Plate 7 Camera Body 8 Lens Unit

───────────────────────────────────────────────────── フロントページの続き (72)発明者 中野 文則 北九州市戸畑区大字中原46−59 新日本製 鐵株式会社機械・プラント事業部内 ─────────────────────────────────────────────────── ─── Continuation of the front page (72) Inventor Fuminori Nakano 46-59 Nakahara, Tobata-ku, Kitakyushu City Nippon Steel Corporation Machinery & Plant Division

Claims (2)

【特許請求の範囲】[Claims] 【請求項1】 対象物を含む1次元画像から前記対象
物のエッジを検出する方法であって、 走査方向に沿う各点の前後所定幅の輝度信号レベルの和
または積分値を求めると共に前記前側所定幅の輝度信号
レベルの和または積分値と前記後側所定幅の輝度信号レ
ベルの和または積分値との差を求め、前記差が最大とな
る点を前記対象物のエッジ位置とすることを特徴とする
エッジ検出方法。
1. A method for detecting an edge of an object from a one-dimensional image containing the object, wherein the sum or integrated value of brightness signal levels of a predetermined width before and after each point along the scanning direction is obtained and the front side is obtained. The difference between the sum or integrated value of the luminance signal level of a predetermined width and the sum or integrated value of the luminance signal level of the rear predetermined width is obtained, and the point where the difference is maximum is set as the edge position of the object. Characteristic edge detection method.
【請求項2】 前記対象物の平均の輝度信号レベルに
応じて前記所定幅を変化させることを特徴とする請求項
1に記載のエッジ検出方法。
2. The edge detecting method according to claim 1, wherein the predetermined width is changed according to an average luminance signal level of the object.
JP10028493A 1993-04-02 1993-04-02 Edge detection method Expired - Fee Related JP2880041B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP10028493A JP2880041B2 (en) 1993-04-02 1993-04-02 Edge detection method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP10028493A JP2880041B2 (en) 1993-04-02 1993-04-02 Edge detection method

Publications (2)

Publication Number Publication Date
JPH06288712A true JPH06288712A (en) 1994-10-18
JP2880041B2 JP2880041B2 (en) 1999-04-05

Family

ID=14269899

Family Applications (1)

Application Number Title Priority Date Filing Date
JP10028493A Expired - Fee Related JP2880041B2 (en) 1993-04-02 1993-04-02 Edge detection method

Country Status (1)

Country Link
JP (1) JP2880041B2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2013011454A (en) * 2011-06-28 2013-01-17 Nissan Motor Co Ltd Apparatus and method for detecting road surface unevenness
CN117911716A (en) * 2024-03-19 2024-04-19 天津医科大学总医院 Arthritis CT image feature extraction method based on machine vision

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2013011454A (en) * 2011-06-28 2013-01-17 Nissan Motor Co Ltd Apparatus and method for detecting road surface unevenness
CN117911716A (en) * 2024-03-19 2024-04-19 天津医科大学总医院 Arthritis CT image feature extraction method based on machine vision

Also Published As

Publication number Publication date
JP2880041B2 (en) 1999-04-05

Similar Documents

Publication Publication Date Title
CN1332556C (en) Image processor
EP1343332A2 (en) Stereoscopic image characteristics examination system
JP3701353B2 (en) Image acquisition device
JP6575232B2 (en) Steel plate shape measuring apparatus and method, and steel plate manufacturing apparatus and method using the same
JP2000011157A (en) Image pickup device
JPH06288712A (en) Edge detecting method
JP2003185419A (en) Method and apparatus for measuring warpage shape
JPH0745504A (en) Focusing position detecting method
CN113805304B (en) Automatic focusing system and method for linear array camera
JPH02194307A (en) Curvature shape measuring instrument for plate-like body
JPH0969973A (en) Position adjusting method for solid-state image pickup element
EP1676238B1 (en) A method for measuring dimensions by means of a digital camera
JPH1065940A (en) Image pickup device
JPH06294616A (en) Edge position correcting method
JP2014204198A (en) Image processing device, control method thereof and program
JP7028086B2 (en) Shape measuring device and shape measuring method
JPH08285550A (en) Instrument for measuring-steel plate displacement
JP6837723B2 (en) Plane shape measuring device
JP2005274325A (en) Optical type defect inspection method for metal band
JPH10326348A (en) Appearance inspection device
JP3252574B2 (en) Image processing device
JP2002139312A (en) Measuring method of earth-carrying quantity
JP3400859B2 (en) Defect pattern detection method and device
JPH05215534A (en) Adjusting method for surface inspecting camera mechanism
JPH08226805A (en) Method for focusing imaging device and non-contact measuring device utilizing the method

Legal Events

Date Code Title Description
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 19990119

LAPS Cancellation because of no payment of annual fees