JPH0846864A - Detection of video cutting point - Google Patents

Detection of video cutting point

Info

Publication number
JPH0846864A
JPH0846864A JP18202494A JP18202494A JPH0846864A JP H0846864 A JPH0846864 A JP H0846864A JP 18202494 A JP18202494 A JP 18202494A JP 18202494 A JP18202494 A JP 18202494A JP H0846864 A JPH0846864 A JP H0846864A
Authority
JP
Japan
Prior art keywords
time
cut point
image
axis
detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
JP18202494A
Other languages
Japanese (ja)
Inventor
Yukinobu Taniguchi
行信 谷口
Yoshinobu Tonomura
佳伸 外村
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nippon Telegraph and Telephone Corp
Original Assignee
Nippon Telegraph and Telephone Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nippon Telegraph and Telephone Corp filed Critical Nippon Telegraph and Telephone Corp
Priority to JP18202494A priority Critical patent/JPH0846864A/en
Publication of JPH0846864A publication Critical patent/JPH0846864A/en
Pending legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/85Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression
    • H04N19/87Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using pre-processing or post-processing specially adapted for video compression involving scene cut or scene change detection in combination with video compression
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/134Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the element, parameter or criterion affecting or controlling the adaptive coding
    • H04N19/142Detection of scene cut or scene change
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/10Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding
    • H04N19/169Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding
    • H04N19/179Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using adaptive coding characterised by the coding unit, i.e. the structural portion or semantic portion of the video signal being the object or the subject of the adaptive coding the unit being a scene or a shot

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Studio Circuits (AREA)

Abstract

PURPOSE:To provide the method of detection of a video cutting point capable of reducing a detection error due to the motion of an object and hand shake in cut and having superior detection performance. CONSTITUTION:A video data string is constituted (101) in a space and time image that is the three-dimensional arrangement of image plane coordinates (x), (y) and a time (t), and the gradient vectors (DELTAxI, DELTAyI, DELTAtI) of the space and time image at a certain time (t) are calculated (102). The gradient vector is made nearly parallel with the axis (t) since the change of luminance and color is small in the directions of (x), (y) and large in the direction of axis (t) at a cutting point due to the steep change of a pattern. Therefore, a value representing the size of the number of gradient vectors set nearly parallel with the axis (t) is found from the count values count1, count2 of a picture element which satisfy a conditional equation by using equation of diff=count2/count1 (103-105). When diff exceeds a thresh-old value theta, this time is judged as the cutting point (106). Such operation is successively performed with respect to the time (t) (107).

Description

【発明の詳細な説明】Detailed Description of the Invention

【0001】[0001]

【産業上の利用分野】映像、すなわち複数枚の画像デー
タの列からそのカット点(シーンが切り替わる点)を検
出する方法に関する。
BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention relates to a method for detecting a cut point (point at which a scene changes) from a video, that is, a sequence of a plurality of image data.

【0002】[0002]

【従来の技術】映像中でシーンが切り替わる点をカット
点という。映像カット点検出方法は、シーンチェンジ検
出とも呼ばれ、さまざまな方法が提案されている。
2. Description of the Related Art The point at which a scene changes in a video is called a cut point. The image cut point detection method is also called scene change detection, and various methods have been proposed.

【0003】代表的な方法として、時間的に隣合う二枚
の画像It,It-1の対応する画素における輝度値の差を
計算し、その絶対値の和をD(t)とし、D(t)があ
る与えられた閾値よりも大きい時、tをカット点とみな
すものがある(大辻、外村、大庭:輝度情報を使った動
画像ブラウジング。電気情報通信学会技術報告、IE9
0−103,1911.)。
As a typical method, a difference in luminance value between corresponding pixels of two temporally adjacent images I t and I t-1 is calculated, and the sum of absolute values thereof is defined as D (t), When D (t) is larger than a given threshold, some consider t as a cut point (Otsuji, Tonomura, Ohiwa: Moving image browsing using luminance information. IEICE Technical Report, IE9
0-103, 1911. ).

【0004】フレーム間差分の代りに画素変化面積、輝
度ヒストグラム差分、ブロック別色相関、χ2検定量な
どがD(t)として使われる(大辻、外村:映像カット
自動検出方式の検討。テレビジョン学会技術報告、Vo
l.16,No.43,pp.7−12)。
Pixel change area, luminance histogram difference, block-wise color correlation, χ 2 test amount, etc. are used as D (t) instead of inter-frame difference (Otsuji, Tonomura: Study of automatic video cut detection method. Technical Report of the Society of John, Vo
l. 16, No. 43, pp. 7-12).

【0005】また、D(t)をそのまま閾値処理するの
ではなく、各種時間フィルタをD(t)に対して作用し
た結果を閾値処理する方法もある。この方法は、映像の
中に激しく動く動体やフラッシュ光があっても誤検出を
生じにくいという特徴を持つ(K.Otsuji an
d Y.Yonomura:ProjectionDe
tecting Filter for Video
Cut Detection.Proc. of AC
M Multimedia 93,1993,pp.2
51−257)。
There is also a method of thresholding the result of the action of various temporal filters on D (t), instead of directly thresholding D (t). This method is characterized in that erroneous detection is unlikely to occur even if there is a moving object or flash light that moves rapidly in the image (K. Otsuji an.
d Y. Yonomura: ProjectionDe
tecting Filter for Video
Cut Detection. Proc. of AC
M Multimedia 93, 1993, pp. Two
51-257).

【0006】[0006]

【発明が解決しようとする課題】以上述べた従来の技術
においては、被写体が動いたり、手ぶれによって画面全
体が動いたりしたときに、上記で説明した変化量が上昇
するために、検出性能が劣化するという問題点があっ
た。
In the conventional technique described above, when the subject moves or the entire screen moves due to camera shake, the amount of change described above increases, and the detection performance deteriorates. There was a problem to do.

【0007】この問題点について図7、図8を使って説
明する。図7(a)に示すような画像データ列を考え
る。この例では、白色(輝度値=255)の四角形(高
さH、幅W)が黒色(輝度値=0)の背景の上をΔxず
つ左へ移動していき、時刻tcで四角形の色が灰色(輝
度値=127)に変化する。ここで、時刻tcがカット
点である。変化量D(t)をここでは時間的に隣合う二
枚の画像It,It-1の対応する画素における輝度値の差
の絶対値の和として算出する。したがって、時刻t(≠
c)における変化量D(t)は、図7(b)で四角形
のずれた部分の面積に黒と白の輝度値の差255をかけ
たものになるので、D(t)=255*2*Δx*Hと
なる。カット点t=tcでは、白色と灰色の輝度値の差
が255−127=128であることから、D(tc
=128*W*Hとなる。図8(a),(b)のグラフ
の横軸は時刻tを表し、縦軸は変化量D(t)を表す。
四角形の移動速度Δxが小さいときには、図8(a)に
示すように、D(t)はD(tc)に比べて十分小さい
のでカット点を正しく検出できる。しかし、移動速度Δ
xが大きくなるにしたがって、図8(b)に示すよう
に、D(t)が全体的に上昇するため、D(tc)とD
(t)の区別が困難となる。このため、従来の方法では
被写体の動きが激しいところをカット点と見誤ることが
あった。カメラの手ぶれによっても、同様の状況が発生
し、カット点検出の性能劣化の一因となっていた。
This problem will be described with reference to FIGS. 7 and 8. Consider an image data string as shown in FIG. In this example, a white (brightness value = 255) quadrangle (height H, width W) is moved to the left by Δx on a black (brightness value = 0) background, and the color of the quadrangle at time t c. Changes to gray (brightness value = 127). Here, the time t c is the cut point. Variation D (t) is adjacent Here temporally two images I t, is calculated as the sum of the absolute value of the difference between the luminance values in corresponding pixels of I t-1. Therefore, time t (≠
The amount of change D (t) in t c ) is obtained by multiplying the area of the displaced portion of the quadrangle in FIG. 7B by the difference 255 between the brightness values of black and white, so that D (t) = 255 * It becomes 2 * Δx * H. At the cut point t = t c , since the difference between the brightness values of white and gray is 255-127 = 128, D (t c )
= 128 * W * H. In the graphs of FIGS. 8A and 8B, the horizontal axis represents time t and the vertical axis represents the amount of change D (t).
When the moving speed Δx of the quadrangle is small, as shown in FIG. 8A, D (t) is sufficiently smaller than D (t c ), so that the cut point can be correctly detected. However, the moving speed Δ
As x increases, as shown in FIG. 8B, D (t) rises overall, so that D (t c ) and D (t c )
It becomes difficult to distinguish (t). For this reason, in the conventional method, the place where the movement of the subject is strong may be mistaken for the cut point. The same situation occurs due to camera shake, which is one of the causes of the deterioration in the performance of cut point detection.

【0008】また、D(t)をそのまま閾値処理するの
ではなく、各種時間フィルタをかけることによって検出
誤りを防ぐ従来の方法では、被写体の動きの連続性を仮
定しているので被写体が突然、非連続的な動きをした場
合にやはり検出誤りを生じていた。
Further, in the conventional method of preventing detection error by applying various time filters instead of directly thresholding D (t), the continuity of the motion of the subject is assumed, so that the subject suddenly appears. In the case of discontinuous movement, a detection error still occurred.

【0009】そこで本発明の目的は、上記従来の技術の
問題点を解決するために、被写体の動きや、カットの手
ぶれによって生じる検出誤りを軽減することのできる検
出性能の良いカット点検出方法を提供することにある。
SUMMARY OF THE INVENTION Therefore, an object of the present invention is to provide a cut point detection method with good detection performance capable of reducing detection errors caused by movement of an object and camera shake of a cut in order to solve the problems of the above-mentioned conventional techniques. To provide.

【0010】[0010]

【課題を解決するための手段】上記の目的を達成するた
め、本発明では、映像を構成する画像データ列からカッ
ト点を検出する映像カット点検出方法において、画像平
面座標x,yと時間tを変数とみなして画像データ列か
ら勾配ベクトルを計算する手順と、該勾配ベクトルのう
ちt軸に平行に近いものの数が多い時刻をカット点とし
て出力する手順と、を有することを特徴とする。
To achieve the above object, in the present invention, in a video cut point detection method for detecting a cut point from an image data string forming a video, image plane coordinates x, y and time t. Is regarded as a variable and a gradient vector is calculated from the image data sequence, and a step of outputting a time point at which a large number of gradient vectors close to the t-axis are output as a cut point.

【0011】また、上記の方法においては、画像データ
列から勾配ベクトルを計算する手順を、画像データ列に
対して平滑化フィルタを作用し、平滑化された時空間画
像の勾配ベクトルを計算するものとするのが、ノイズの
影響を排除して検出性能を安定にする点で好適である。
Further, in the above method, the procedure of calculating a gradient vector from an image data string is performed by applying a smoothing filter to the image data string to calculate the gradient vector of the smoothed spatiotemporal image. Is preferable in that the influence of noise is eliminated and the detection performance is stabilized.

【0012】さらに、上記の方法においては、平滑化フ
ィルタをディジタルフィルタで構成するのが、計算量を
低減して検出を容易にする点で好適である。
Further, in the above method, it is preferable to configure the smoothing filter with a digital filter in order to reduce the amount of calculation and facilitate detection.

【0013】[0013]

【作用】映像を構成する画像データ列を画像平面座標
x,yと時間tの3次元配列として構成した時空間画像
の勾配ベクトルは、輝度変化を例にするとカット点のと
ころでは、x,y方向には小さくt軸方向には絵柄が急
に変わるので大きいためt軸に対し平行に近くなる。こ
れに対して、カット点間のショット内では、勾配ベクト
ルとt軸がなす各度は被写体の動きによって決まる(被
写体が静止していれば、t軸に垂直になる)が、いずれ
にしても、勾配ベクトルとt軸が平行になることは稀で
ある。本発明の映像カット点検出方法は、このように勾
配ベクトがカット点において平行に近くなる性質を利用
して、各時刻の勾配ベクトルを逐次計算し、そのうちt
軸に平行に近い勾配ベクトルが多い時刻をカット点とし
て検出することにより、カット点の検出性能を向上させ
る。
The gradient vector of a spatiotemporal image, in which the image data string forming the image is formed as a three-dimensional array of image plane coordinates x, y and time t, is x, y at the cut point in the case of a luminance change. Since the pattern is small in the direction and suddenly changes in the t-axis direction, it is close to parallel to the t-axis because it is large. On the other hand, in the shot between the cut points, each degree formed by the gradient vector and the t-axis is determined by the movement of the subject (if the subject is stationary, it is perpendicular to the t-axis), but in any case, , The gradient vector and the t-axis are rarely parallel. The image cut point detection method of the present invention utilizes the property that the gradient vector becomes nearly parallel at the cut point in this way, and sequentially calculates the gradient vector at each time, and t
The detection performance of the cut point is improved by detecting the time when there are many gradient vectors near the axis as the cut point.

【0014】[0014]

【実施例】以下、本発明の一実施例を、図面に基づいて
詳細に説明する。
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS An embodiment of the present invention will be described in detail below with reference to the drawings.

【0015】まず、本発明の原理を図6を用いて説明す
る。本発明では、映像を構成する画像データ列からカッ
ト点を検出する映像カット点検出方法において、画像平
面座標x,yと時間tを変数とみなして3次元の配列を
構成する。このように画像データ列を3次元ボリューム
と考えたものを時空間画像と呼ぶ。この時空間画像の画
像データ列から勾配ベクトル(ΔxI,ΔyI,ΔtI)
を時刻tについて逐次計算し、その勾配ベクトルのうち
t軸に平行に近いものの数が多い時刻tをカット点とし
て出力する。本発明は、カット点において、勾配ベクト
ルがt軸に平行に近くなるという性質を利用している。
First, the principle of the present invention will be described with reference to FIG. According to the present invention, in the video cut point detection method of detecting the cut points from the image data sequence forming the video, the image plane coordinates x, y and the time t are regarded as variables to form a three-dimensional array. Such an image data string considered as a three-dimensional volume is called a spatiotemporal image. Gradient vector (Δ x I, Δ y I, Δ t I) from the image data sequence of this spatiotemporal image
Is sequentially calculated at time t, and the time t in which the number of the gradient vectors close to the t-axis is large is output as a cut point. The present invention utilizes the property that the gradient vector becomes close to parallel to the t-axis at the cut point.

【0016】なお、本発明は、時空間画像の勾配ベクト
ル(ΔxI,ΔyI,ΔtI)のうちΔtIの大きさだけで
なく、空間方向の差分ΔxI,ΔyIを考慮するという点
で、従来の技術と異なる。ΔtIは、ある画素に着目し
たとき、時間tの前後で輝度あるいは色が変化するかを
定量化したものであり、従来の技術では、時間方向の差
分、すなわち|ΔtI|のみを見ていて、空間方向の差
分ΔxI,ΔyIについては考慮していなかった。
[0016] The present invention is the gradient vector of the space-time image (Δ x I, Δ y I , Δ t I) as well as the magnitude of delta t I of the difference between the spatial direction Δ x I, Δ y It is different from the conventional technique in that I is taken into consideration. Δt I is a quantification of whether the luminance or the color changes before and after the time t when focusing on a certain pixel. In the conventional technique, only the difference in the time direction, that is, | Δt I | not look, the difference Δ x I spatial direction, for Δ y I did not take into account.

【0017】時空間画像の勾配ベクトルがカット点のと
ころでt軸に平行に近くなるという性質を、図6を使っ
て説明する。図6は、時空間画像をxt平面に平行な面
で切断したときの様子を示している。カットで区切られ
た連続映像区間、すなわちショットの中では、被写体の
動きが帯状の流れとして現れてくる。カット点では絵柄
が急に変わるので、t軸に垂直なエッジが現れる。勾配
ベクトル(図中では矢印で表示)はエッジとほぼ垂直に
交わるので、カット点では72に示すように、t軸にほ
ぼ平行になっている(すべてが厳密に平行になるわけで
はない)。カット点での画素70を例にとって説明する
と、画素70の近傍でx方向の輝度変化は小さいので|
ΔxI|は小さくなる。t方向の輝度変化は大きいので
|ΔtI|は大きくなり、画素70における勾配ベクト
ルはt軸にほぼ平行になるというわけである。ショット
内では、勾配ベクトルとt軸がなす各度は被写体の動き
によって決まる(被写体が静止していれば、t軸に垂直
になる)が、いずれにしても、71に示すように勾配ベ
クトルとt軸が平行になることは稀である。本発明で
は、勾配ベクトルのうちt軸に平行に近いものが多い時
刻をカット点として検出するので、検出性能が向上し、
被写体の動きや手ぶれがあってもそれをカット点と見誤
ることが少なくなる。
The property that the gradient vector of the spatiotemporal image becomes nearly parallel to the t-axis at the cut point will be described with reference to FIG. FIG. 6 shows a state in which the spatiotemporal image is cut along a plane parallel to the xt plane. In a continuous video section divided by cuts, that is, in a shot, the movement of the subject appears as a band-shaped flow. Since the pattern suddenly changes at the cut point, an edge perpendicular to the t-axis appears. Since the gradient vector (indicated by an arrow in the drawing) intersects the edge almost perpendicularly, it is almost parallel to the t-axis (not all are exactly parallel) at the cut point, as shown at 72. Taking the pixel 70 at the cut point as an example, the brightness change in the x direction near the pixel 70 is small.
Δ x I | becomes smaller. Since the change in luminance in the t direction is large, | Δt I | becomes large, and the gradient vector in the pixel 70 is almost parallel to the t axis. Within the shot, the degrees formed by the gradient vector and the t-axis are determined by the movement of the subject (if the subject is stationary, the angle is perpendicular to the t-axis), but in any case, as shown by 71, The t-axes are rarely parallel. According to the present invention, since a time point that is mostly parallel to the t-axis among the gradient vectors is detected as a cut point, the detection performance is improved,
Even if there is movement or camera shake of the subject, it is less likely to mistake it as a cut point.

【0018】次に、このような原理に基づく本発明の一
実施例を説明する。図1は、本実施例を示す処理フロー
図である。
Next, an embodiment of the present invention based on such a principle will be described. FIG. 1 is a process flow chart showing the present embodiment.

【0019】ここでは、画像データ列はディジタル化さ
れており、x,y座標と時刻tについて輝度値I(x,
y,t)を定めるような3次元の配列(時空間画像)と
考えることにする。なお、ここでは、一実施例として輝
度値のみを考慮するが、色情報例えばRGB値を考える
こともできる。
Here, the image data string is digitized, and the brightness value I (x,
Let us consider it as a three-dimensional array (spatiotemporal image) that defines y, t). Although only the luminance value is considered here as an example, color information such as RGB values may be considered.

【0020】ステップ101では、時空間画像に対して
平滑化作用を持つフィルタを作用することによって平滑
化時空間画像を構成する。平滑化することによって、安
定に勾配ベクトルを求めることができるようになる(平
滑化しないとノイズによって勾配を安定に求めることが
できない)。平滑化作用を持つフィルタとしては、例え
ばガウシアンフィルタ
In step 101, a smoothed spatiotemporal image is constructed by operating a filter having a smoothing action on the spatiotemporal image. By smoothing, it becomes possible to stably obtain the gradient vector (without smoothing, the gradient cannot be stably obtained due to noise). As a filter having a smoothing action, for example, a Gaussian filter

【0021】[0021]

【数1】 [Equation 1]

【0022】を利用することができる。ガウシアンフィ
ルタGと時空間画像Iの畳み込み(convoluti
on)をとったものが、平滑化時空間画像I′(=G*
I)となる。σx,σy,σtはそれぞれx,y,t方向
に関する平滑化のスケールである。σが大きいほどフィ
ルタ後の時空間画像は滑らかになり、ノイズが除去され
る反面、細部の特徴が失われる。計算量を削減するため
に、次のようなディジタルフィルタを利用することもで
きる。
Can be used. Convolution of the Gaussian filter G and the spatiotemporal image I (convoluti)
on) is the smoothed spatiotemporal image I '(= G *
I). σ x , σ y , and σ t are smoothing scales in the x, y, and t directions, respectively. The larger σ is, the smoother the spatiotemporal image after filtering is, and the noise is removed, but the fine features are lost. In order to reduce the amount of calculation, the following digital filter can be used.

【0023】[0023]

【数2】 [Equation 2]

【0024】ただし、W,H,Tは平滑化のスケールを
表す。
However, W, H, and T represent smoothing scales.

【0025】ステップ102では、時刻tについて、平
滑化された時空間画像の各画素についてその勾配(Δx
I,ΔyI,ΔtI)を算出する。ガウシアンフィルタの
場合、
In step 102, at time t, the gradient (Δ x
I, Δ y I, Δ t I) are calculated. For a Gaussian filter,

【0026】[0026]

【数3】 (Equation 3)

【0027】が成り立つので、平滑化フィルタを畳み込
んでから微分するかわりに、微分フィルタ∂G/∂xを
畳み込んでもよい。また、その他の微分作用をもつフィ
ルタ、例えば
Since the following holds, instead of convolving the smoothing filter and then differentiating, the differential filter ∂G / ∂x may be convoluted. Also, other filters with differentiating effects, such as

【0028】[0028]

【数4】 [Equation 4]

【0029】を使ってもよい。May be used.

【0030】ステップ103では、勾配ベクトルが |ΔxI|<θ1and|ΔyI|<θ1 (1) の条件を満たす領域(領域Aと呼ぶ)の画素数をカウン
トし、それをcount1とする。|ΔxI|は近傍に
縦方向のエッジが存在すると大きな値をとり、|Δy
|は近傍に横方向のエッジが存在すると大きな値をとる
ので、(1)式を満たす領域Aの近傍にエッジがないと
いうことができる。図2(a)の画像に対して、領域A
を求めたものを図2(b)に斜線で示す。定性的には被
写体のエッジ近傍を除いた領域が領域Aとなる。(1)
式の変わりに |ΔxI|+|ΔyI|<θ あるいは、 √(|ΔI|2+|ΔyI|2)<θ1 などの式を使ってもよい。
[0030] At step 103, the gradient vector is | Δ x I | <θ 1 and | Δ y I | counts the number of pixels <theta satisfies region 1 (1) (referred to as region A), it count1. │Δ x I │ takes a large value when there are vertical edges in the vicinity, and │Δ y I
Since | has a large value when there is a horizontal edge in the vicinity, it can be said that there is no edge in the vicinity of the region A that satisfies the expression (1). In the image of FIG. 2A, the area A
What was calculated is shown by the diagonal lines in FIG. Qualitatively, the area A is the area excluding the vicinity of the edge of the subject. (1)
Instead of equation | Δ x I | + | Δ y I | <θ 1 or, √ (| Δ x I | 2 + | Δ y I | 2) < may use an expression such theta 1.

【0031】ステップ104では、勾配ベクトルが
(1)の条件と |ΔtI|>θ2 (2) の条件をともに満たす領域(領域Bと呼ぶ)の画素数を
カウントし、それをcount2とする。条件(2)の
みを満たす領域を領域Cと呼ぶことにすると、領域Cと
領域Aの共通部分が領域Bになる。従来、領域Cの面積
を全画素数で割ったものをdiff’とし、diff’
がある閾値を越えたときをカット点とみなすという方法
があった。この方法では、図4に示すように、被写体の
動きが激しいとカット点でなくてもdiff’が増加す
るため、被写体の動きや手ぶれによって検出誤りが多い
という問題点があった。被写体の動きや手ぶれがある
と、被写体のエッジ近傍でフレーム間差分が大きくなる
(|ΔtI|が大きくなる)ためである。そこで、本発
明の一実施例では、領域Cの面積(diff’)ではな
く領域Bの面積をカウントするようにする。条件(1)
と条件(2)が成り立つということは、幾何学的に解釈
すると、勾配ベクトルがt軸に平行に近いということで
ある。例として図3(a)の画面から図3(b)の画像
にシーンが切り替わる場合を考える。図3(c)に領域
Aを斜線で示し領域Bを黒色で示す。カット点では領域
Aのうち領域Bが大きな部分を占めるという性質を利用
してカット点を検出する。
In step 104, the number of pixels in a region (referred to as region B) in which the gradient vector satisfies both the condition (1) and the condition | Δ t I |> θ 2 (2) is counted as count2. To do. When the area that satisfies only the condition (2) is called area C, the common part of area C and area A becomes area B. Conventionally, the area of the region C divided by the total number of pixels is defined as diff ', and diff'
There was a method to consider when a certain threshold was exceeded as a cut point. In this method, as shown in FIG. 4, when the movement of the subject is strong, the diff 'increases even at the cut point, so that there is a problem that many detection errors occur due to the movement of the subject or camera shake. This is because if there is motion or camera shake of the subject, the inter-frame difference increases (| Δ t I | increases) near the edge of the subject. Therefore, in one embodiment of the present invention, not the area (diff ') of the area C but the area of the area B is counted. Condition (1)
The condition (2) is satisfied that the gradient vector is parallel to the t-axis when viewed geometrically. As an example, consider the case where the scene switches from the screen of FIG. 3A to the image of FIG. In FIG. 3 (c), the area A is shaded and the area B is black. At the cut point, the cut point is detected by utilizing the property that the area B occupies a large portion of the area A.

【0032】ステップ105では、diff=coun
t2/count1を計算する。すなわち、領域Bの面
積と領域Aの面積の比をdiffとする。0≦diff
≦1でありカット点のところでdiffは1に近い値を
とるので、ステップ106でdiff>θ3が成り立つ
ときカット点ありと判定し、そうでないときカット点な
しと判定する。図5にdiffのグラフの典型例を示
す。被写体の動きや手ぶれがあっても、diffは図4
とは異なり上昇しないので、閾値処理によってカット点
を安定に検出できる。
In step 105, diff = count
Calculate t2 / count1. That is, the ratio of the area of the region B to the area of the region A is defined as diff. 0 ≦ diff
Since ≦ 1 and diff takes a value close to 1 at the cut point, it is determined in step 106 that there is a cut point when diff> θ 3 is true, and otherwise it is determined that there is no cut point. FIG. 5 shows a typical example of the diff graph. Even if there is movement of the subject or camera shake, diff is
Unlike the above, since it does not rise, the cut point can be stably detected by the threshold processing.

【0033】ステップ106ではtを進めてステップ1
02に戻る。
At step 106, t is advanced to step 1
Return to 02.

【0034】[0034]

【発明の効果】以上説明したように、本発明の映像カッ
ト点検出方法によれば、映像を画像平面座標と時間を変
数とする時空間画像として考え、その時空間画像の勾配
ベクトルを計算して時間方向の差分だけでなく空間方向
の差分も考慮してカット点を検出するようにしたので、
カット点の検出性能が向上し、時空間画像の被写体の動
きや手ぶれによって生じる検出誤りを軽減できる効果が
ある。
As described above, according to the video cut point detection method of the present invention, a video is considered as a spatiotemporal image having image plane coordinates and time as variables, and the gradient vector of the spatiotemporal image is calculated. Since the cut point is detected by considering not only the difference in the time direction but also the difference in the spatial direction,
There is an effect that the detection performance of the cut point is improved and the detection error caused by the movement of the subject and the camera shake in the spatiotemporal image can be reduced.

【0035】また、上記において、時空間画像に平滑化
フィルタを作用させるようにした場合には、特に、ノイ
ズの影響を排除してカット点の検出を安定にする。
Further, in the above, when the smoothing filter is made to act on the spatiotemporal image, in particular, the influence of noise is eliminated and the detection of the cut point is stabilized.

【0036】さらに、上記において、平滑化フィルタを
ディジタルフィルタで構成した場合には、特に、計算量
を削減できる。
Further, in the above, when the smoothing filter is composed of a digital filter, the amount of calculation can be reduced especially.

【図面の簡単な説明】[Brief description of drawings]

【図1】本発明の一実施例の処理フロー図。FIG. 1 is a process flow chart of an embodiment of the present invention.

【図2】(a),(b)は、上記実施例における条件
(1)を満たす領域(領域A)の説明図。
2A and 2B are explanatory views of a region (region A) that satisfies the condition (1) in the above-described embodiment.

【図3】(a),(b),(c)は、上記実施例におけ
る条件(1)を満たす領域(領域A)と、条件(1),
(2)を満たす領域(領域B)の説明図。
3 (a), (b), and (c) are a region (region A) satisfying the condition (1) in the above-described embodiment and the condition (1),
Explanatory drawing of the area | region (area | region B) which satisfy | fills (2).

【図4】上記実施例における条件(2)のみを満たす領
域(領域C)の面積(diff’)の典型的な時間変化
を示す図。
FIG. 4 is a diagram showing a typical time change of an area (diff ′) of a region (region C) satisfying only condition (2) in the above-mentioned embodiment.

【図5】上記領域Bの面積(diff)の典型的な時間
変化を示す図。
FIG. 5 is a diagram showing a typical time change of the area (diff) of the region B.

【図6】本発明の原理を説明するための時空間画像の切
断面と勾配ベクトルの説明図。
FIG. 6 is an explanatory diagram of a cutting plane and a gradient vector of a spatiotemporal image for explaining the principle of the present invention.

【図7】従来方法の問題点を説明するための図であっ
て、(a)は入力画像列、(b)は変化量D(t)の算
出方法の説明図。
7A and 7B are diagrams for explaining the problems of the conventional method, in which FIG. 7A is an input image sequence, and FIG. 7B is an explanatory diagram of a method of calculating a variation D (t).

【図8】従来方法の問題点を説明するための図であっ
て、(a)は移動量Δが小さい時の変化量D(t)のグ
ラフ、(b)は移動量Δが大きい時の変化量D(t)の
グラフ。
8A and 8B are diagrams for explaining problems of the conventional method, in which FIG. 8A is a graph of a change amount D (t) when the movement amount Δ is small, and FIG. 8B is a graph when the movement amount Δ is large. The graph of change amount D (t).

【符号の説明】[Explanation of symbols]

101〜107…ステップ 101-107 ... Step

Claims (3)

【特許請求の範囲】[Claims] 【請求項1】 映像を構成する画像データ列からカット
点を検出する映像カット点検出方法において、 画像平面座標x,yと時間tを変数とみなして画像デー
タ列から勾配ベクトルを計算する手順と、 該勾配ベクトルのうちt軸に平行に近いものの数が多い
時刻をカット点として出力する手順と、を有することを
特徴とする映像カット点検出方法。
1. A video cut point detection method for detecting a cut point from an image data string forming a video, wherein a gradient vector is calculated from the image data string by regarding the image plane coordinates x, y and time t as variables. A step of outputting, as a cut point, a time at which a large number of the gradient vectors that are close to the t-axis are output, as a cut point detection method.
【請求項2】 画像データ列から勾配ベクトルを計算す
る手順が、 画像データ列に対して平滑化フィルタを作用し、平滑化
された時空間画像の勾配ベクトルを計算するものである
ことを特徴とする請求項1記載の映像カット点検出方
法。
2. A procedure of calculating a gradient vector from an image data string is a step of operating a smoothing filter on the image data string to calculate a gradient vector of a smoothed spatiotemporal image. The image cut point detection method according to claim 1.
【請求項3】 平滑化フィルタがディジタルフィルタで
構成されていることを特徴とする請求項2記載の映像カ
ット点検出方法。
3. The video cut point detection method according to claim 2, wherein the smoothing filter is a digital filter.
JP18202494A 1994-08-03 1994-08-03 Detection of video cutting point Pending JPH0846864A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP18202494A JPH0846864A (en) 1994-08-03 1994-08-03 Detection of video cutting point

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP18202494A JPH0846864A (en) 1994-08-03 1994-08-03 Detection of video cutting point

Publications (1)

Publication Number Publication Date
JPH0846864A true JPH0846864A (en) 1996-02-16

Family

ID=16111012

Family Applications (1)

Application Number Title Priority Date Filing Date
JP18202494A Pending JPH0846864A (en) 1994-08-03 1994-08-03 Detection of video cutting point

Country Status (1)

Country Link
JP (1) JPH0846864A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100525678B1 (en) * 2000-06-06 2005-11-03 Method and system for compressing motion image information

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR100525678B1 (en) * 2000-06-06 2005-11-03 Method and system for compressing motion image information

Similar Documents

Publication Publication Date Title
US7221776B2 (en) Video stabilizer
US7406208B2 (en) Edge enhancement process and system
EP1112661B1 (en) Tracking semantic objects in vector image sequences
US7095786B1 (en) Object tracking using adaptive block-size matching along object boundary and frame-skipping when object motion is low
US8416993B2 (en) Object boundary accurate motion detection using hierarchical block splitting and motion segmentation
KR20180084085A (en) METHOD, APPARATUS AND ELECTRONIC DEVICE
US20160171664A1 (en) Image processing apparatus and image processing method
US7519230B2 (en) Background motion vector detection
US20110115790A1 (en) Apparatus and method for converting 2d image signals into 3d image signals
JP2008518331A (en) Understanding video content through real-time video motion analysis
US20150022677A1 (en) System and method for efficient post-processing video stabilization with camera path linearization
US20130271666A1 (en) Dominant motion estimation for image sequence processing
CA2279797A1 (en) A method for temporal interpolation of an image sequence using object-based image analysis
JPH08202879A (en) Method for change of continuous video images belonging to sequence of mutually interrelated images as well as apparatus and method for replacement of expression of targetdiscriminated by set of object points by matched expression of predetermined and stored pattern of same geometrical shape in continuous tv frames of same sequence
US7949205B2 (en) Image processing unit with fall-back
US20070092111A1 (en) Motion vector field re-timing
US20100208140A1 (en) Image processing apparatus, image processing method and storage medium storing image processing program
JP4913801B2 (en) Shielding object image identification apparatus and method
JP2009533887A (en) Motion vector field correction apparatus and method
JP2005509230A (en) Occlusion detector and method for detecting an occlusion region
JP3404803B2 (en) Moving image recording medium, still image extracting device, moving image recording device, and still image automatic extracting method
Gangal et al. An improved motion-compensated restoration method for damaged color motion picture films
JP5059855B2 (en) Global motion estimation method
JPH0846864A (en) Detection of video cutting point
Lagendijk et al. Video enhancement and restoration