US20110274367A1 - Method and Apparatus for Processing Video Data - Google Patents

Method and Apparatus for Processing Video Data Download PDF

Info

Publication number
US20110274367A1
US20110274367A1 US13/186,552 US201113186552A US2011274367A1 US 20110274367 A1 US20110274367 A1 US 20110274367A1 US 201113186552 A US201113186552 A US 201113186552A US 2011274367 A1 US2011274367 A1 US 2011274367A1
Authority
US
United States
Prior art keywords
frame
target
pixels
target pixels
frames
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US13/186,552
Inventor
Wei Hu
Qiao Huang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to US13/186,552 priority Critical patent/US20110274367A1/en
Publication of US20110274367A1 publication Critical patent/US20110274367A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion

Definitions

  • Some methods of object detection directly detect each occurrence of an object based on the pixel-wise or block-wise color difference between consecutive frames. Other procedures entail first detecting several occurrences (samples) of the object. A template (usually a frame) is then learned for the object by extracting common characteristics from these acquired samples. For example, some methods use motion feature technique and apply dynamic programming to match the object movement. The extracted object template is then used to scan the whole video to find all occurrences of the object. Each occurrence of the object should contain the same foreground pixels (i.e., object pixels) and different background pixels (i.e., non-object pixels). Thus, foreground pixels may be extracted to represent the object itself. Background pixels may not describe the object and may introduce noise.
  • Still other methods may be used for extracting the foreground pixels of an object.
  • the object may be highlighted and located at the center of the frame. Thus, only the center and bright pixels are extracted.
  • Other methods may use motion information, assuming pixels that move faster than others are foreground pixels. However, pixels around the margin of the object may be returned. These marginal pixels do not provide accurate information for the object.
  • FIG. 1 is a representation of video data in one embodiment of the invention.
  • FIG. 2 is a flow diagram of a method of processing video data in one embodiment of the invention.
  • FIG. 3 includes a system for use with embodiments of the invention.
  • FIG. 1 shows multiple occurrences of the same object (e.g., logo) 101 displayed in three video segments.
  • these segments may include “instant replays” in of a sporting event.
  • the first replay includes a first sequence of frames 110 , 120 , 130
  • the second replay includes a second sequence of frames 140 , 150 , 160
  • the third replay includes a third sequence of frames 170 , 180 , 190 .
  • Each replay is preceded by logo 101 floating across the video frames.
  • accurate localization and retrieval of logo 101 i.e., object 101
  • logos 101 may be good candidates for “markers” because each logo 101 occurrence contains consecutive image sequences, the duration of the segment is normally less than one second, the logo may occur many times in a video, the logo may keep the same shape, color, duration and motion model each time, and the background may change greatly for each occurrence of the logo.
  • a background FIG. 102 may be included in the first sequence of frames.
  • a different background FIG. 103 may be included in the second sequence of frames.
  • yet another background FIG. 104 may be included.
  • FIG. 2 shows a flow diagram 200 of a method of processing video data in one embodiment of the invention.
  • M sequences are aligned.
  • the first, second, and third sequences of FIG. 1 are aligned.
  • Various methods for aligning sequences are known to those of ordinary skill in the art.
  • alignment may be obtained with global programming and the Needleman-Wunsch-Sellers algorithm.
  • a representative frame for an object is found. More specifically, in one embodiment of the invention, for each pair of the example sequences i and j (i.e., any two of the three sequences of FIG. 1 ), compute the match score S seq (i, j). In one embodiment of the invention, use of motion activity or motion direction is used. However, other methods are known to those of ordinary skill in the art. In one embodiment of the invention, the sum score S seq for each sequence is calculated. For example, such sums may be based on each sequence compared to other aligned sequences. Then, identify the two sequences, when there are more than two sequences, with the highest and second highest summed score. For example, the first and second sequences of FIG. 1 may be selected.
  • This frame may be labeled the k-frame.
  • frame 120 can be identified as a k-frame in FIG. 1 .
  • frames 150 , 180 may be selected as k-frames. Together, the k-frames form a “k-frame set.”
  • C1 pixels (candidate object pixels) are obtained.
  • the color difference score for each pair of frames is computed.
  • sum up all N C M 2 values for each pixel in the frame.
  • N 3 with three sequences.
  • N 6 if there are four sequences.
  • the cumulated color difference in these N pairs can be computed as
  • R, G, and B are the colors of the pixel.
  • the object 101 may be substantially unchanged while the background 102 , 103 , 104 changes greatly.
  • the smaller the accumulated value D of a pixel the more possible this pixel is an O-pixel (object pixel) that describes the object 101 foreground.
  • all the pixels' cumulated value Ds are sorted in ascending order.
  • the top 30%-50% pixels are the candidate O-pixels, called “C1-pixels.”
  • the N-frame is selected. From FIG. 1 , we can see some of the background pixels 105 may also be considered C1pixels due to their static nature. These background pixels 105 may be, for example, a TV channel logo for a broadcast corporation that does not change frame-to-frame. Those background pixels 105 , shown in every frame including both object frames and non-object frames, are noise and should not be the O-pixels. Thus, in one embodiment of the invention, the background pixels 105 may be eliminated. A reference frame may be used to help remove most of the non O-pixels (e.g., 105 ) include among the C1-pixels. To do so, one may randomly select a frame preceding the k-frame within, for example, 50-100 frame distances.
  • C2 pixels (candidate object pixels) are obtained in one embodiment of the invention. To do so, compute the color difference score between the N-frame and each frame of the k-frame set. Then, sum up all M values for each pixel. During a short time span, the background usually changes little. Thus, most of the changes come from the occurrence of the object. Therefore, pixels with bigger difference scores have higher possibility to be the O-pixels. Sort these summed scores for each pixel in descending order. For example, in one embodiment of the invention the top 30%-50% pixels are considered as candidate O-pixels, called “C2 pixels.”
  • the O-pixels are obtained from C1 pixels and C2 pixels in one embodiment of the invention. If a pixel is in both C1 and C2 lists, it is identified as an O-pixel. In some embodiments of the invention, the O-pixels contain 20% pixels of the whole frame image. The method ends in block 270 .
  • the method and related apparatus may be used to gain a pixel-wise object representation.
  • the method may be used to extract object pixels from multiple occurrences in, for example, video data.
  • the proposed method may concern logo detection in sports videos.
  • the idea can be easily generalized to other video types or object types where the object occurs multiple times with little change. This approach can benefit video-content analysis and object representation and retrieval applications.
  • the methods may remove the non-object pixels and get a high-accuracy pixel-wise representation of the object of interest.
  • embodiments of the invention may be implemented in a computer program. As such, these embodiments may be stored on a storage medium having stored thereon instructions which can be used to program a computer system to perform the embodiments.
  • the storage medium may include, but is not limited to, any type of disk including floppy disks, optical disks, compact disk read-only memories (CD-ROMs), compact disk rewritables (CD-RWs), and magneto-optical disks, semiconductor devices such as read-only memories (ROMs), random access memories (RAMs), erasable programmable read-only memories (EPROMs), electrically erasable programmable read-only memories (EEPROMs), flash memories, magnetic or optical cards, or any type of media suitable for storing electronic instructions.
  • ROMs read-only memories
  • RAMs random access memories
  • EPROMs erasable programmable read-only memories
  • EEPROMs electrically erasable programmable read-only memories
  • flash memories magnetic or optical cards, or any type of media suitable
  • embodiments may be implemented as software modules executed by a programmable control device.
  • a programmable control device may be a computer processor or a custom designed state machine.
  • Custom designed state machines may be embodied in a hardware device such as a printed circuit board having discrete logic, integrated circuits, or specially designed application specific integrated circuits (ASICs).
  • FIG. 3 is a block diagram of a representative data processing system, namely computer system 300 with which embodiments of the invention may be used.
  • computer system 300 includes a processor 310 , which may include a general-purpose or special-purpose processor such as a microprocessor, microcontroller, ASIC, a programmable gate array (PGA), and the like.
  • the processor 310 may be coupled over a host bus 315 to a memory controller 330 in one embodiment, which may be coupled to a system memory 320 via a memory bus 325 .
  • the memory controller 330 may also be coupled over a bus 333 to a display 337 .
  • the memory controller 330 may also be coupled (via a link 338 ) to an input/output (I/O) controller 339 that is coupled to an input/output (I/O) expansion bus 342 .
  • the I/O expansion bus 342 may be coupled to an I/O controller 346 that controls access to one or more I/O devices. As shown in FIG. 3 , these devices may include, in one embodiment of the invention, storage devices such as a disk drive 350 and input devices, such as keyboard 352 and mouse 354 .
  • the I/O controller 339 may also be coupled to, for example, a hard disk drive or server 356 .
  • the I/O controller 339 may be coupled to a PCI bus 344 and a network interface card (NIC) 362 .
  • NIC network interface card

Abstract

In one embodiment of the invention, the invention my concern a method including receiving digital images, each image including object pixels and non-object pixels; and filtering the non-object pixels from the object pixels. Other embodiments are described and claimed.

Description

  • This application is a continuation of U.S. patent application Ser. No. 11/648,294, filed Dec. 29, 2006, the content of which is hereby incorporated by reference.
  • BACKGROUND
  • Learning to recognize key objects in video data and then extracting the pixels that compose those objects is a component in content based video processing. Some methods of object detection directly detect each occurrence of an object based on the pixel-wise or block-wise color difference between consecutive frames. Other procedures entail first detecting several occurrences (samples) of the object. A template (usually a frame) is then learned for the object by extracting common characteristics from these acquired samples. For example, some methods use motion feature technique and apply dynamic programming to match the object movement. The extracted object template is then used to scan the whole video to find all occurrences of the object. Each occurrence of the object should contain the same foreground pixels (i.e., object pixels) and different background pixels (i.e., non-object pixels). Thus, foreground pixels may be extracted to represent the object itself. Background pixels may not describe the object and may introduce noise.
  • Still other methods may be used for extracting the foreground pixels of an object. The object may be highlighted and located at the center of the frame. Thus, only the center and bright pixels are extracted. Other methods may use motion information, assuming pixels that move faster than others are foreground pixels. However, pixels around the margin of the object may be returned. These marginal pixels do not provide accurate information for the object.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings, incorporated in and constituting a part of this specification, illustrate one or more implementations consistent with the principles of the invention and, together with the description of the invention, explain such implementations. The drawings are not necessarily to scale, the emphasis instead being placed upon illustrating the principles of the invention. In the drawings:
  • FIG. 1 is a representation of video data in one embodiment of the invention.
  • FIG. 2 is a flow diagram of a method of processing video data in one embodiment of the invention.
  • FIG. 3 includes a system for use with embodiments of the invention.
  • DETAILED DESCRIPTION
  • The following description refers to the accompanying drawings. Among the various drawings the same reference numbers may be used to identify the same or similar elements. While the following description provides a thorough understanding of the various aspects of the claimed invention by setting forth specific details such as particular structures, architectures, interfaces, techniques, etc., such details are provided for purposes of explanation and should not be viewed as limiting. Moreover, those of skill in the art will, in light of the present disclosure, appreciate that various aspects of the invention claimed may be practiced in other examples or implementations that depart from these specific details. At certain junctures in the following disclosure descriptions of well known devices, circuits, and methods have been omitted to avoid clouding the description of the present invention with unnecessary detail.
  • FIG. 1 shows multiple occurrences of the same object (e.g., logo) 101 displayed in three video segments. In one embodiment of the invention, these segments may include “instant replays” in of a sporting event. The first replay includes a first sequence of frames 110, 120, 130, the second replay includes a second sequence of frames 140, 150, 160, and the third replay includes a third sequence of frames 170, 180, 190. Each replay is preceded by logo 101 floating across the video frames. Thus, accurate localization and retrieval of logo 101 (i.e., object 101) may be a reliable method to find instant replays or other portions of video data in one embodiment of the invention. These logos 101 may be good candidates for “markers” because each logo 101 occurrence contains consecutive image sequences, the duration of the segment is normally less than one second, the logo may occur many times in a video, the logo may keep the same shape, color, duration and motion model each time, and the background may change greatly for each occurrence of the logo. For example, in the first sequence of frames a background FIG. 102 may be included. In the second sequence of frames a different background FIG. 103 may be included. In the third sequence of frames yet another background FIG. 104 may be included.
  • FIG. 2 shows a flow diagram 200 of a method of processing video data in one embodiment of the invention. In block 210, M sequences are aligned. For example, the first, second, and third sequences of FIG. 1 are aligned. Various methods for aligning sequences are known to those of ordinary skill in the art. For example, in one embodiment of the invention, alignment may be obtained with global programming and the Needleman-Wunsch-Sellers algorithm.
  • In block 220, a representative frame for an object is found. More specifically, in one embodiment of the invention, for each pair of the example sequences i and j (i.e., any two of the three sequences of FIG. 1), compute the match score Sseq(i, j). In one embodiment of the invention, use of motion activity or motion direction is used. However, other methods are known to those of ordinary skill in the art. In one embodiment of the invention, the sum score Sseq for each sequence is calculated. For example, such sums may be based on each sequence compared to other aligned sequences. Then, identify the two sequences, when there are more than two sequences, with the highest and second highest summed score. For example, the first and second sequences of FIG. 1 may be selected. Match these two selected sequences again if the score Sseq was not retained in, for example, memory. Then find a frame that matches best. This frame may be labeled the k-frame. For example, frame 120 can be identified as a k-frame in FIG. 1. Next, find a frame in each sequence that corresponding to the k-frame. For example, frames 150, 180 may be selected as k-frames. Together, the k-frames form a “k-frame set.”
  • In block 230, C1 pixels (candidate object pixels) are obtained. In one embodiment of the invention, for the identified k-frame set, which contains M frames, the color difference score for each pair of frames is computed. Then, sum up all N=CM 2 values for each pixel in the frame. In FIG. 1, N=3 with three sequences. As another example, N=6 if there are four sequences. For a particular pixel, the cumulated color difference in these N pairs can be computed as
  • D = N Δ R + Δ G + Δ B ,
  • where R, G, and B are the colors of the pixel. In video, the object 101 may be substantially unchanged while the background 102, 103, 104 changes greatly. Thus, the smaller the accumulated value D of a pixel, the more possible this pixel is an O-pixel (object pixel) that describes the object 101 foreground. Next, in one embodiment of the invention, all the pixels' cumulated value Ds are sorted in ascending order. In one embodiment of the invention, the top 30%-50% pixels are the candidate O-pixels, called “C1-pixels.”
  • In block 240, the N-frame is selected. From FIG. 1, we can see some of the background pixels 105 may also be considered C1pixels due to their static nature. These background pixels 105 may be, for example, a TV channel logo for a broadcast corporation that does not change frame-to-frame. Those background pixels 105, shown in every frame including both object frames and non-object frames, are noise and should not be the O-pixels. Thus, in one embodiment of the invention, the background pixels 105 may be eliminated. A reference frame may be used to help remove most of the non O-pixels (e.g., 105) include among the C1-pixels. To do so, one may randomly select a frame preceding the k-frame within, for example, 50-100 frame distances. One must make sure the selected frame is not in an object track (i.e., the frame is not a part of an occurrence of the object). Selecting a frame within, for example, 50-100 frame distances, may accomplish this. This selected frame is the “N-frame”.
  • In block 250, C2 pixels (candidate object pixels) are obtained in one embodiment of the invention. To do so, compute the color difference score between the N-frame and each frame of the k-frame set. Then, sum up all M values for each pixel. During a short time span, the background usually changes little. Thus, most of the changes come from the occurrence of the object. Therefore, pixels with bigger difference scores have higher possibility to be the O-pixels. Sort these summed scores for each pixel in descending order. For example, in one embodiment of the invention the top 30%-50% pixels are considered as candidate O-pixels, called “C2 pixels.”
  • In block 260, the O-pixels are obtained from C1 pixels and C2 pixels in one embodiment of the invention. If a pixel is in both C1 and C2 lists, it is identified as an O-pixel. In some embodiments of the invention, the O-pixels contain 20% pixels of the whole frame image. The method ends in block 270.
  • In one embodiment of the invention, the method and related apparatus may be used to gain a pixel-wise object representation. The method may be used to extract object pixels from multiple occurrences in, for example, video data. The proposed method may concern logo detection in sports videos. However, the idea can be easily generalized to other video types or object types where the object occurs multiple times with little change. This approach can benefit video-content analysis and object representation and retrieval applications. The methods may remove the non-object pixels and get a high-accuracy pixel-wise representation of the object of interest.
  • As those of ordinary skill in the art will appreciate, embodiments of the invention may be implemented in a computer program. As such, these embodiments may be stored on a storage medium having stored thereon instructions which can be used to program a computer system to perform the embodiments. The storage medium may include, but is not limited to, any type of disk including floppy disks, optical disks, compact disk read-only memories (CD-ROMs), compact disk rewritables (CD-RWs), and magneto-optical disks, semiconductor devices such as read-only memories (ROMs), random access memories (RAMs), erasable programmable read-only memories (EPROMs), electrically erasable programmable read-only memories (EEPROMs), flash memories, magnetic or optical cards, or any type of media suitable for storing electronic instructions. Similarly, embodiments may be implemented as software modules executed by a programmable control device. A programmable control device may be a computer processor or a custom designed state machine. Custom designed state machines may be embodied in a hardware device such as a printed circuit board having discrete logic, integrated circuits, or specially designed application specific integrated circuits (ASICs).
  • FIG. 3 is a block diagram of a representative data processing system, namely computer system 300 with which embodiments of the invention may be used. In one embodiment, computer system 300 includes a processor 310, which may include a general-purpose or special-purpose processor such as a microprocessor, microcontroller, ASIC, a programmable gate array (PGA), and the like. The processor 310 may be coupled over a host bus 315 to a memory controller 330 in one embodiment, which may be coupled to a system memory 320 via a memory bus 325. The memory controller 330 may also be coupled over a bus 333 to a display 337.
  • The memory controller 330 may also be coupled (via a link 338) to an input/output (I/O) controller 339 that is coupled to an input/output (I/O) expansion bus 342. The I/O expansion bus 342 may be coupled to an I/O controller 346 that controls access to one or more I/O devices. As shown in FIG. 3, these devices may include, in one embodiment of the invention, storage devices such as a disk drive 350 and input devices, such as keyboard 352 and mouse 354. The I/O controller 339 may also be coupled to, for example, a hard disk drive or server 356. Furthermore, the I/O controller 339 may be coupled to a PCI bus 344 and a network interface card (NIC) 362. Although the description makes reference to specific components of the system 300, it is contemplated that numerous modifications and variations of the described and illustrated embodiments may be possible.
  • While the present invention has been described with respect to a limited number of embodiments, those skilled in the art will appreciate numerous modifications and variations therefrom. It is intended that the appended claims cover all such modifications and variations that falls within the true spirit and scope of this present invention.

Claims (20)

1. A method comprising:
filtering non-target pixels of a non-target object from target pixels of a target object;
wherein (a) the target object and the non-target object both change locations in adjacent neighboring frames of a first frame sequence, and (b) the target object is not included in another frame located outside the first frame sequence.
2. The method of claim 1, including filtering the non-target pixels from the target pixels based on the non-target pixels in first and second frames being less similar than the target pixels in the first and second frames.
3. The method of claim 2, wherein the first frame is included in the first frame sequence and the second frame is included in a second frame sequence non-adjacent to the first frame sequence.
4. The method of claim 3, wherein the first and second frame sequences are (a) separated by a frame that does not include the target object, and (b) included in a K-frame set.
5. The method of claim 1, including filtering non-target pixels of a static object, included in one of the adjacent neighboring frames, from the target pixels based on additional non-target pixels in an additional frame being more similar to the non-object pixels of the static object than to the target pixels.
6. The method of claim 5, wherein the additional frame is included in an N-frame and is not included in an object track.
7. The method of claim 1, including searching a digital video for digital images that match the target pixels.
8. An article comprising a non-transitory medium storing instructions that enable a processor-based system to:
filter non-target pixels of a non-target object from target pixels of a target object;
wherein (a) the target object and the non-target object both change locations in adjacent neighboring frames of a first frame sequence, and (b) the target object is not included in another frame located outside the first frame sequence.
9. The article of claim 8, storing instructions that enable the system to filter the non-target pixels from the target pixels based on the non-target pixels in first and second frames being less similar than the target pixels in the first and second frames.
10. The article of claim 9, wherein the first frame is included in the first frame sequence and the second frame is included in a second frame sequence non-adjacent to the first frame sequence.
11. The article of claim 10, wherein the first and second frame sequences are (a) separated by a frame that does not include the target object, and (b) included in a K-frame set.
12. The article of claim 8, storing instructions that enable the system to filter non-target pixels of a static object, included in one of the adjacent neighboring frames, from the target pixels based on additional non-target pixels in an additional frame being more similar to the non-object pixels of the static object than to the target pixels.
13. The article of claim 12, wherein the additional frame is included in an N-frame and is not included in an object track.
14. The article of claim 8, storing instructions that enable the system to search a digital video for digital images that match the target pixels.
15. A system comprising:
a memory;
a processor, coupled to the memory, to filter non-target pixels of a non-target object from target pixels of a target object; wherein (a) the target object and the non-target object both change locations in adjacent neighboring frames of a first frame sequence, and (b) the target object is not included in another frame located outside the first frame sequence.
16. The system of claim 15, wherein the processor is to filter the non-target pixels from the target pixels based on the non-target pixels in first and second frames being less similar than the target pixels in the first and second frames.
17. The system of claim 16, wherein the first frame is included in the first frame sequence and the second frame is included in a second frame sequence non-adjacent to the first frame sequence.
18. The system of claim 17, wherein the first and second frame sequences are (a) separated by a frame that does not include the target object, and (b) included in a K-frame set.
19. The system of claim 15, wherein the processor is to filter non-target pixels of a static object, included in one of the adjacent neighboring frames, from the target pixels based on additional non-target pixels in an additional frame being more similar to the non-object pixels of the static object than to the target pixels.
20. The system of claim 19, wherein the additional frame is included in an N-frame and is not included in an object track.
US13/186,552 2006-12-29 2011-07-20 Method and Apparatus for Processing Video Data Abandoned US20110274367A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/186,552 US20110274367A1 (en) 2006-12-29 2011-07-20 Method and Apparatus for Processing Video Data

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US11/648,294 US8009925B2 (en) 2006-12-29 2006-12-29 Method and apparatus for processing video data
US13/186,552 US20110274367A1 (en) 2006-12-29 2011-07-20 Method and Apparatus for Processing Video Data

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US11/648,294 Continuation US8009925B2 (en) 2006-12-29 2006-12-29 Method and apparatus for processing video data

Publications (1)

Publication Number Publication Date
US20110274367A1 true US20110274367A1 (en) 2011-11-10

Family

ID=39584115

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/648,294 Expired - Fee Related US8009925B2 (en) 2006-12-29 2006-12-29 Method and apparatus for processing video data
US13/186,552 Abandoned US20110274367A1 (en) 2006-12-29 2011-07-20 Method and Apparatus for Processing Video Data

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/648,294 Expired - Fee Related US8009925B2 (en) 2006-12-29 2006-12-29 Method and apparatus for processing video data

Country Status (1)

Country Link
US (2) US8009925B2 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102637253B (en) * 2011-12-30 2014-02-19 清华大学 Video foreground object extracting method based on visual saliency and superpixel division
US20190275600A1 (en) * 2018-03-07 2019-09-12 Powertech Technology Inc. Flux transfer tool and flux transfer method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
Alan J. Lipton,Local Application of Optic Flow to Analyse Rigid versus Non-Rigid Motion, CMU 1999 *
Michael Oren et al, Pedestrian Detection Using Wavelet Templates, IEEE 1997 *
S.M. Smith, Real-Time Motion Segmentation and Object Tracking, Oxford Centre for Functional Magnetic Resonance Imaging of the Brain. 1995 *

Also Published As

Publication number Publication date
US8009925B2 (en) 2011-08-30
US20080159648A1 (en) 2008-07-03

Similar Documents

Publication Publication Date Title
US7783118B2 (en) Method and apparatus for determining motion in images
US10318813B1 (en) Digital video fingerprinting using motion segmentation
US8761582B2 (en) Video editing device and video editing system
US9171197B2 (en) Facial tracking method
US8488682B2 (en) System and method for extracting text captions from video and generating video summaries
US6456730B1 (en) Moving object detection apparatus and method
JP4626886B2 (en) Method and apparatus for locating and extracting captions in digital images
EP2034426A1 (en) Moving image analyzing, method and system
US7783106B2 (en) Video segmentation combining similarity analysis and classification
US20080143880A1 (en) Method and apparatus for detecting caption of video
EP1840798A1 (en) Method for classifying digital image data
KR20080095787A (en) Apparatus and method of determining similar image
Oh et al. Content-based scene change detection and classification technique using background tracking
US11710315B2 (en) Method, electronic apparatus and storage medium for detecting a static logo of a video
US8009925B2 (en) Method and apparatus for processing video data
Cózar et al. Logotype detection to support semantic-based video annotation
Heng et al. High accuracy flashlight scene determination for shot boundary detection
Zhou et al. Uhp-sot++: An unsupervised lightweight single object tracker
Halima et al. A comprehensive method for Arabic video text detection, localization, extraction and recognition
Lu et al. An accumulation algorithm for video shot boundary detection
CN112380970B (en) Video target detection method based on local area search
CN113888562A (en) Target detection method and device, computing equipment and storage medium
Li et al. An integration text extraction approach in video frame
Ngo et al. Motion Driven Approaches to Shot Boundary Detection, Low-Level Feature Extraction and BBC Rushes Characterization at TRECVID 2005.
Timofte et al. Automatic stave discovery for musical facsimiles

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION