WO2008152556A1 - Procédé et appareil pour générer automatiquement des résumés d'un fichier multimédia - Google Patents

Procédé et appareil pour générer automatiquement des résumés d'un fichier multimédia Download PDF

Info

Publication number
WO2008152556A1
WO2008152556A1 PCT/IB2008/052250 IB2008052250W WO2008152556A1 WO 2008152556 A1 WO2008152556 A1 WO 2008152556A1 IB 2008052250 W IB2008052250 W IB 2008052250W WO 2008152556 A1 WO2008152556 A1 WO 2008152556A1
Authority
WO
WIPO (PCT)
Prior art keywords
segments
multimedia file
content
generating
summaries
Prior art date
Application number
PCT/IB2008/052250
Other languages
English (en)
Inventor
Johannes Weda
Marco E. Campanella
Mauro Barbieri
Prarthana Shrestha
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics N.V. filed Critical Koninklijke Philips Electronics N.V.
Priority to CN2008800203066A priority Critical patent/CN101743596B/zh
Priority to EP08763246A priority patent/EP2156438A1/fr
Priority to JP2010511756A priority patent/JP2010531561A/ja
Priority to US12/663,529 priority patent/US20100185628A1/en
Publication of WO2008152556A1 publication Critical patent/WO2008152556A1/fr

Links

Classifications

    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/738Presentation of query results
    • G06F16/739Presentation of query results in form of a video summary, e.g. the video summary being a video sequence, a composite still image or having synthesized frames
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/02Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
    • G11B27/031Electronic editing of digitised analogue information signals, e.g. audio or video signals
    • G11B27/034Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel

Definitions

  • the present invention relates to a method and apparatus for automatically generating a plurality of summaries of a multimedia file. In particular, but not exclusively, it relates to generating summaries of captured video.
  • Summary generation is particularly useful, for example, to people who regularly capture video.
  • Today an increasing number of people regularly capture video. This is due to the cheap, easy and effortless availability of video cameras in dedicated devices (such as a camcorders), or video cameras embedded in cell phones. As a result a user's collection of video recordings can become excessively large, making reviewing and browsing increasingly difficult.
  • the raw video material may be lengthy and rather boring to watch. It may be desirable to edit the raw material to show occurrence of major events. Since video is a massive stream of data, it is difficult to access, split, change, extract parts from and merge, in other words, to edit at a "scene" level, i.e. groups of shots naturally belonging together to create a scene. To assist users in a cheap and easy manner, several commercial software packages are available that allow users to edit their recordings.
  • One example of such a known software package is an extensive and powerful tool known as a non-linear video editing tool that gives the user full control on a frame level.
  • the user needs to be familiar with technical and aesthetic aspects of composing desired video footage out of the raw material. Specific examples of such software packages are "Adobe Premiere” and "Ulead Video Studio 9", which can be found at www.ulead.com/vs.
  • a further known system for summarizing a visual recording is disclosed by US 2004/0052505.
  • multiple visual summaries are created from a single visual recording such that segments in a first summary of the visual recording are not included in other summaries created from the same visual recording.
  • the summaries are created according to an automated technique and the multiple summaries can be stored for selection or creation of a final summary.
  • the summaries are created using the same selection technique and contain similar content. The user, in considering the content, which has been excluded, must view all the summaries which is time consuming and cumbersome.
  • the problems with the known systems mentioned above are that they do not give the user easy access, control or an overview of segments excluded from the automatically generated summaries. This is a particular problem for large summary compressions (i.e. summaries that only include a small fraction of the original multimedia file) as the user is required to view all of the multimedia file and compare it to the automatically generated summary in order to determine segments that have been excluded. This forms a difficult and cumbersome problem for the user.
  • the present invention seeks to provide a method for automatically generating a plurality of summaries of a multimedia file that overcomes the disadvantages associated with known methods.
  • the present invention seeks to extend the known systems by not only automatically generating a first summary, but also generating a summary of the segments of the multimedia file not included in the first summary.
  • the invention therefore extends the second group of software packages discussed earlier by providing more control and overview to the user, without entering the complicated field of non- linear editing.
  • a method for automatically generating a plurality of summaries of a multimedia file comprising the steps of: generating a first summary of a multimedia file; generating at least one second summary of the multimedia file, the at least one second summary including content excluded from the first summary, wherein the content of the at least one second summary is selected such that it is semantically different to the content of the first summary.
  • apparatus for automatically generating a plurality of summaries of a multimedia file comprising: means for generating a first summary of a multimedia file; and means for generating at least one second summary of the multimedia file, the at least one second summary including content excluded from the first summary, wherein the content of the at least one second summary is selected such that it is semantically different to the content of the first summary.
  • the user is provided with a first summary and also at least one second summary including the segments of the multimedia file that were omitted from the first summary.
  • the method for generating a summary of a multimedia file is not merely a general content summarization algorithm, but further enables the generation of a summary of the missing segments of a multimedia file.
  • the missing segments being selected such that they are semantically different to the segments selected for the first summary giving the user a clear indication of the overall content of the file and providing the user with a different view of a summary of the content of the file.
  • the content of the at least one second summary may be selected such that it is most semantically different to the content of the first summary.
  • the summary of the missing segments is such that it focuses on the segments of the multimedia file that mostly differ from the segments included in the first summary, thus the user is provided with a summarized view of a more complete range of the content of the file.
  • the multimedia file is divided into a plurality of segments and the step of generating at least one second summary comprises the steps of: determining a measure of a semantic distance between segments included in the first summary and segments excluded from the first summary; including segments in the at least one second summary having a measure of a semantic distance above a threshold.
  • the multimedia file is divided into a plurality of segments and the step of generating at least one second summary comprises the steps of: determining a measure of a semantic distance between segments included in the first summary and segments excluded from the first summary; including segments in the at least one second summary having a highest measure of a semantic distance.
  • the at least one second summary covers the content excluded from the first summary efficiently, without overloading the user with too many details. This is important if the multimedia file is much longer than the first summary, which means that the number of segments not included in the first summary is much higher than the segments included in the first summary.
  • the at least one second summary more compact to allow the user efficient and effective browsing and selecting, which takes into account the attention and time capabilities of the user.
  • the semantic distance may be determined from the audio and/or visual content of the plurality of segments of the multimedia file.
  • the semantic distance may be determined from the color histograms distances and/or temporal distance of the plurality of segments of the multimedia file.
  • the semantic difference may be determined from location data, and/or person data, and/or focus object data.
  • the missing segments can be found by looking for a person, a location and a focus object (i.e. objects taking up a large part of multiple frames) that are not present in the included segments.
  • the method may further comprise the steps of: selecting at least one segment of the at least one second summary; and incorporating the selected at least one segment into the first summary.
  • selecting at least one segment of the at least one second summary may be included in the first summary, creating a more personalized summary.
  • the segments included in the at least one second summary may be grouped such that the content of the segments is similar.
  • a plurality of second summaries may be organized in accordance with their degree of similarity to the content of the first summary for browsing the plurality of second summaries. In this way, the plurality of second summaries are efficiently and effectively shown to a user.
  • the invention can be applied to hard disk recorders, camcorders, video-editing software. Due to its simplicity, the user interface can easily be implemented in consumer products such as hard disk recorders.
  • Fig. 1 is a flowchart of a known method for automatically generating a plurality of summaries of a multimedia file according to prior art
  • Fig. 2 is a simplified schematic of apparatus according to an embodiment of the present invention
  • Fig. 3 is a flowchart of a method for automatically generating a plurality of summaries of a multimedia file according to an embodiment of the present invention.
  • the multimedia file is first imported, step 102.
  • the multimedia file is then segmented according to features (for example, low-level audiovisual features) extracted from the multimedia file, step 104.
  • features for example, low-level audiovisual features
  • the user can set parameters for segmentation, (such as presence of faces and camera motion) and can also manually indicate which segments should definitely end up in the summary, step 106.
  • the system automatically generates a summary of the content of the multimedia file based on internal and/or user-defined settings, step 108. This step involves selecting segments to include in the summary of the multimedia file.
  • the generated summary is then shown to the user, step 110.
  • the user By viewing the summary, the user is able to see which segments have been included in the summary. However, the user has no way of knowing which segments have been excluded from the summary, unless the user views the entire multimedia file and compares it with the generated summary.
  • the apparatus 200 of an embodiment of the present invention comprises an input terminal 202 for input of a multimedia file.
  • the multimedia file is input into a segmenting means 204 via the input terminal 202.
  • the output of the segmenting means 204 is connected to a first generating means 206.
  • the output of the first generating means 206 is output on the output terminal 208.
  • the output of the first generating means 206 is also connected to a measuring means 210.
  • the output of the measuring means 210 is connected to a second generating means 212.
  • the output of the second generating means 212 is output on the output terminal 214.
  • the apparatus 200 also comprises another input terminal 216 for input into the measuring means 210.
  • a multimedia file is imported and input on the input terminal 202, step 302.
  • the segmenting means 204 receives the multimedia file via the input terminal 202.
  • the segmenting means 204 divides the multimedia file into a plurality of segments, step 304.
  • a user may, for example, set parameters for segmentation that indicate which segments they wish be included in the summary, step 306.
  • the segmenting means 204 inputs the plurality of segments into the first generating means 206.
  • the first generating means 206 generates a first summary of the multimedia file (step 308) and outputs the generated summary on the first output terminal 208 (step 310).
  • the first generating means 206 inputs the segments included in the generated summary and the segments excluded from the generated summary into the measuring means 210.
  • the measuring means 210 determines a measure of a semantic distance between segments included in the first summary and segments excluded from the first summary.
  • the second summary generated by the second generating means 212 is then based on the segments determined to be semantically different from the segments included in the first summary. Therefore, it is possible to establish if two video segments contain correlated or uncorrelated semantics. If the semantic distance between segments included in the first summary and segments excluded from the first summary is determined to be low, the segments have similar semantic content.
  • the measuring means 210 may determine the semantic distance, for example, from the audio and/or visual content of the plurality of segments of the multimedia file. Further the semantic distance may be based on location data which may be generated independently, for example, GPS data or from recognition of objects captured by images of the multimedia file. The semantic distance may be based on person data which may be derived automatically from facial recognition of persons captured by images of the multimedia file. The semantic distance may be based on focus object data, i.e. objects which take up a large part of multiple frames.
  • At least one of the one or more segments is preferably included in the second summary.
  • the measuring means 210 may determine the semantic distance from the color histograms distances and/or temporal distance of the plurality of segments of the multimedia file.
  • the semantic distance between segments i and j is given by,
  • ⁇ 1 ' J' is the semantic distance between segments i and j
  • c ⁇ ' ⁇ > ' is the color histograms distance between segments i and j
  • ⁇ ⁇ ⁇ > J 1S the temporal distance between i and j
  • ⁇ > f ⁇ L 1 i is an appropriate function to combine the two distances.
  • the function J f ⁇ L 1 J may be given by,
  • the output of the measuring means 210 is input into the second generating means 212.
  • the second generating means 212 generates at least one second summary of the multimedia file, step 314.
  • the second generating means 212 generates the at least one second summary such that it includes content excluded from the first summary that was determined to be semantically different to the content of the first summary by the measuring means 210 (step 312).
  • the second generating means 212 generates at least one second summary that includes segments having a measure of a semantic distance above a threshold. This means that only segments that have uncorrelated semantic content with the first summary are included in the second summary.
  • the second generating means 212 generates at least one second summary that includes segments having a highest measure of a semantic distance.
  • i each segment included in the first summary S and c is the representative segment for cluster C.
  • the second generating means 212 uses the distance * ⁇ ' ⁇ to rank the clusters of the segments excluded from the first summary on the basis of the semantic distance they have with the first summary S. Then, the second generating means 212 generates at least one second summary that includes segments having a highest measure of a semantic distance (i.e. segments that differ the most from the segments of the first summary).
  • the second generating means 212 generates at least one second summary that includes segments having similar content.
  • the second generating means 212 may generate the at least one second summary using a correlation dimension.
  • the second generating means 212 positions the segments on a correlation scale according to their correlation with the segments included in the first summary.
  • the second generating means 212 could then identify segments that are very similar, rather similar, or totally different from the segments included in the first summary and thus generates at least one second summary according to a degree of similarity selected by the user.
  • the second generating means 212 organizes the second summaries in accordance with their degree of similarity to the content of the first summary for browsing the plurality of second summaries, step 316.
  • the second generating means 212 may cluster the segments excluded from the first summary and organize them according to the semantic distance between segments D ⁇ l 'J> , (as defined, for example, in equation (I)).
  • the second generating means 212 may cluster segments that are close to each other according to a semantic distance such that each cluster contains segments having the same semantic distance.
  • the second generating means 212 then outputs the most relevant clusters with respect to the degree of similarity specified by the user on the second output terminal 214, step 318. In this way, the user is not required to browse a large number of second summaries, which would be cumbersome and time consuming. Examples of clustering techniques can be found in "Self- organizing formation of topologically correct feature maps", T. Kohonen, Biological Cybernetics 43(1), pp. 59-69, 1982 and "Pattern Recognition Principles", J. T. Tou and R. C. Gonzalez, Addison- Wesley Publishing Co, 1974.
  • the second generating means 212 may cluster and organize the segments in a hierarchical way such that the main clusters contain other clusters.
  • the second generating means 212 then outputs the main clusters on the second output terminal 214 (step 318).
  • the user only has to browse a small number of main clusters. Then, if they desire, the user can explore each of the other clusters in more and more detail with a few interactions. This makes browsing a plurality of second summaries very easy. The user is able to view the first summary output on the first output terminal
  • step 310) and the at least one second summary output on the second output terminal 214 (step 318).
  • the user can provide feedback via the input terminal 216, step 320. For example, the user may review the second summary and select segments to be included in the first summary.
  • the user feedback is input into the measuring means 210 via the input terminal 216.
  • the measuring means 210 selects at least one segment of the at least one second summary such that the feedback of the user is taken into account, step 322.
  • the measuring means 210 inputs the selected at least one segment into the first generating means 206.
  • the first generating means 206 then incorporates the selected at least one segment into the first summary (step 308) and outputs the first summary of the first output terminal 208 (step 310). While the invention has been described in connection with preferred embodiments, it will be understood that modifications thereof within the principles outlined above will be evident to those skilled in the art, and thus the invention is not limited to the preferred embodiments but is intended to encompass such modifications. The invention resides in each and every novel characteristic feature and each and every combination of characteristic features. Reference numerals in the claims do not limit their protective scope. Use of the verb "to comprise” and its conjugations does not exclude the presence of elements other than those stated in the claims. Use of the article "a” or “an” preceding an element does not exclude the presence of a plurality of such elements.
  • 'Means' as will be apparent to a person skilled in the art, are meant to include any hardware (such as separate or integrated circuits or electronic elements) or software (such as programs or parts of programs) which perform in operation or are designed to perform a specified function, be it solely or in conjunction with other functions, be it in isolation or in co-operation with other elements.
  • the invention can be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the apparatus claim enumerating several means, several of these means can be embodied by one and the same item of hardware.
  • 'Computer program product' is to be understood to mean any software product stored on a computer-readable medium, such as a floppy disk, downloadable via a network, such as the Internet, or marketable in any other manner.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Theoretical Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Television Signal Processing For Recording (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

La présente invention concerne une pluralité de résumés d'un fichier multimédia qui sont générés automatiquement. Un premier résumé d'un fichier multimédia est généré (étape 308). Au moins un second résumé du fichier multimédia est ensuite généré (étape 314). Le second résumé comprend le contenu exclu du premier. Le contenu du second résumé est sélectionné de sorte qu'il soit sémantiquement différent du contenu du premier (étape 312).
PCT/IB2008/052250 2007-06-15 2008-06-09 Procédé et appareil pour générer automatiquement des résumés d'un fichier multimédia WO2008152556A1 (fr)

Priority Applications (4)

Application Number Priority Date Filing Date Title
CN2008800203066A CN101743596B (zh) 2007-06-15 2008-06-09 用于自动生成多媒体文件概要的方法和设备
EP08763246A EP2156438A1 (fr) 2007-06-15 2008-06-09 Procédé et appareil pour générer automatiquement des résumés d'un fichier multimédia
JP2010511756A JP2010531561A (ja) 2007-06-15 2008-06-09 マルチメディアファイルのサマリを自動的に生成する方法及び装置
US12/663,529 US20100185628A1 (en) 2007-06-15 2008-06-09 Method and apparatus for automatically generating summaries of a multimedia file

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP07110324.6 2007-06-15
EP07110324 2007-06-15

Publications (1)

Publication Number Publication Date
WO2008152556A1 true WO2008152556A1 (fr) 2008-12-18

Family

ID=39721940

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/IB2008/052250 WO2008152556A1 (fr) 2007-06-15 2008-06-09 Procédé et appareil pour générer automatiquement des résumés d'un fichier multimédia

Country Status (6)

Country Link
US (1) US20100185628A1 (fr)
EP (1) EP2156438A1 (fr)
JP (1) JP2010531561A (fr)
KR (1) KR20100018070A (fr)
CN (1) CN101743596B (fr)
WO (1) WO2008152556A1 (fr)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012019305A (ja) * 2010-07-07 2012-01-26 Nippon Telegr & Teleph Corp <Ntt> 映像要約装置,映像要約方法および映像要約プログラム

Families Citing this family (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3876107A1 (fr) 2013-03-15 2021-09-08 Factual Inc. Appareil, systèmes et procédés pour analyser des mouvements d'entités cibles
US10095783B2 (en) 2015-05-25 2018-10-09 Microsoft Technology Licensing, Llc Multiple rounds of results summarization for improved latency and relevance
CN105228033B (zh) * 2015-08-27 2018-11-09 联想(北京)有限公司 一种视频处理方法及电子设备
US10321196B2 (en) * 2015-12-09 2019-06-11 Rovi Guides, Inc. Methods and systems for customizing a media asset with feedback on customization
KR102592904B1 (ko) * 2016-02-19 2023-10-23 삼성전자주식회사 영상 요약 장치 및 방법
WO2017142143A1 (fr) * 2016-02-19 2017-08-24 Samsung Electronics Co., Ltd. Procédé et appareil permettant de fournir des informations de résumé d'une vidéo
DE102018202514A1 (de) * 2018-02-20 2019-08-22 Bayerische Motoren Werke Aktiengesellschaft System und Verfahren zur automatischen Erstellung eines Videos einer Fahrt

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040052505A1 (en) 2002-05-28 2004-03-18 Yesvideo, Inc. Summarization of a visual recording
US20050002647A1 (en) * 2003-07-02 2005-01-06 Fuji Xerox Co., Ltd. Systems and methods for generating multi-level hypervideo summaries

Family Cites Families (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3823333B2 (ja) * 1995-02-21 2006-09-20 株式会社日立製作所 動画像の変化点検出方法、動画像の変化点検出装置、動画像の変化点検出システム
JP3240871B2 (ja) * 1995-03-07 2001-12-25 松下電器産業株式会社 映像要約方法
JPH10232884A (ja) * 1996-11-29 1998-09-02 Media Rinku Syst:Kk 映像ソフトウェア処理方法及び映像ソフトウェア処理装置
JP2000285243A (ja) * 1999-01-29 2000-10-13 Sony Corp 信号処理方法及び映像音声処理装置
JP2001014306A (ja) * 1999-06-30 2001-01-19 Sony Corp 電子文書処理方法及び電子文書処理装置並びに電子文書処理プログラムが記録された記録媒体
US7016540B1 (en) * 1999-11-24 2006-03-21 Nec Corporation Method and system for segmentation, classification, and summarization of video images
AUPQ535200A0 (en) * 2000-01-31 2000-02-17 Canon Kabushiki Kaisha Extracting key frames from a video sequence
US7042493B2 (en) * 2000-04-07 2006-05-09 Paolo Prandoni Automated stroboscoping of video sequences
US7296231B2 (en) * 2001-08-09 2007-11-13 Eastman Kodak Company Video structuring by probabilistic merging of video segments
US20030117428A1 (en) * 2001-12-20 2003-06-26 Koninklijke Philips Electronics N.V. Visual summary of audio-visual program features
US7333712B2 (en) * 2002-02-14 2008-02-19 Koninklijke Philips Electronics N.V. Visual summary for scanning forwards and backwards in video content
US7184955B2 (en) * 2002-03-25 2007-02-27 Hewlett-Packard Development Company, L.P. System and method for indexing videos based on speaker distinction
JP4067326B2 (ja) * 2002-03-26 2008-03-26 富士通株式会社 動画像内容表示装置
JP2003330941A (ja) * 2002-05-08 2003-11-21 Olympus Optical Co Ltd 類似画像分類装置
FR2845179B1 (fr) * 2002-09-27 2004-11-05 Thomson Licensing Sa Procede de regroupement d'images d'une sequence video
US7143352B2 (en) * 2002-11-01 2006-11-28 Mitsubishi Electric Research Laboratories, Inc Blind summarization of video content
JP2004187029A (ja) * 2002-12-04 2004-07-02 Toshiba Corp 要約映像追いかけ再生装置
US20040181545A1 (en) * 2003-03-10 2004-09-16 Yining Deng Generating and rendering annotated video files
US20050257242A1 (en) * 2003-03-14 2005-11-17 Starz Entertainment Group Llc Multicast video edit control
JP4344534B2 (ja) * 2003-04-30 2009-10-14 セコム株式会社 画像処理システム
KR100590537B1 (ko) * 2004-02-18 2006-06-15 삼성전자주식회사 복수 영상의 요약 방법 및 장치
JP2005277445A (ja) * 2004-03-22 2005-10-06 Fuji Xerox Co Ltd 会議映像処理装置、会議映像処理方法およびプログラム
US7302451B2 (en) * 2004-05-07 2007-11-27 Mitsubishi Electric Research Laboratories, Inc. Feature identification of events in multimedia
JP4140579B2 (ja) * 2004-08-11 2008-08-27 ソニー株式会社 画像処理装置および方法、撮影装置、並びにプログラム
JP4641450B2 (ja) * 2005-05-23 2011-03-02 日本電信電話株式会社 非定常画像検出方法、非定常画像検出装置、及び非定常画像検出プログラム
US7555149B2 (en) * 2005-10-25 2009-06-30 Mitsubishi Electric Research Laboratories, Inc. Method and system for segmenting videos using face detection

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20040052505A1 (en) 2002-05-28 2004-03-18 Yesvideo, Inc. Summarization of a visual recording
US20050002647A1 (en) * 2003-07-02 2005-01-06 Fuji Xerox Co., Ltd. Systems and methods for generating multi-level hypervideo summaries

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
ANDREAS GIRGENSOHN, JOHN BORECZKY, LYNN WILCOX: "Keyframe-Based User Interfaces for Digital Video", IEEE COMPUTER MAGAZINE, vol. 34, no. 9, September 2001 (2001-09-01), pages 61 - 67, XP002494967 *
BENINI, S. ET AL.: "Extraction of Significant Video Summaries by Dendrogram Analysis", 2006 IEEE INT. CONF. ON IMAGE PROCESSING, 11 October 2006 (2006-10-11), pages 133 - 136, XP031048591, DOI: doi:10.1109/ICIP.2006.312377
GIRGENSOHN, A. ET AL.: "Keyframe-based User Interfaces for Digital Video", IEEE COMPUTER MAGAZINE, vol. 34, no. 9, September 2001 (2001-09-01), pages 61 - 67, XP002494967, DOI: doi:10.1109/2.947093
S. BENINI, A. BIANCHETTI, R. LEONARDI, P. MIGLIORATI: "Extraction of Significant Video Summaries by Dendrogram Analysis", 2006 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, 11 October 2006 (2006-10-11), pages 133 - 136, XP002494966 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2012019305A (ja) * 2010-07-07 2012-01-26 Nippon Telegr & Teleph Corp <Ntt> 映像要約装置,映像要約方法および映像要約プログラム

Also Published As

Publication number Publication date
KR20100018070A (ko) 2010-02-16
CN101743596B (zh) 2012-05-30
EP2156438A1 (fr) 2010-02-24
JP2010531561A (ja) 2010-09-24
CN101743596A (zh) 2010-06-16
US20100185628A1 (en) 2010-07-22

Similar Documents

Publication Publication Date Title
Li et al. An overview of video abstraction techniques
US7383508B2 (en) Computer user interface for interacting with video cliplets generated from digital video
Truong et al. Video abstraction: A systematic review and classification
US7702185B2 (en) Use of image similarity in annotating groups of visual images in a collection of visual images
RU2440606C2 (ru) Способ и устройство автоматического генерирования сводки множества изображений
US7697792B2 (en) Process-response statistical modeling of a visual image for use in determining similarity between visual images
US8316301B2 (en) Apparatus, medium, and method segmenting video sequences based on topic
US8750681B2 (en) Electronic apparatus, content recommendation method, and program therefor
US20100185628A1 (en) Method and apparatus for automatically generating summaries of a multimedia file
US20060020597A1 (en) Use of image similarity in summarizing a collection of visual images
EP1557837A1 (fr) Elimination des redondances dans un système d&#39;aperçu vidéo opérant de manière adaptative selon le contenu
US20060015495A1 (en) Use of image similarity in image searching via a network of computational apparatus
US20060015497A1 (en) Content-based indexing or grouping of visual images, with particular use of image similarity to effect same
US20060015494A1 (en) Use of image similarity in selecting a representative visual image for a group of visual images
Chen et al. Tiling slideshow
EP2530605A1 (fr) Dispositif de traitement de données
Jiang et al. Automatic consumer video summarization by audio and visual analysis
WO2002082328A2 (fr) Meta-donnees de camera pour classification de contenu
KR20070118635A (ko) 오디오 및/또는 비주얼 데이터의 서머라이제이션
Dimitrova et al. Video keyframe extraction and filtering: a keyframe is not a keyframe to everyone
JP2009123095A (ja) 映像解析装置及び映像解析方法
EP1527453A1 (fr) Procede, systeme et produit de programme pour obtenir une table de contenus fondee sur le contenu
El-Bendary et al. PCA-based home videos annotation system
Fersini et al. Multimedia summarization in law courts: a clustering-based environment for browsing and consulting judicial folders
Adami et al. The ToCAI description scheme for indexing and retrieval of multimedia documents

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200880020306.6

Country of ref document: CN

121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 08763246

Country of ref document: EP

Kind code of ref document: A1

WWE Wipo information: entry into national phase

Ref document number: 2008763246

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2010511756

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 12663529

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 119/CHENP/2010

Country of ref document: IN

ENP Entry into the national phase

Ref document number: 20107000745

Country of ref document: KR

Kind code of ref document: A