WO2008152556A1 - Procédé et appareil pour générer automatiquement des résumés d'un fichier multimédia - Google Patents
Procédé et appareil pour générer automatiquement des résumés d'un fichier multimédia Download PDFInfo
- Publication number
- WO2008152556A1 WO2008152556A1 PCT/IB2008/052250 IB2008052250W WO2008152556A1 WO 2008152556 A1 WO2008152556 A1 WO 2008152556A1 IB 2008052250 W IB2008052250 W IB 2008052250W WO 2008152556 A1 WO2008152556 A1 WO 2008152556A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- segments
- multimedia file
- content
- generating
- summaries
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
- G06F16/738—Presentation of query results
- G06F16/739—Presentation of query results in form of a video summary, e.g. the video summary being a video sequence, a composite still image or having synthesized frames
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/02—Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
- G11B27/031—Electronic editing of digitised analogue information signals, e.g. audio or video signals
- G11B27/034—Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
Definitions
- the present invention relates to a method and apparatus for automatically generating a plurality of summaries of a multimedia file. In particular, but not exclusively, it relates to generating summaries of captured video.
- Summary generation is particularly useful, for example, to people who regularly capture video.
- Today an increasing number of people regularly capture video. This is due to the cheap, easy and effortless availability of video cameras in dedicated devices (such as a camcorders), or video cameras embedded in cell phones. As a result a user's collection of video recordings can become excessively large, making reviewing and browsing increasingly difficult.
- the raw video material may be lengthy and rather boring to watch. It may be desirable to edit the raw material to show occurrence of major events. Since video is a massive stream of data, it is difficult to access, split, change, extract parts from and merge, in other words, to edit at a "scene" level, i.e. groups of shots naturally belonging together to create a scene. To assist users in a cheap and easy manner, several commercial software packages are available that allow users to edit their recordings.
- One example of such a known software package is an extensive and powerful tool known as a non-linear video editing tool that gives the user full control on a frame level.
- the user needs to be familiar with technical and aesthetic aspects of composing desired video footage out of the raw material. Specific examples of such software packages are "Adobe Premiere” and "Ulead Video Studio 9", which can be found at www.ulead.com/vs.
- a further known system for summarizing a visual recording is disclosed by US 2004/0052505.
- multiple visual summaries are created from a single visual recording such that segments in a first summary of the visual recording are not included in other summaries created from the same visual recording.
- the summaries are created according to an automated technique and the multiple summaries can be stored for selection or creation of a final summary.
- the summaries are created using the same selection technique and contain similar content. The user, in considering the content, which has been excluded, must view all the summaries which is time consuming and cumbersome.
- the problems with the known systems mentioned above are that they do not give the user easy access, control or an overview of segments excluded from the automatically generated summaries. This is a particular problem for large summary compressions (i.e. summaries that only include a small fraction of the original multimedia file) as the user is required to view all of the multimedia file and compare it to the automatically generated summary in order to determine segments that have been excluded. This forms a difficult and cumbersome problem for the user.
- the present invention seeks to provide a method for automatically generating a plurality of summaries of a multimedia file that overcomes the disadvantages associated with known methods.
- the present invention seeks to extend the known systems by not only automatically generating a first summary, but also generating a summary of the segments of the multimedia file not included in the first summary.
- the invention therefore extends the second group of software packages discussed earlier by providing more control and overview to the user, without entering the complicated field of non- linear editing.
- a method for automatically generating a plurality of summaries of a multimedia file comprising the steps of: generating a first summary of a multimedia file; generating at least one second summary of the multimedia file, the at least one second summary including content excluded from the first summary, wherein the content of the at least one second summary is selected such that it is semantically different to the content of the first summary.
- apparatus for automatically generating a plurality of summaries of a multimedia file comprising: means for generating a first summary of a multimedia file; and means for generating at least one second summary of the multimedia file, the at least one second summary including content excluded from the first summary, wherein the content of the at least one second summary is selected such that it is semantically different to the content of the first summary.
- the user is provided with a first summary and also at least one second summary including the segments of the multimedia file that were omitted from the first summary.
- the method for generating a summary of a multimedia file is not merely a general content summarization algorithm, but further enables the generation of a summary of the missing segments of a multimedia file.
- the missing segments being selected such that they are semantically different to the segments selected for the first summary giving the user a clear indication of the overall content of the file and providing the user with a different view of a summary of the content of the file.
- the content of the at least one second summary may be selected such that it is most semantically different to the content of the first summary.
- the summary of the missing segments is such that it focuses on the segments of the multimedia file that mostly differ from the segments included in the first summary, thus the user is provided with a summarized view of a more complete range of the content of the file.
- the multimedia file is divided into a plurality of segments and the step of generating at least one second summary comprises the steps of: determining a measure of a semantic distance between segments included in the first summary and segments excluded from the first summary; including segments in the at least one second summary having a measure of a semantic distance above a threshold.
- the multimedia file is divided into a plurality of segments and the step of generating at least one second summary comprises the steps of: determining a measure of a semantic distance between segments included in the first summary and segments excluded from the first summary; including segments in the at least one second summary having a highest measure of a semantic distance.
- the at least one second summary covers the content excluded from the first summary efficiently, without overloading the user with too many details. This is important if the multimedia file is much longer than the first summary, which means that the number of segments not included in the first summary is much higher than the segments included in the first summary.
- the at least one second summary more compact to allow the user efficient and effective browsing and selecting, which takes into account the attention and time capabilities of the user.
- the semantic distance may be determined from the audio and/or visual content of the plurality of segments of the multimedia file.
- the semantic distance may be determined from the color histograms distances and/or temporal distance of the plurality of segments of the multimedia file.
- the semantic difference may be determined from location data, and/or person data, and/or focus object data.
- the missing segments can be found by looking for a person, a location and a focus object (i.e. objects taking up a large part of multiple frames) that are not present in the included segments.
- the method may further comprise the steps of: selecting at least one segment of the at least one second summary; and incorporating the selected at least one segment into the first summary.
- selecting at least one segment of the at least one second summary may be included in the first summary, creating a more personalized summary.
- the segments included in the at least one second summary may be grouped such that the content of the segments is similar.
- a plurality of second summaries may be organized in accordance with their degree of similarity to the content of the first summary for browsing the plurality of second summaries. In this way, the plurality of second summaries are efficiently and effectively shown to a user.
- the invention can be applied to hard disk recorders, camcorders, video-editing software. Due to its simplicity, the user interface can easily be implemented in consumer products such as hard disk recorders.
- Fig. 1 is a flowchart of a known method for automatically generating a plurality of summaries of a multimedia file according to prior art
- Fig. 2 is a simplified schematic of apparatus according to an embodiment of the present invention
- Fig. 3 is a flowchart of a method for automatically generating a plurality of summaries of a multimedia file according to an embodiment of the present invention.
- the multimedia file is first imported, step 102.
- the multimedia file is then segmented according to features (for example, low-level audiovisual features) extracted from the multimedia file, step 104.
- features for example, low-level audiovisual features
- the user can set parameters for segmentation, (such as presence of faces and camera motion) and can also manually indicate which segments should definitely end up in the summary, step 106.
- the system automatically generates a summary of the content of the multimedia file based on internal and/or user-defined settings, step 108. This step involves selecting segments to include in the summary of the multimedia file.
- the generated summary is then shown to the user, step 110.
- the user By viewing the summary, the user is able to see which segments have been included in the summary. However, the user has no way of knowing which segments have been excluded from the summary, unless the user views the entire multimedia file and compares it with the generated summary.
- the apparatus 200 of an embodiment of the present invention comprises an input terminal 202 for input of a multimedia file.
- the multimedia file is input into a segmenting means 204 via the input terminal 202.
- the output of the segmenting means 204 is connected to a first generating means 206.
- the output of the first generating means 206 is output on the output terminal 208.
- the output of the first generating means 206 is also connected to a measuring means 210.
- the output of the measuring means 210 is connected to a second generating means 212.
- the output of the second generating means 212 is output on the output terminal 214.
- the apparatus 200 also comprises another input terminal 216 for input into the measuring means 210.
- a multimedia file is imported and input on the input terminal 202, step 302.
- the segmenting means 204 receives the multimedia file via the input terminal 202.
- the segmenting means 204 divides the multimedia file into a plurality of segments, step 304.
- a user may, for example, set parameters for segmentation that indicate which segments they wish be included in the summary, step 306.
- the segmenting means 204 inputs the plurality of segments into the first generating means 206.
- the first generating means 206 generates a first summary of the multimedia file (step 308) and outputs the generated summary on the first output terminal 208 (step 310).
- the first generating means 206 inputs the segments included in the generated summary and the segments excluded from the generated summary into the measuring means 210.
- the measuring means 210 determines a measure of a semantic distance between segments included in the first summary and segments excluded from the first summary.
- the second summary generated by the second generating means 212 is then based on the segments determined to be semantically different from the segments included in the first summary. Therefore, it is possible to establish if two video segments contain correlated or uncorrelated semantics. If the semantic distance between segments included in the first summary and segments excluded from the first summary is determined to be low, the segments have similar semantic content.
- the measuring means 210 may determine the semantic distance, for example, from the audio and/or visual content of the plurality of segments of the multimedia file. Further the semantic distance may be based on location data which may be generated independently, for example, GPS data or from recognition of objects captured by images of the multimedia file. The semantic distance may be based on person data which may be derived automatically from facial recognition of persons captured by images of the multimedia file. The semantic distance may be based on focus object data, i.e. objects which take up a large part of multiple frames.
- At least one of the one or more segments is preferably included in the second summary.
- the measuring means 210 may determine the semantic distance from the color histograms distances and/or temporal distance of the plurality of segments of the multimedia file.
- the semantic distance between segments i and j is given by,
- ⁇ 1 ' J' is the semantic distance between segments i and j
- c ⁇ ' ⁇ > ' is the color histograms distance between segments i and j
- ⁇ ⁇ ⁇ > J 1S the temporal distance between i and j
- ⁇ > f ⁇ L 1 i is an appropriate function to combine the two distances.
- the function J f ⁇ L 1 J may be given by,
- the output of the measuring means 210 is input into the second generating means 212.
- the second generating means 212 generates at least one second summary of the multimedia file, step 314.
- the second generating means 212 generates the at least one second summary such that it includes content excluded from the first summary that was determined to be semantically different to the content of the first summary by the measuring means 210 (step 312).
- the second generating means 212 generates at least one second summary that includes segments having a measure of a semantic distance above a threshold. This means that only segments that have uncorrelated semantic content with the first summary are included in the second summary.
- the second generating means 212 generates at least one second summary that includes segments having a highest measure of a semantic distance.
- i each segment included in the first summary S and c is the representative segment for cluster C.
- the second generating means 212 uses the distance * ⁇ ' ⁇ to rank the clusters of the segments excluded from the first summary on the basis of the semantic distance they have with the first summary S. Then, the second generating means 212 generates at least one second summary that includes segments having a highest measure of a semantic distance (i.e. segments that differ the most from the segments of the first summary).
- the second generating means 212 generates at least one second summary that includes segments having similar content.
- the second generating means 212 may generate the at least one second summary using a correlation dimension.
- the second generating means 212 positions the segments on a correlation scale according to their correlation with the segments included in the first summary.
- the second generating means 212 could then identify segments that are very similar, rather similar, or totally different from the segments included in the first summary and thus generates at least one second summary according to a degree of similarity selected by the user.
- the second generating means 212 organizes the second summaries in accordance with their degree of similarity to the content of the first summary for browsing the plurality of second summaries, step 316.
- the second generating means 212 may cluster the segments excluded from the first summary and organize them according to the semantic distance between segments D ⁇ l 'J> , (as defined, for example, in equation (I)).
- the second generating means 212 may cluster segments that are close to each other according to a semantic distance such that each cluster contains segments having the same semantic distance.
- the second generating means 212 then outputs the most relevant clusters with respect to the degree of similarity specified by the user on the second output terminal 214, step 318. In this way, the user is not required to browse a large number of second summaries, which would be cumbersome and time consuming. Examples of clustering techniques can be found in "Self- organizing formation of topologically correct feature maps", T. Kohonen, Biological Cybernetics 43(1), pp. 59-69, 1982 and "Pattern Recognition Principles", J. T. Tou and R. C. Gonzalez, Addison- Wesley Publishing Co, 1974.
- the second generating means 212 may cluster and organize the segments in a hierarchical way such that the main clusters contain other clusters.
- the second generating means 212 then outputs the main clusters on the second output terminal 214 (step 318).
- the user only has to browse a small number of main clusters. Then, if they desire, the user can explore each of the other clusters in more and more detail with a few interactions. This makes browsing a plurality of second summaries very easy. The user is able to view the first summary output on the first output terminal
- step 310) and the at least one second summary output on the second output terminal 214 (step 318).
- the user can provide feedback via the input terminal 216, step 320. For example, the user may review the second summary and select segments to be included in the first summary.
- the user feedback is input into the measuring means 210 via the input terminal 216.
- the measuring means 210 selects at least one segment of the at least one second summary such that the feedback of the user is taken into account, step 322.
- the measuring means 210 inputs the selected at least one segment into the first generating means 206.
- the first generating means 206 then incorporates the selected at least one segment into the first summary (step 308) and outputs the first summary of the first output terminal 208 (step 310). While the invention has been described in connection with preferred embodiments, it will be understood that modifications thereof within the principles outlined above will be evident to those skilled in the art, and thus the invention is not limited to the preferred embodiments but is intended to encompass such modifications. The invention resides in each and every novel characteristic feature and each and every combination of characteristic features. Reference numerals in the claims do not limit their protective scope. Use of the verb "to comprise” and its conjugations does not exclude the presence of elements other than those stated in the claims. Use of the article "a” or “an” preceding an element does not exclude the presence of a plurality of such elements.
- 'Means' as will be apparent to a person skilled in the art, are meant to include any hardware (such as separate or integrated circuits or electronic elements) or software (such as programs or parts of programs) which perform in operation or are designed to perform a specified function, be it solely or in conjunction with other functions, be it in isolation or in co-operation with other elements.
- the invention can be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the apparatus claim enumerating several means, several of these means can be embodied by one and the same item of hardware.
- 'Computer program product' is to be understood to mean any software product stored on a computer-readable medium, such as a floppy disk, downloadable via a network, such as the Internet, or marketable in any other manner.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Television Signal Processing For Recording (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2008800203066A CN101743596B (zh) | 2007-06-15 | 2008-06-09 | 用于自动生成多媒体文件概要的方法和设备 |
EP08763246A EP2156438A1 (fr) | 2007-06-15 | 2008-06-09 | Procédé et appareil pour générer automatiquement des résumés d'un fichier multimédia |
JP2010511756A JP2010531561A (ja) | 2007-06-15 | 2008-06-09 | マルチメディアファイルのサマリを自動的に生成する方法及び装置 |
US12/663,529 US20100185628A1 (en) | 2007-06-15 | 2008-06-09 | Method and apparatus for automatically generating summaries of a multimedia file |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
EP07110324.6 | 2007-06-15 | ||
EP07110324 | 2007-06-15 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2008152556A1 true WO2008152556A1 (fr) | 2008-12-18 |
Family
ID=39721940
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2008/052250 WO2008152556A1 (fr) | 2007-06-15 | 2008-06-09 | Procédé et appareil pour générer automatiquement des résumés d'un fichier multimédia |
Country Status (6)
Country | Link |
---|---|
US (1) | US20100185628A1 (fr) |
EP (1) | EP2156438A1 (fr) |
JP (1) | JP2010531561A (fr) |
KR (1) | KR20100018070A (fr) |
CN (1) | CN101743596B (fr) |
WO (1) | WO2008152556A1 (fr) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2012019305A (ja) * | 2010-07-07 | 2012-01-26 | Nippon Telegr & Teleph Corp <Ntt> | 映像要約装置,映像要約方法および映像要約プログラム |
Families Citing this family (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3876107A1 (fr) | 2013-03-15 | 2021-09-08 | Factual Inc. | Appareil, systèmes et procédés pour analyser des mouvements d'entités cibles |
US10095783B2 (en) | 2015-05-25 | 2018-10-09 | Microsoft Technology Licensing, Llc | Multiple rounds of results summarization for improved latency and relevance |
CN105228033B (zh) * | 2015-08-27 | 2018-11-09 | 联想(北京)有限公司 | 一种视频处理方法及电子设备 |
US10321196B2 (en) * | 2015-12-09 | 2019-06-11 | Rovi Guides, Inc. | Methods and systems for customizing a media asset with feedback on customization |
KR102592904B1 (ko) * | 2016-02-19 | 2023-10-23 | 삼성전자주식회사 | 영상 요약 장치 및 방법 |
WO2017142143A1 (fr) * | 2016-02-19 | 2017-08-24 | Samsung Electronics Co., Ltd. | Procédé et appareil permettant de fournir des informations de résumé d'une vidéo |
DE102018202514A1 (de) * | 2018-02-20 | 2019-08-22 | Bayerische Motoren Werke Aktiengesellschaft | System und Verfahren zur automatischen Erstellung eines Videos einer Fahrt |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040052505A1 (en) | 2002-05-28 | 2004-03-18 | Yesvideo, Inc. | Summarization of a visual recording |
US20050002647A1 (en) * | 2003-07-02 | 2005-01-06 | Fuji Xerox Co., Ltd. | Systems and methods for generating multi-level hypervideo summaries |
Family Cites Families (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3823333B2 (ja) * | 1995-02-21 | 2006-09-20 | 株式会社日立製作所 | 動画像の変化点検出方法、動画像の変化点検出装置、動画像の変化点検出システム |
JP3240871B2 (ja) * | 1995-03-07 | 2001-12-25 | 松下電器産業株式会社 | 映像要約方法 |
JPH10232884A (ja) * | 1996-11-29 | 1998-09-02 | Media Rinku Syst:Kk | 映像ソフトウェア処理方法及び映像ソフトウェア処理装置 |
JP2000285243A (ja) * | 1999-01-29 | 2000-10-13 | Sony Corp | 信号処理方法及び映像音声処理装置 |
JP2001014306A (ja) * | 1999-06-30 | 2001-01-19 | Sony Corp | 電子文書処理方法及び電子文書処理装置並びに電子文書処理プログラムが記録された記録媒体 |
US7016540B1 (en) * | 1999-11-24 | 2006-03-21 | Nec Corporation | Method and system for segmentation, classification, and summarization of video images |
AUPQ535200A0 (en) * | 2000-01-31 | 2000-02-17 | Canon Kabushiki Kaisha | Extracting key frames from a video sequence |
US7042493B2 (en) * | 2000-04-07 | 2006-05-09 | Paolo Prandoni | Automated stroboscoping of video sequences |
US7296231B2 (en) * | 2001-08-09 | 2007-11-13 | Eastman Kodak Company | Video structuring by probabilistic merging of video segments |
US20030117428A1 (en) * | 2001-12-20 | 2003-06-26 | Koninklijke Philips Electronics N.V. | Visual summary of audio-visual program features |
US7333712B2 (en) * | 2002-02-14 | 2008-02-19 | Koninklijke Philips Electronics N.V. | Visual summary for scanning forwards and backwards in video content |
US7184955B2 (en) * | 2002-03-25 | 2007-02-27 | Hewlett-Packard Development Company, L.P. | System and method for indexing videos based on speaker distinction |
JP4067326B2 (ja) * | 2002-03-26 | 2008-03-26 | 富士通株式会社 | 動画像内容表示装置 |
JP2003330941A (ja) * | 2002-05-08 | 2003-11-21 | Olympus Optical Co Ltd | 類似画像分類装置 |
FR2845179B1 (fr) * | 2002-09-27 | 2004-11-05 | Thomson Licensing Sa | Procede de regroupement d'images d'une sequence video |
US7143352B2 (en) * | 2002-11-01 | 2006-11-28 | Mitsubishi Electric Research Laboratories, Inc | Blind summarization of video content |
JP2004187029A (ja) * | 2002-12-04 | 2004-07-02 | Toshiba Corp | 要約映像追いかけ再生装置 |
US20040181545A1 (en) * | 2003-03-10 | 2004-09-16 | Yining Deng | Generating and rendering annotated video files |
US20050257242A1 (en) * | 2003-03-14 | 2005-11-17 | Starz Entertainment Group Llc | Multicast video edit control |
JP4344534B2 (ja) * | 2003-04-30 | 2009-10-14 | セコム株式会社 | 画像処理システム |
KR100590537B1 (ko) * | 2004-02-18 | 2006-06-15 | 삼성전자주식회사 | 복수 영상의 요약 방법 및 장치 |
JP2005277445A (ja) * | 2004-03-22 | 2005-10-06 | Fuji Xerox Co Ltd | 会議映像処理装置、会議映像処理方法およびプログラム |
US7302451B2 (en) * | 2004-05-07 | 2007-11-27 | Mitsubishi Electric Research Laboratories, Inc. | Feature identification of events in multimedia |
JP4140579B2 (ja) * | 2004-08-11 | 2008-08-27 | ソニー株式会社 | 画像処理装置および方法、撮影装置、並びにプログラム |
JP4641450B2 (ja) * | 2005-05-23 | 2011-03-02 | 日本電信電話株式会社 | 非定常画像検出方法、非定常画像検出装置、及び非定常画像検出プログラム |
US7555149B2 (en) * | 2005-10-25 | 2009-06-30 | Mitsubishi Electric Research Laboratories, Inc. | Method and system for segmenting videos using face detection |
-
2008
- 2008-06-09 WO PCT/IB2008/052250 patent/WO2008152556A1/fr active Application Filing
- 2008-06-09 JP JP2010511756A patent/JP2010531561A/ja active Pending
- 2008-06-09 CN CN2008800203066A patent/CN101743596B/zh not_active Expired - Fee Related
- 2008-06-09 US US12/663,529 patent/US20100185628A1/en not_active Abandoned
- 2008-06-09 KR KR1020107000745A patent/KR20100018070A/ko not_active Application Discontinuation
- 2008-06-09 EP EP08763246A patent/EP2156438A1/fr not_active Withdrawn
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20040052505A1 (en) | 2002-05-28 | 2004-03-18 | Yesvideo, Inc. | Summarization of a visual recording |
US20050002647A1 (en) * | 2003-07-02 | 2005-01-06 | Fuji Xerox Co., Ltd. | Systems and methods for generating multi-level hypervideo summaries |
Non-Patent Citations (4)
Title |
---|
ANDREAS GIRGENSOHN, JOHN BORECZKY, LYNN WILCOX: "Keyframe-Based User Interfaces for Digital Video", IEEE COMPUTER MAGAZINE, vol. 34, no. 9, September 2001 (2001-09-01), pages 61 - 67, XP002494967 * |
BENINI, S. ET AL.: "Extraction of Significant Video Summaries by Dendrogram Analysis", 2006 IEEE INT. CONF. ON IMAGE PROCESSING, 11 October 2006 (2006-10-11), pages 133 - 136, XP031048591, DOI: doi:10.1109/ICIP.2006.312377 |
GIRGENSOHN, A. ET AL.: "Keyframe-based User Interfaces for Digital Video", IEEE COMPUTER MAGAZINE, vol. 34, no. 9, September 2001 (2001-09-01), pages 61 - 67, XP002494967, DOI: doi:10.1109/2.947093 |
S. BENINI, A. BIANCHETTI, R. LEONARDI, P. MIGLIORATI: "Extraction of Significant Video Summaries by Dendrogram Analysis", 2006 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, 11 October 2006 (2006-10-11), pages 133 - 136, XP002494966 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2012019305A (ja) * | 2010-07-07 | 2012-01-26 | Nippon Telegr & Teleph Corp <Ntt> | 映像要約装置,映像要約方法および映像要約プログラム |
Also Published As
Publication number | Publication date |
---|---|
KR20100018070A (ko) | 2010-02-16 |
CN101743596B (zh) | 2012-05-30 |
EP2156438A1 (fr) | 2010-02-24 |
JP2010531561A (ja) | 2010-09-24 |
CN101743596A (zh) | 2010-06-16 |
US20100185628A1 (en) | 2010-07-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Li et al. | An overview of video abstraction techniques | |
US7383508B2 (en) | Computer user interface for interacting with video cliplets generated from digital video | |
Truong et al. | Video abstraction: A systematic review and classification | |
US7702185B2 (en) | Use of image similarity in annotating groups of visual images in a collection of visual images | |
RU2440606C2 (ru) | Способ и устройство автоматического генерирования сводки множества изображений | |
US7697792B2 (en) | Process-response statistical modeling of a visual image for use in determining similarity between visual images | |
US8316301B2 (en) | Apparatus, medium, and method segmenting video sequences based on topic | |
US8750681B2 (en) | Electronic apparatus, content recommendation method, and program therefor | |
US20100185628A1 (en) | Method and apparatus for automatically generating summaries of a multimedia file | |
US20060020597A1 (en) | Use of image similarity in summarizing a collection of visual images | |
EP1557837A1 (fr) | Elimination des redondances dans un système d'aperçu vidéo opérant de manière adaptative selon le contenu | |
US20060015495A1 (en) | Use of image similarity in image searching via a network of computational apparatus | |
US20060015497A1 (en) | Content-based indexing or grouping of visual images, with particular use of image similarity to effect same | |
US20060015494A1 (en) | Use of image similarity in selecting a representative visual image for a group of visual images | |
Chen et al. | Tiling slideshow | |
EP2530605A1 (fr) | Dispositif de traitement de données | |
Jiang et al. | Automatic consumer video summarization by audio and visual analysis | |
WO2002082328A2 (fr) | Meta-donnees de camera pour classification de contenu | |
KR20070118635A (ko) | 오디오 및/또는 비주얼 데이터의 서머라이제이션 | |
Dimitrova et al. | Video keyframe extraction and filtering: a keyframe is not a keyframe to everyone | |
JP2009123095A (ja) | 映像解析装置及び映像解析方法 | |
EP1527453A1 (fr) | Procede, systeme et produit de programme pour obtenir une table de contenus fondee sur le contenu | |
El-Bendary et al. | PCA-based home videos annotation system | |
Fersini et al. | Multimedia summarization in law courts: a clustering-based environment for browsing and consulting judicial folders | |
Adami et al. | The ToCAI description scheme for indexing and retrieval of multimedia documents |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200880020306.6 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 08763246 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008763246 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010511756 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12663529 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 119/CHENP/2010 Country of ref document: IN |
|
ENP | Entry into the national phase |
Ref document number: 20107000745 Country of ref document: KR Kind code of ref document: A |