WO2012108090A1 - 映像間対応関係表示システム及び映像間対応関係表示方法 - Google Patents
映像間対応関係表示システム及び映像間対応関係表示方法 Download PDFInfo
- Publication number
- WO2012108090A1 WO2012108090A1 PCT/JP2011/077379 JP2011077379W WO2012108090A1 WO 2012108090 A1 WO2012108090 A1 WO 2012108090A1 JP 2011077379 W JP2011077379 W JP 2011077379W WO 2012108090 A1 WO2012108090 A1 WO 2012108090A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- correspondence
- section
- inter
- derived
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims description 22
- 230000004048 modification Effects 0.000 claims abstract description 52
- 238000012986 modification Methods 0.000 claims abstract description 52
- 230000004075 alteration Effects 0.000 claims description 12
- 230000004044 response Effects 0.000 claims description 2
- 239000013598 vector Substances 0.000 description 28
- 238000010586 diagram Methods 0.000 description 10
- 238000000605 extraction Methods 0.000 description 7
- 230000008569 process Effects 0.000 description 5
- 230000006870 function Effects 0.000 description 3
- 239000000284 extract Substances 0.000 description 2
- 230000002123 temporal effect Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 239000000463 material Substances 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/34—Indicating arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
Definitions
- the present invention relates to an inter-video correspondence display system and an inter-video correspondence display method.
- an edited video is created by editing a coverage video generated by the coverage.
- a plurality of edited videos are used when creating a broadcast video that is actually used when a program is broadcast.
- the coverage video is not only used for one edited video, but may be variously edited as necessary to create a plurality of edited videos.
- the edited video is not only used for one broadcast video but also used for a plurality of broadcast videos.
- Patent Document 1 a method for displaying the correspondence between videos by registering the correspondence between material video, edited video, and on-air video in a database has been proposed. ing. In this method, when a video is created, the correspondence between the video and another video is registered in a database. Therefore, it is assumed that the correspondence relationship is registered in the database for the created video.
- Patent Literature 2 discloses a technique for searching for content similar to a certain content by using content feature information.
- Patent Document 3 discloses a method of determining content identity based on content feature amounts and grouping the content.
- Patent Document 4 discloses a technique that can determine the identity between videos with high discriminating power even when there is such a modification.
- Patent Document 1 Although a correspondence relationship between videos is displayed, when a derived video is created using a part of the original video, each section of the derived video It is not possible to easily grasp whether there is any alteration from the original video.
- a broadcast video may be created by modifying the edited video, such as adding a telop, but it is easy to know which section of the broadcast video has been modified from the edited video. Can not do it. Therefore, for example, in order to confirm what kind of alteration has been made from the edited video in the broadcast video, it is necessary to reproduce the entire section of the broadcast video while comparing it with the edited video.
- Patent Documents 2 to 4 disclose a method for determining the similarity and identity between videos, but a method for easily grasping whether or not there is a modification from the original video. Is not disclosed at all.
- the present invention has been made in view of such circumstances, and an object of the present invention is to make it possible to easily grasp whether or not there is a modification from the original video in each section of the derived video created using the original video. .
- An inter-video correspondence display system provides a correspondence between sections between at least one original video and a derived video created using at least a section of at least one original video.
- a relation information storage unit for storing relation information including information indicating a relation and information indicating the presence or absence of alteration from the original video in a section having a correspondence relation of the derived video; and relation information stored in the relation information storage unit
- the correspondence display section for displaying the correspondence of the section between the derived video and the at least one original video in a display mode according to the presence or absence of alteration from the original video in the section having the correspondence of the derived video And comprising.
- the “unit” does not simply mean a physical means, but includes a case where the function of the “unit” is realized by software. Also, even if the functions of one “unit” or device are realized by two or more physical means or devices, the functions of two or more “units” or devices are realized by one physical means or device. May be.
- FIG. 1 is a diagram showing a configuration of an inter-video correspondence presentation system that is an embodiment of the present invention.
- the inter-video correspondence presentation system 10 is a system that presents the presence / absence of modification in each section of a derived video together with the correspondence between videos between videos having a correspondence in the creation process.
- An amount storage unit 22, a correspondence relationship determination unit 24, a relationship information storage unit 26, and a correspondence relationship display unit 28 are included.
- the feature amount storage unit 22 includes a coverage video feature amount storage unit 30, an edited video feature amount storage unit 32, and a broadcast video feature amount storage unit 34.
- the inter-video correspondence presentation system 10 refers to the video database 40.
- the video database 40 includes a coverage video database (DB) 42, an edited video database (DB) 44, and a broadcast video database (DB) 46.
- DB coverage video database
- DB broadcast video database
- Such an inter-video correspondence presentation system 10 is configured using one or a plurality of information processing apparatuses, and the feature amount extraction unit 20, the correspondence judgment unit 24, and the correspondence display 28 are stored in a memory.
- the program can be realized by the processor executing the program.
- the feature amount storage unit 22 and the relationship information storage unit 26 can be realized using a storage area such as a memory or a storage device.
- the video having a corresponding relationship in the creation process includes, for example, a coverage video used in a broadcasting station, an edited video created using at least a part of the coverage video, and at least a part of the edited video. It is a broadcast video created by using it.
- the broadcast video is a derived video
- the coverage video and the edited video are original videos.
- the video having a corresponding relationship in the creation process is not limited to the video used in the broadcasting station.
- the video on the broadcast video or the recording medium is the original video and the illegal video is a derived video.
- the description will focus on video used in a broadcasting station, but the correspondence presentation system shown in the present embodiment can be applied to any video as long as it has a correspondence in the creation process. It is.
- FIG. 2 is a diagram illustrating a configuration example of the coverage video DB 42.
- video data of the coverage video is stored in association with the video identifier and the attribute information.
- the video identifier is information for identifying the coverage video, and for example, a file name or the like can be used.
- the attribute information is information indicating various attributes relating to the coverage video, and includes, for example, a video title, creation date and time, time of the entire video, and the like.
- FIG. 3 is a diagram illustrating a configuration example of the edited video DB 44.
- the edited video DB 44 stores the video data of the edited video in association with the video identifier and attribute information.
- the video identifier and attribute information are the same as those of the coverage video DB 42.
- FIG. 4 is a diagram illustrating a configuration example of the broadcast video DB 46.
- the broadcast video DB 46 video data of the broadcast video is stored in association with the video identifier, attribute information, and frequency information.
- the video identifier and attribute information are the same as those of the coverage video DB 42.
- information indicating the number of times the broadcast video has been broadcast is set in the frequency information.
- the value of frequency information can be increased each time rebroadcasting is performed.
- the rebroadcast broadcast video may be stored in the broadcast video DB 46 as a video different from the original broadcast video.
- the attribute information can include information regarding the broadcast date and time. If broadcast video is distributed in response to a viewer's request, the frequency of viewing the broadcast video may be set in the frequency information.
- the coverage video, the edited video, and the broadcast video are stored in separate databases.
- these videos may be stored together in one database.
- information indicating the type of video may be given to each video.
- the feature quantity extraction unit 20 refers to the video database 40, extracts a feature quantity vector from each of a plurality of frame images constituting the video, and stores the feature quantity vector in the feature quantity storage unit 22.
- the feature vectors of the coverage video, the edited video, and the broadcast video are stored in the coverage video feature amount storage unit 30, the edited video feature amount storage unit 32, and the broadcast video feature amount storage unit 34, respectively.
- the feature vectors of the coverage video, the edited video, and the broadcast video may be stored together in one storage unit. In this case, information indicating the type of video may be given to each feature vector.
- the feature amount vector indicates the feature amount for each predetermined section constituting the video.
- a feature amount vector is generated for each frame image constituting a video by a method described in International Publication No. 2010/0884714.
- the feature quantity vector generated by this method is a set of N feature quantities corresponding to N (N ⁇ 2) partial areas defined in the frame image.
- Each partial area corresponding to each dimension of the feature quantity vector includes, for example, a plurality of partial areas in the frame image.
- the feature-value of each dimension can be produced
- FIG. 5 is a diagram showing an example of area division of a frame image. As shown in FIG. 5, for example, each frame image can be divided into 32 ⁇ 32 1024 regions (divided regions). The partial area corresponding to each dimension in the feature vector is constituted by a combination of one or more divided areas.
- FIG. 6 is a diagram showing an image when extracting feature values in the Mth dimension.
- the partial areas corresponding to the Mth dimension are two partial areas 62 and 64.
- the feature quantity extraction unit 20 may generate an M-th dimension feature quantity based on the difference between the feature quantity (area feature quantity) of the partial area 62 and the feature quantity (area feature quantity) of the partial area 64. it can.
- the feature amount of each of the partial areas 62 and 64 can be calculated using an arbitrary method such as an average value or a median value of pixel values in each partial area.
- the feature quantity extraction unit 20 can generate the Mth dimension feature quantity by quantizing the difference between the area feature quantities of the partial areas 62 and 64 into, for example, three values ( ⁇ 1, 0, 1). .
- the feature quantity extraction unit 20 can generate an N-dimensional feature quantity vector by generating feature quantities for each dimension (first dimension to Nth dimension). Note that the feature value calculation method for each dimension described here is an example, and a feature value vector can be generated by any method.
- FIG. 7 is a diagram illustrating an example of a feature amount vector stored in the coverage video feature amount storage unit 30.
- the feature vector is stored in association with a video identifier and sequence information indicating the temporal order of frame images.
- the sequence information may be any information that can grasp the order of the feature vector, and for example, a frame number or the like can be used. Note that the sequence information may not be provided as long as the temporal order of the feature vector can be identified by the data storage structure or the like.
- the edited video feature quantity storage unit 32 and the broadcast video feature quantity storage unit 34 can also be configured in the same manner as the coverage video feature quantity storage unit 30.
- the correspondence determination unit 24 determines the correspondence between videos based on the feature vector stored in the feature storage 22. Specifically, the correspondence between the coverage video (original video) and the edited video (derived video) and the correspondence between the edited video (original video) and the broadcast video (derived video) are determined. Thereby, the correspondence between the coverage video (original video) and the broadcast video (derived video) is also indirectly determined via the edited video. The correspondence determining unit 24 can also directly determine the correspondence between the coverage video and the broadcast video without using the edited video.
- FIG. 8 is a diagram illustrating a configuration example of the correspondence determination unit 24.
- the correspondence determination unit 24 can include a feature amount comparison unit 70, a frame selection unit 72, and a relationship information output unit 74.
- the feature quantity comparison unit 70 compares the feature quantity vector of the video that is the candidate for the original video and the feature quantity vector of the video that is the candidate for the derived video, for example, for each frame. For example, the feature amount comparison unit 70 compares the feature amount vector of the coverage video with the feature amount vector of the edited video. In addition, the feature amount comparison unit 70 compares the feature amount vector of the edited video with the feature amount vector of the broadcast video.
- the frame selection unit 72 selects a frame image determined to have a correspondence (identity) between videos based on the comparison result in the feature amount comparison unit 70. For determining the identity between frame images, for example, the number of dimensions in which feature quantities match between two feature quantity vectors, the number of dimensions in which feature quantities do not match, and the size of feature quantity vectors are compared. Can be done.
- the relationship information output unit 74 outputs relationship information that is information indicating a section having a correspondence relationship between videos. Specifically, the relationship information output unit 74 generates relationship information indicating a section in which the frame selection unit 72 determines that the frame images are identical, and stores the relationship information in the relationship information storage unit 26. Further, the relationship information output unit 74 outputs information indicating the degree of modification in a section having a correspondence relationship between videos based on the comparison result in the feature amount comparison unit 70.
- the degree of modification indicates the degree of modification made by adding a telop or the like when a derived image is created using the original image, and is the same as, for example, the above-described determination of identity between frame images. It can obtain by the method of.
- FIG. 9 is a diagram illustrating an example of relationship information stored in the relationship information storage unit 26.
- the relationship information includes information about the original video, information about the derived video, and information indicating the degree of modification.
- the information related to the original video or the derived video includes a video type, a video identifier, and section information.
- the video type is information indicating whether it is a coverage video, an edited video, or a broadcast video.
- the section information is information indicating a video section. For example, the playback time or frame number of the section in the video can be used.
- FIG. 9 shows that the broadcast video B1 is created using the edited videos E1 to E3. More specifically, the section “15:00 to 33:00” of the edited video E1 is used as the section “00:00 to 15:00” of the broadcast video B1, and the modification degree is “5”. It has become. That is, it can be seen that the section “00:00 to 15:00” of the broadcast video B1 is a modification of the section “15:00 to 33:00” of the edited video E1. Similarly, the section “15:00 to 20:00” of the broadcast video B1 is a modification of the section “05:00 to 10:00” of the edited video E2, and “20” of the broadcast video B1.
- the section from 3:00 to 33:00 is a modification of the section from “03:00 to 13:00” of the edited video E3.
- FIG. 9 only one value is shown as the degree of modification in the section with the corresponding relationship.
- a plurality of sections (divided sections) included in this section are shown. ) May be set for the degree of modification.
- the section “00:00 to 40:00” of the edited video E1 is created from the section “10:00 to 50:00” of the coverage video C1, and the modification degree is “0”. It has become. That is, it can be seen that the edited video E1 is an extracted portion of “10:00 to 50:00” of the coverage video C1. Similarly, the edited videos E2 and E3 are obtained by extracting the section “02:00 to 17:00” of the coverage video C2 and the section “25:00 to 55:00” of the coverage video C2, respectively. Recognize.
- the correspondence relationship display unit 28 converts the correspondence relationship between the derived video and the original video to the degree of modification in each section of the derived video based on the relationship information stored in the relationship information storage unit 26. Display in the corresponding display mode.
- FIG. 10 is a diagram illustrating an example of a screen output by the correspondence display unit 28.
- the correspondence display unit 28 can display the correspondence between videos based on the relationship information stored in the relationship information storage unit 26.
- the screen 80 includes areas 82 to 94.
- the correspondence display 28 can display information such as the video identifier, title, and creation date / time of the broadcast video in the area 82 by referring to the broadcast video DB 46.
- the correspondence display unit 28 may display information on all broadcast videos stored in the broadcast video DB 46 in the area 82, or the broadcast video in which the relationship information is stored in the relationship information storage unit 26. Information regarding only the broadcast video that has been modified may be displayed in the area 82. In the area 82, one broadcast video can be selected by a user operation such as clicking.
- the correspondence display 28 can display information related to the broadcast video selected in the area 82 in another area.
- a representative thumbnail image of the selected broadcast video is displayed.
- a plurality of thumbnail images of the broadcast video are displayed.
- the correspondence display unit 28 can display a thumbnail image of a broadcast video by referring to the broadcast video DB 46, for example.
- the timeline of the broadcast video is displayed, and in the area 90, the correspondence relationship between the broadcast video and the edited video is displayed in accordance with the time axis of the broadcast video timeline.
- the timeline (00:00 to 33:00) of the broadcast video B1 is shown in the area 88, and the section “00:00 to 15 of the broadcast video B1 is shown in the area 90.
- 0:00 ” is created from the edited video E1
- the section“ 15:00 to 20:00 ”of the broadcast video B1 is created from the edited video E2
- the section“ 20:00 ”of the broadcast video B1 is created.
- “-30: 00” is created from the edited video E3.
- the correspondence display unit 28 displays the correspondence between the broadcast video and the edited video in the area 90
- the correspondence display unit 28 can also display the section of each edited video based on the relationship information. For example, it can be displayed that the section of the edited video E1 is “15:00 to 33:00”. The same applies to the edited videos E2 and E3.
- thumbnail images of edited videos in the section corresponding to the broadcast video are displayed.
- the correspondence display unit 28 can display a thumbnail image of the edited video by referring to the edited video DB 44, for example.
- the correspondence relationship display unit 28 can grasp the correspondence relationship between the edited video and the coverage video based on the relationship information stored in the relationship information storage unit 26. Therefore, the correspondence display unit 28 can display the thumbnail image of the coverage video corresponding to the edited video in the region 92 by referring to the coverage video DB 42.
- the degree of modification from the edited video in each section of the broadcast video is displayed in a graph in accordance with the timeline of the broadcast video. From this graph, for example, it can be seen that the degree of modification in the range surrounded by the circle 96 is high. Then, referring to the timeline of the broadcast video, it can be seen that the degree of modification in the vicinity of the broadcast video section “00:00 to 05:00” is high. Furthermore, by referring to the thumbnail image in the area 92, it is possible to grasp what kind of scene is the section with a high degree of modification.
- the correspondence display unit 28 displays the correspondence in the region 90 and displays the degree of modification in the region 94, so that the correspondence between the sections between the broadcast video and the edited video can be obtained for both regions as a whole.
- the display mode is displayed in accordance with the degree of modification in the section having the corresponding relationship.
- the display mode according to the degree of modification is not limited to the display of the degree of modification as shown in FIG. 10 as long as the degree of modification or the presence or absence of modification can be identified.
- the modified section may be highlighted.
- the strength or color of the highlight may be changed or the size of the thumbnail image may be changed according to the degree of modification.
- the correspondence display unit 28 can output a screen that displays the broadcast video and the edited video of the selected section.
- FIG. 11 shows an example.
- the screen 120 includes an area 122 where a broadcast video is displayed and an area 123 where an edited video or a coverage video is displayed.
- the correspondence relationship display unit 28 acquires the section information of the selected section from the relationship information, acquires the video in the section from the video database 40, and displays it in the regions 122 and 123.
- the correspondence display unit 28 may display an image of a section selected in the region 86 or the region 94 instead of the region 92.
- FIG. 12 is a flowchart showing an example of processing in the inter-video correspondence presentation system 10.
- the feature quantity extraction unit 20 extracts a feature quantity vector for each video stored in the video database 40 and stores it in the feature quantity storage unit 22 (S1201).
- the feature quantity comparison unit 70 refers to the feature quantity storage unit 22 and compares the feature quantity vectors between the videos (S1202).
- the frame selection unit 72 selects a frame image having a correspondence relationship (identity) based on the comparison result in the feature amount comparison unit 70 (S1203).
- the relationship information output unit 74 stores the relationship information about the section including the selected frame image in the relationship information storage unit 26 (S1204).
- the correspondence relationship display unit 28 displays the correspondence relationship of the sections between the derived video (broadcast video) and the original video (edited video or coverage video) based on the relationship information stored in the relationship information storage unit 26. Then, it is displayed in a display mode corresponding to the degree of modification from the original video in each section of the derived video (S1205).
- the section correspondence between the derived video and the original video is displayed in a display mode according to the modification degree in each section of the derived video.
- the presence or absence of alteration in each section of the video can be easily grasped.
- this embodiment is for making an understanding of this invention easy, and is not for limiting and interpreting this invention.
- the present invention can be changed / improved without departing from the spirit thereof, and the present invention includes equivalents thereof.
- the correspondence display system (Supplementary note 2) The correspondence display system according to supplementary note 1, wherein the information indicating the presence / absence of modification in the section having the correspondence relation is information indicating the presence / absence of modification in each of the plurality of divided sections included in the section.
- the correspondence display unit displays the correspondence of the sections in a display mode according to the presence or absence of alteration in each of the plurality of divided sections included in the section based on the relationship information.
- Correspondence display system. (Supplementary note 3) The inter-video correspondence table system according to supplementary note 1 or 2, wherein the correspondence display unit displays the presence or absence of the modification in accordance with the time axis of the derived video. system.
- the inter-video correspondence display system according to any one of supplementary notes 1 to 3, wherein the correspondence display unit displays the correspondence for the at least one original video as the derived video.
- Display system for inter-video correspondences that displays along the time axis.
- the information indicating presence / absence of modification in the section includes information indicating the degree of modification in the section
- the correspondence display unit includes: The inter-video correspondence display system for displaying the correspondence of the sections in a display mode corresponding to the degree of modification in each section based on the relation information.
- (Supplementary note 6) The inter-video correspondence display system according to any one of supplementary notes 1 to 5, wherein the correspondence display unit includes a storage unit that stores thumbnail images for each section of the original video.
- An inter-video correspondence display system that refers to and displays the correspondence in a display mode corresponding to the presence or absence of the alteration, and displays a thumbnail image of the original video in the section having the correspondence.
- (Supplementary note 7) The inter-video correspondence display system according to any one of supplementary notes 1 to 6, wherein the relation information storage unit stores the relation information for a plurality of the derived videos, and displays the correspondence relation display.
- the unit displays information for selecting one of the plurality of derived videos, and the correspondence relationship of the one derived video selected by a user input in accordance with the display depends on whether or not the modification is performed.
- Display system for displaying correspondence between images. (Supplementary note 8) The inter-video correspondence display system according to any one of supplementary notes 1 to 7, wherein the correspondence relation display unit selects one of the sections in which the correspondence is displayed.
- the inter-video correspondence display system according to any one of supplementary notes 1 to 8, wherein the feature quantity storage unit stores the feature quantity for each predetermined section constituting the video in association with each video. And a feature amount comparison unit that compares feature amounts for each section of a plurality of videos including the derived image and the original image, and outputs the relation information based on a comparison result of the feature amounts for each section.
- An inter-video correspondence display system further comprising: a relation information output unit stored in the relation information storage unit.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
(付記1)少なくとも1つの元映像と、前記少なくとも1つの元映像の少なくとも一部の区間を利用して作成された派生映像との間における、区間の対応関係を示す情報と、前記派生映像の前記対応関係のある区間における前記元映像からの改変の有無を示す情報とを含む関係情報を記憶する関係情報記憶部と、前記関係情報記憶部に記憶されている前記関係情報に基づいて、前記派生映像と前記少なくとも1つの元映像との間における前記区間の対応関係を、前記派生映像の前記対応関係のある区間における前記元映像からの改変の有無に応じた表示態様で表示する対応関係表示部と、を備える映像間対応関係表示システム。
(付記2)付記1に記載の対応関係表示システムであって、前記対応関係のある区間における改変の有無を示す情報は、該区間に含まれる複数の分割区間のそれぞれにおける改変の有無を示す情報を含み、前記対応関係表示部は、前記関係情報に基づいて、前記区間の対応関係を、該区間に含まれる複数の分割区間のそれぞれにおける改変の有無に応じた表示態様で表示する、映像間対応関係表示システム。
(付記3)付記1または2に記載の映像間対応関係表システムであって、前記対応関係表示部は、前記改変の有無を前記派生映像の時間軸に合わせて表示する、映像間対応関係表示システム。
(付記4)付記1~3の何れか一項に記載の映像間対応関係表示システムであって、前記対応関係表示部は、前記少なくとも1つの元映像のそれぞれについて、前記対応関係を前記派生映像の時間軸に合わせて表示する、映像間対応関係表示システム。
(付記5)付記1~4に記載の映像間対応関係表示システムであって、前記区間における改変の有無を示す情報は、該区間における改変の度合いを示す情報を含み、前記対応関係表示部は、前記関係情報に基づいて、前記区間の対応関係を、各区間における改変の度合いに応じた表示態様で表示する、映像間対応関係表示システム。
(付記6)付記1~5の何れか一項に記載の映像間対応関係表示システムであって、前記対応関係表示部は、前記元映像の区間ごとのサムネイル画像が記憶されている記憶部を参照し、前記対応関係を前記改変の有無に応じた表示態様で表示するともに、前記対応関係のある区間における前記元映像のサムネイル画像を表示する、映像間対応関係表示システム。
(付記7)付記1~6の何れか一項に記載の映像間対応関係表示システムであって、前記関係情報記憶部は、複数の前記派生映像に対する前記関係情報を記憶し、前記対応関係表示部は、前記複数の派生映像のうちの1つを選択するための情報を表示し、該表示に応じてユーザ入力により選択される1つの派生映像について、前記対応関係を前記改変の有無に応じた表示態様で表示する、映像間対応関係表示システム。
(付記8)付記1~7の何れか一項に記載の映像間対応関係表示システムであって、前記対応関係表示部は、前記対応関係が表示された区間のうちの1つの区間を選択するユーザ入力に応じて、前記派生映像及び元映像が記憶されている記憶部を参照し、該選択された区間における前記派生映像及び前記元映像を表示する、映像間対応関係表示システム。
(付記9)付記1~8の何れか一項に記載の映像間対応関係表示システムであって、映像を構成する所定の区間ごとの特徴量を映像ごとに対応づけて記憶する特徴量記憶部と、前記派生映像及び前記元映像を含む、複数の映像に対する前記区間ごとの特徴量を比較する特徴量比較部と、前記区間ごとの特徴量の比較結果に基づいて前記関係情報を出力して前記関係情報記憶部に格納する関係情報出力部と、をさらに備える映像間対応関係表示システム。
(付記10)少なくとも1つの元映像と、前記少なくとも1つの元映像の少なくとも一部の区間を利用して作成された派生映像との間における、区間の対応関係を示す情報と、前記派生映像の前記対応関係のある区間における前記元映像からの改変の有無を示す情報とを含む関係情報を関係情報記憶部に記憶し、前記関係情報記憶部に記憶されている前記関係情報に基づいて、前記派生映像と前記少なくとも1つの元映像との間における前記区間の対応関係を、前記派生映像の前記対応関係のある区間における前記元映像からの改変の有無に応じた表示態様で表示する、映像間対応関係表示方法。
20 特徴量抽出部
22 特徴量記憶部
24 対応関係判定部
26 関係情報記憶部
28 対応関係表示部
30 取材映像特徴量記憶部
32 編集映像特徴量記憶部
34 放送映像特徴量記憶部
40 映像データベース
42 取材映像データベース
44 編集映像データベース
46 放送映像データベース
70 特徴量比較部
72 フレーム選択部
74 関係情報出力部
Claims (10)
- 少なくとも1つの元映像と、前記少なくとも1つの元映像の少なくとも一部の区間を利用して作成された派生映像との間における、区間の対応関係を示す情報と、前記派生映像の前記対応関係のある区間における前記元映像からの改変の有無を示す情報とを含む関係情報を記憶する関係情報記憶部と、
前記関係情報記憶部に記憶されている前記関係情報に基づいて、前記派生映像と前記少なくとも1つの元映像との間における前記区間の対応関係を、前記派生映像の前記対応関係のある区間における前記元映像からの改変の有無に応じた表示態様で表示する対応関係表示部と、
を備える映像間対応関係表示システム。 - 請求項1に記載の対応関係表示システムであって、
前記対応関係のある区間における改変の有無を示す情報は、該区間に含まれる複数の分割区間のそれぞれにおける改変の有無を示す情報を含み、
前記対応関係表示部は、前記関係情報に基づいて、前記区間の対応関係を、該区間に含まれる複数の分割区間のそれぞれにおける改変の有無に応じた表示態様で表示する、
映像間対応関係表示システム。 - 請求項1または2に記載の映像間対応関係表システムであって、
前記対応関係表示部は、前記改変の有無を前記派生映像の時間軸に合わせて表示する、
映像間対応関係表示システム。 - 請求項1~3の何れか一項に記載の映像間対応関係表示システムであって、
前記対応関係表示部は、前記少なくとも1つの元映像のそれぞれについて、前記対応関係を前記派生映像の時間軸に合わせて表示する、
映像間対応関係表示システム。 - 請求項1~4に記載の映像間対応関係表示システムであって、
前記区間における改変の有無を示す情報は、該区間における改変の度合いを示す情報を含み、
前記対応関係表示部は、前記関係情報に基づいて、前記区間の対応関係を、各区間における改変の度合いに応じた表示態様で表示する、
映像間対応関係表示システム。 - 請求項1~5の何れか一項に記載の映像間対応関係表示システムであって、
前記対応関係表示部は、前記元映像の区間ごとのサムネイル画像が記憶されている記憶部を参照し、前記対応関係を前記改変の有無に応じた表示態様で表示するともに、前記対応関係のある区間における前記元映像のサムネイル画像を表示する、
映像間対応関係表示システム。 - 請求項1~6の何れか一項に記載の映像間対応関係表示システムであって、
前記関係情報記憶部は、複数の前記派生映像に対する前記関係情報を記憶し、
前記対応関係表示部は、
前記複数の派生映像のうちの1つを選択するための情報を表示し、
該表示に応じてユーザ入力により選択される1つの派生映像について、前記対応関係を前記改変の有無に応じた表示態様で表示する、
映像間対応関係表示システム。 - 請求項1~7の何れか一項に記載の映像間対応関係表示システムであって、
前記対応関係表示部は、前記対応関係が表示された区間のうちの1つの区間を選択するユーザ入力に応じて、前記派生映像及び元映像が記憶されている記憶部を参照し、該選択された区間における前記派生映像及び前記元映像を表示する、
映像間対応関係表示システム。 - 請求項1~8の何れか一項に記載の映像間対応関係表示システムであって、
映像を構成する所定の区間ごとの特徴量を映像ごとに対応づけて記憶する特徴量記憶部と、
前記派生映像及び前記元映像を含む、複数の映像に対する前記区間ごとの特徴量を比較する特徴量比較部と、
前記区間ごとの特徴量の比較結果に基づいて前記関係情報を出力して前記関係情報記憶部に格納する関係情報出力部と、
をさらに備える映像間対応関係表示システム。 - 少なくとも1つの元映像と、前記少なくとも1つの元映像の少なくとも一部の区間を利用して作成された派生映像との間における、区間の対応関係を示す情報と、前記派生映像の前記対応関係のある区間における前記元映像からの改変の有無を示す情報とを含む関係情報を関係情報記憶部に記憶し、
前記関係情報記憶部に記憶されている前記関係情報に基づいて、前記派生映像と前記少なくとも1つの元映像との間における前記区間の対応関係を、前記派生映像の前記対応関係のある区間における前記元映像からの改変の有無に応じた表示態様で表示する、
映像間対応関係表示方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/983,957 US9473734B2 (en) | 2011-02-10 | 2011-11-28 | Inter-video corresponding relationship display system and inter-video corresponding relationship display method |
JP2012556756A JP5854232B2 (ja) | 2011-02-10 | 2011-11-28 | 映像間対応関係表示システム及び映像間対応関係表示方法 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2011027192 | 2011-02-10 | ||
JP2011-027192 | 2011-02-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012108090A1 true WO2012108090A1 (ja) | 2012-08-16 |
Family
ID=46638319
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/077379 WO2012108090A1 (ja) | 2011-02-10 | 2011-11-28 | 映像間対応関係表示システム及び映像間対応関係表示方法 |
Country Status (3)
Country | Link |
---|---|
US (1) | US9473734B2 (ja) |
JP (1) | JP5854232B2 (ja) |
WO (1) | WO2012108090A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018156344A (ja) * | 2017-03-17 | 2018-10-04 | 日本放送協会 | 映像ストリームの一致判定プログラム |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9792528B2 (en) * | 2012-01-30 | 2017-10-17 | Nec Corporation | Information processing system, information processing method, information processing apparatus and control method and control program thereof, and communication terminal and control method and control program thereof |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10149673A (ja) * | 1996-09-20 | 1998-06-02 | Sony Corp | 編集システム、編集方法及び表示装置 |
JPH10224665A (ja) * | 1997-02-04 | 1998-08-21 | Sony Corp | 送出システム |
JPH11266422A (ja) * | 1998-03-18 | 1999-09-28 | Toshiba Corp | 放送番組管理装置、放送番組管理方法、及び放送番組管理処理プログラムを記録した記録媒体 |
JP2004172787A (ja) * | 2002-11-19 | 2004-06-17 | Nec Corp | 映像閲覧システムおよび方法 |
JP2005203881A (ja) * | 2004-01-13 | 2005-07-28 | Sony Corp | 情報処理装置および方法、並びにプログラム |
WO2010073695A1 (ja) * | 2008-12-25 | 2010-07-01 | 日本電気株式会社 | 編集情報提示装置、編集情報提示方法、プログラム、及び記録媒体 |
JP2010282376A (ja) * | 2009-06-03 | 2010-12-16 | Nippon Telegr & Teleph Corp <Ntt> | メタデータ管理装置、メタデータ管理方法、及び、メタデータ管理プログラム |
Family Cites Families (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
DE69222102T2 (de) * | 1991-08-02 | 1998-03-26 | Grass Valley Group | Bedienerschnittstelle für Videoschnittsystem zur Anzeige und interaktive Steuerung von Videomaterial |
DE69322470T2 (de) * | 1992-08-12 | 1999-07-15 | Ibm | System und Verfahren zur Lokalisierung von Videosegmentübergängen |
DE69428180T2 (de) * | 1993-12-13 | 2002-05-02 | Sony Corp | Schnittsysteme und -verfahren |
US5600375A (en) * | 1994-09-08 | 1997-02-04 | Intel Corporation | Rendering an inter verses intra video encoding decision based upon a vertical gradient measure of target video frames |
GB2354104A (en) * | 1999-09-08 | 2001-03-14 | Sony Uk Ltd | An editing method and system |
CN1273989C (zh) * | 1999-11-10 | 2006-09-06 | 汤姆森特许公司 | 在dvd录像机中实时产生菜单的方法 |
JP2001157183A (ja) * | 1999-11-25 | 2001-06-08 | Nec Corp | テレビ電話システム |
WO2002008948A2 (en) * | 2000-07-24 | 2002-01-31 | Vivcom, Inc. | System and method for indexing, searching, identifying, and editing portions of electronic multimedia files |
JP4228662B2 (ja) | 2002-11-19 | 2009-02-25 | 日本電気株式会社 | 映像閲覧システムおよび方法 |
JP2005148801A (ja) | 2003-11-11 | 2005-06-09 | Nippon Telegr & Teleph Corp <Ntt> | 関連情報提供方法、関連情報提供装置、関連情報提供プログラム、および、関連情報提供プログラムの記録媒体 |
US7925669B2 (en) * | 2004-10-13 | 2011-04-12 | Sony Corporation | Method and apparatus for audio/video attribute and relationship storage and retrieval for efficient composition |
US8135750B2 (en) * | 2005-04-22 | 2012-03-13 | Microsoft Corporation | Efficiently describing relationships between resources |
JP2007164633A (ja) | 2005-12-15 | 2007-06-28 | Nippon Telegr & Teleph Corp <Ntt> | コンテンツ検索方法及び装置及びプログラム |
JP4270240B2 (ja) * | 2006-08-14 | 2009-05-27 | コニカミノルタビジネステクノロジーズ株式会社 | 画像表示装置、画像表示方法、および画像表示プログラム |
US8364020B2 (en) * | 2007-09-28 | 2013-01-29 | Motorola Mobility Llc | Solution for capturing and presenting user-created textual annotations synchronously while playing a video recording |
US8797233B2 (en) * | 2008-08-20 | 2014-08-05 | The Regents Of The University Of California | Systems, methods, and devices for dynamic management of data streams updating displays |
US20100162286A1 (en) * | 2008-11-24 | 2010-06-24 | Digitalsmiths Corporation | Systems and methods for analyzing trends in video consumption based on embedded video metadata |
CN102292745B (zh) | 2009-01-23 | 2015-08-12 | 日本电气株式会社 | 图像签名提取设备 |
JP4857453B2 (ja) * | 2009-07-01 | 2012-01-18 | テクトロニクス・インターナショナル・セールス・ゲーエムベーハー | パケットとシリアル・データの対応関係確認方法 |
US8743178B2 (en) * | 2010-01-05 | 2014-06-03 | Dolby Laboratories Licensing Corporation | Multi-view video format control |
RU2534958C2 (ru) * | 2010-03-04 | 2014-12-10 | ТиЭйчИкс ЛТД. | Электронный адаптер для селективной модификации аудио- или видеоданных для использования с выходным устройством |
US8606073B2 (en) * | 2010-05-12 | 2013-12-10 | Woodman Labs, Inc. | Broadcast management system |
US20120002884A1 (en) * | 2010-06-30 | 2012-01-05 | Alcatel-Lucent Usa Inc. | Method and apparatus for managing video content |
KR101110202B1 (ko) * | 2010-08-02 | 2012-02-16 | (주)엔써즈 | 동영상 데이터들의 상호 관계에 기초한 데이터베이스 형성 방법 및 데이터베이스 형성 시스템 |
US8798400B2 (en) * | 2010-10-21 | 2014-08-05 | International Business Machines Corporation | Using near-duplicate video frames to analyze, classify, track, and visualize evolution and fitness of videos |
US20130031479A1 (en) * | 2011-07-25 | 2013-01-31 | Flowers Harriett T | Web-based video navigation, editing and augmenting apparatus, system and method |
-
2011
- 2011-11-28 US US13/983,957 patent/US9473734B2/en active Active
- 2011-11-28 JP JP2012556756A patent/JP5854232B2/ja active Active
- 2011-11-28 WO PCT/JP2011/077379 patent/WO2012108090A1/ja active Application Filing
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH10149673A (ja) * | 1996-09-20 | 1998-06-02 | Sony Corp | 編集システム、編集方法及び表示装置 |
JPH10224665A (ja) * | 1997-02-04 | 1998-08-21 | Sony Corp | 送出システム |
JPH11266422A (ja) * | 1998-03-18 | 1999-09-28 | Toshiba Corp | 放送番組管理装置、放送番組管理方法、及び放送番組管理処理プログラムを記録した記録媒体 |
JP2004172787A (ja) * | 2002-11-19 | 2004-06-17 | Nec Corp | 映像閲覧システムおよび方法 |
JP2005203881A (ja) * | 2004-01-13 | 2005-07-28 | Sony Corp | 情報処理装置および方法、並びにプログラム |
WO2010073695A1 (ja) * | 2008-12-25 | 2010-07-01 | 日本電気株式会社 | 編集情報提示装置、編集情報提示方法、プログラム、及び記録媒体 |
JP2010282376A (ja) * | 2009-06-03 | 2010-12-16 | Nippon Telegr & Teleph Corp <Ntt> | メタデータ管理装置、メタデータ管理方法、及び、メタデータ管理プログラム |
Non-Patent Citations (1)
Title |
---|
EIJI KASUTANI ET AL.: "Video Material Archive System for Efficient Video Editing based on MPEG-7", IPSJ SIG NOTES 2003-AVM-40, vol. 2003, no. 24, 7 March 2003 (2003-03-07), pages 49 - 54 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2018156344A (ja) * | 2017-03-17 | 2018-10-04 | 日本放送協会 | 映像ストリームの一致判定プログラム |
Also Published As
Publication number | Publication date |
---|---|
JPWO2012108090A1 (ja) | 2014-07-03 |
JP5854232B2 (ja) | 2016-02-09 |
US20130314601A1 (en) | 2013-11-28 |
US9473734B2 (en) | 2016-10-18 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5651231B2 (ja) | コンテンツを決定し検索するためのメディアフィンガープリント | |
KR101382499B1 (ko) | 영상 태깅 방법 및 이를 사용하는 영상 재생 장치. | |
CN101150699B (zh) | 信息处理装置和信息处理方法 | |
US10014029B2 (en) | Video processing apparatus and method | |
US20160196252A1 (en) | Smart multimedia processing | |
KR20090089878A (ko) | 개요 및 리포트를 이미 포함하는 시청각 도큐먼트의 새로운 개요를 생성하기 위한 방법 및 상기 방법을 구현할 수 있는 수신기 | |
JP4215681B2 (ja) | 動画像処理装置及びその方法 | |
US8467666B2 (en) | Recording medium production method, recording medium, and player for recording medium | |
JP2007141092A (ja) | 情報提示装置、情報提示方法、情報提示プログラム及び情報記録媒体 | |
JP2007272468A (ja) | 情報処理装置および方法、並びにプログラム | |
US20100053149A1 (en) | Display apparatus and display method | |
JP5854232B2 (ja) | 映像間対応関係表示システム及び映像間対応関係表示方法 | |
US20130097491A1 (en) | Information processing apparatus, information processing method, and computer readable medium | |
JP6037443B2 (ja) | 映像間対応関係表示システム及び映像間対応関係表示方法 | |
WO2014103374A1 (ja) | 情報管理装置、サーバ及び制御方法 | |
US9224069B2 (en) | Program, method and apparatus for accumulating images that have associated text information | |
JP2008166895A (ja) | 映像表示装置及びその制御方法、プログラム、記録媒体 | |
JP2007213574A (ja) | コンテンツ検索方法および装置 | |
KR20090126525A (ko) | 재생 위치에 따라 디지털 컨텐츠를 관리하는 방법과 장치및 실행하는 방법 및 장치 | |
US20140189769A1 (en) | Information management device, server, and control method | |
KR20050025483A (ko) | 내용 기반 검색을 위한 개선된 질의 방법 및 그 방법을구현하기 위한 프로그램을 저장한 기록매체 | |
JP2004342302A (ja) | 光記憶媒体の専用位置に格納されるメニューアイテムデータに関連する付加データを格納するための方法及びメニューアイテムデータ及びこのメニューアイテムデータに関連する付加データを有するメニュー関連データを格納又は光記憶媒体から検索するための装置 | |
JP2005148801A (ja) | 関連情報提供方法、関連情報提供装置、関連情報提供プログラム、および、関連情報提供プログラムの記録媒体 | |
JP2004348553A (ja) | メタデータ編集方法、装置、及びプログラム | |
JP2011193386A (ja) | 電子機器及び画像処理方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11858304 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13983957 Country of ref document: US |
|
ENP | Entry into the national phase |
Ref document number: 2012556756 Country of ref document: JP Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 11858304 Country of ref document: EP Kind code of ref document: A1 |