WO2011125744A1 - コンテンツ検索装置および方法、並びにプログラム - Google Patents
コンテンツ検索装置および方法、並びにプログラム Download PDFInfo
- Publication number
- WO2011125744A1 WO2011125744A1 PCT/JP2011/058022 JP2011058022W WO2011125744A1 WO 2011125744 A1 WO2011125744 A1 WO 2011125744A1 JP 2011058022 W JP2011058022 W JP 2011058022W WO 2011125744 A1 WO2011125744 A1 WO 2011125744A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- content
- search
- displayed
- images
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 43
- 238000000605 extraction Methods 0.000 claims abstract description 28
- 239000000284 extract Substances 0.000 claims abstract description 12
- 230000002123 temporal effect Effects 0.000 claims description 20
- 230000008859 change Effects 0.000 claims description 7
- 238000004364 calculation method Methods 0.000 claims description 5
- 239000003086 colorant Substances 0.000 claims description 3
- 238000012545 processing Methods 0.000 description 22
- 230000008569 process Effects 0.000 description 17
- 238000010586 diagram Methods 0.000 description 12
- 238000004891 communication Methods 0.000 description 3
- 230000012447 hatching Effects 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000005236 sound signal Effects 0.000 description 2
- 238000004590 computer program Methods 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000009191 jumping Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/102—Programmed access in sequence to addressed parts of tracks of operating record carriers
- G11B27/105—Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/5838—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using colour
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/58—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/583—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/5854—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using shape and object relationship
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/73—Querying
- G06F16/738—Presentation of query results
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/70—Information retrieval; Database structures therefor; File system structures therefor of video data
- G06F16/78—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/783—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
- G06F16/7847—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content
- G06F16/785—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using low-level visual features of the video content using colour or luminescence
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/34—Indicating arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/234—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
- H04N21/23418—Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving operations for analysing video streams, e.g. detecting features or characteristics
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/44—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs
- H04N21/4402—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display
- H04N21/440263—Processing of video elementary streams, e.g. splicing a video clip retrieved from local storage with an incoming video stream or rendering scenes according to encoded video stream scene graphs involving reformatting operations of video signals for household redistribution, storage or real-time display by altering the spatial resolution, e.g. for displaying on a connected PDA
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/472—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content
- H04N21/47217—End-user interface for requesting content, additional data or services; End-user interface for interacting with content, e.g. for content reservation or setting reminders, for requesting event notification, for manipulating displayed content for controlling playback functions for recorded or on-demand content, e.g. using progress bars, mode or play-point indicators or bookmarks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/47—End-user applications
- H04N21/482—End-user interface for program selection
- H04N21/4828—End-user interface for program selection for searching program descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/8146—Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics
- H04N21/8153—Monomedia components thereof involving graphical data, e.g. 3D object, 2D graphics comprising still images, e.g. texture, background image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
- H04N21/8456—Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
- H04N5/93—Regeneration of the television signal or of selected parts thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N7/00—Television systems
- H04N7/16—Analogue secrecy systems; Analogue subscription systems
- H04N7/173—Analogue secrecy systems; Analogue subscription systems with two-way working, e.g. subscriber sending a programme selection signal
- H04N7/17309—Transmission or handling of upstream communications
- H04N7/17318—Direct or substantially direct transmission and handling of requests
Definitions
- the present invention relates to a content search apparatus and method, and a program, and more particularly, a content search apparatus and a content search apparatus that display a search result for each of a plurality of images so that the content of moving image content can be grasped more easily.
- the present invention relates to a method and a program.
- a waveform image such as an audio signal or a change in the luminance value of a pixel
- a timeline display an example of displaying a waveform image such as an audio signal or a change in the luminance value of a pixel
- an index video is recorded by dropping a normal video frame.
- the index video is played back for the selected video.
- the conventional technique has a problem that the user cannot easily find a scene including a desired image or the user cannot easily search for a desired characteristic scene.
- the content can be searched as an image to be searched for an individual image, a portion where a plurality of images are sequentially displayed, or a time ratio during which a desired image is displayed in the content is set to other images. It was not possible to compare with the percentage of time displayed.
- the display becomes complicated in the case of content that is long in time, and such display makes it easy for the user to understand the content of the content Very Hard to think.
- the present invention has been made in view of such a situation, and displays search results for each of a plurality of images so that the contents of moving image content can be more easily grasped.
- One aspect of the present invention is a search image input unit that receives a plurality of search image inputs that are still images, an image feature amount extraction unit that respectively extracts image feature amounts of the plurality of search images in a predetermined manner, Similarity calculation means for calculating a similarity between the image feature amount extracted in advance from a still image of a frame constituting the content and each image feature amount of the search image for the content of the moving image; By comparing the degree of similarity with a predetermined threshold to specify a frame of the still image similar to each of the search images, the specified frame, and extracted from the content in advance Based on the metadata, the content is timed with information representing the temporal position of the still image frame similar to each of the search images. Displayed on the axis which is the content retrieval device and a time line display means for generating display data for displaying the timeline.
- the timeline display means generates display data for displaying a screen having a search image display unit that displays a list of the plurality of search images and a timeline display unit that displays the content on a timeline,
- each of the plurality of search images is displayed surrounded by display frames of different colors, and in the timeline display unit, the temporal position of the frame of the still image similar to each of the search images Can be displayed as a color bar of the same color as the color of the display frame.
- the timeline display means displays each of a plurality of contents on the timeline display unit on the same screen in the timeline, and the still image frame similar to each of the search images in each of the plurality of contents.
- the information indicating the temporal position of the display frame can be displayed as a color bar having the same color as the color of the display frame.
- the timeline display unit generates display data for displaying a screen further including a moving image display unit that displays a moving image obtained by reproducing the content when the reproduction of the content displayed on the timeline is instructed. To be able to.
- the timeline display means identifies a scene constituting the content based on information on a scene change point of the content included in the metadata, and each of the scenes corresponds to a temporal length of the scene.
- the content can be displayed on the timeline by displaying the graphic for displaying the image with the horizontal or vertical direction of the screen as the time axis.
- the timeline display means can display a graphic for displaying the scene in a representative color of the scene specified by a predetermined method to display the content on the timeline.
- the search image input unit receives a plurality of search image inputs that are still images
- the image feature amount extraction unit extracts the image feature amounts of the plurality of search images, respectively, by a predetermined method.
- the similarity calculation means calculates a similarity between the image feature amount extracted in advance from a still image of a frame constituting the content and each image feature amount of the search image for the content of the video
- the specifying unit specifies the frame of the still image similar to each of the search images by comparing the calculated similarity with a predetermined threshold
- a timeline display unit is configured to specify the specified Information representing a temporal position of a frame of the still image similar to each of the search images based on metadata extracted in advance from the content.
- a content search method comprising the step of generating display data for the display the content on the time axis to display the timeline as well.
- a computer includes a search image input unit that receives a plurality of input search images that are still images, and an image feature amount extraction that respectively extracts image feature amounts of the plurality of search images using a predetermined method.
- a plurality of search image inputs that are still images are accepted, and image feature amounts of the plurality of search images are extracted in a predetermined manner, and the content is configured for moving image content.
- FIG. 5 is a diagram illustrating an example when a search is performed on the screens of FIGS. 2 to 4. It is a figure which shows the example of the screen as which the search result was displayed in the screen shown by FIG.
- FIG. 16 is a block diagram illustrating a configuration example of a personal computer.
- FIG. 1 is a block diagram showing a configuration example of a content processing apparatus according to an embodiment of the present invention.
- the content processing apparatus 10 displays the input moving image content on the timeline, searches for a predetermined image from the content as necessary, and displays the search result on the timeline.
- the content processing apparatus 10 is provided with a content input unit 21, a metadata extraction unit 22, an image feature quantity extraction unit 23, a metadata database 24, and an image feature quantity database 25.
- the content processing apparatus 10 includes a search image input unit 26, a search unit 27, and an output unit 28.
- the content input unit 21 receives input of content data.
- the content is a moving image content, and includes audio, captions, and the like as necessary. These contents are, for example, contents edited as part of a broadcast program.
- the metadata extraction unit 22 analyzes the content data supplied from the content input unit 21 and extracts metadata from the content.
- the metadata is, for example, information on the scene change point of the content, information on the time necessary for displaying the timeline described later, and information on the representative image of the content.
- the representative image of the content included in the metadata is a frame image (still image) corresponding to the scene where the level of the audio signal is the highest among the scenes constituting the content, and has a predetermined method. It is assumed that a representative image has been extracted. Information such as the representative image data and the frame number of the frame corresponding to the representative image is also extracted by the metadata extraction unit.
- the metadata extracted by the metadata extraction unit is stored in the metadata database 24 in association with content identification information, for example.
- the image feature amount extraction unit 23 analyzes the data of the content supplied from the content input unit 21, and extracts the image feature amount from the content.
- the image feature amount is information used to obtain a similarity with the search image in the processing of the search unit 27 described later.
- the image feature amount is, for example, information that divides each still image of one frame constituting the content into a plurality of predetermined regions and describes a representative color of each region. Further, the image feature amount may be information of a histogram of pixel values of still images for one frame.
- the image feature amount extracted by the image feature amount extraction unit 23 is stored in the image feature amount database 25 in association with, for example, content identification information and a frame number.
- the image feature quantity extraction unit 23 extracts the image feature quantity from the search image input from the search image input unit 26 in the same manner.
- the search image input unit 26 accepts input of a search image that is a still image.
- the search image is, for example, an image arbitrarily selected by the user, and is input to search for an image similar to the search image from the content input from the content input unit 21.
- the search unit 27 compares the image feature amount of the search image extracted by the image feature amount extraction unit 23 with the image feature amount stored in the image feature amount database 25 by a predetermined method. As a result, the similarity between the image feature amount of the search image and each image feature amount of the still image for one frame constituting the content stored in the image feature amount database 25 is calculated as a numerical value. ing.
- the search unit 27 specifies, for example, still images having image feature amounts whose similarity to the image feature amount of the search image is equal to or greater than a predetermined threshold, and outputs the frame numbers and the like of these still images. To supply.
- the output unit 28 reads content metadata from the metadata database 24 and generates display data necessary for timeline display of the content.
- the output unit 28 reads the metadata of the content from the metadata database 24, and displays the search result on the timeline based on the frame number of the still image supplied from the search unit 27 and the read metadata. Display data to generate.
- the display data output from the output unit 28 is supplied to a display (not shown) or the like and displayed as an image as described later.
- FIG. 2 is a diagram showing an example of a screen displayed on a display (not shown) based on display data output from the output unit 28.
- the screen shown in FIG. 2 has a moving image display unit 51.
- the content image input from the content input unit 21 is displayed as a moving image in the moving image display unit 51.
- the screen shown in FIG. 2 has a representative image display unit 52.
- the screen shown in FIG. 2 has a timeline display unit 53.
- the content is displayed on the timeline so that the horizontal direction in the figure corresponds to the time axis of the content. That is, the content is displayed corresponding to the time axis so that the left end of the timeline display unit 53 corresponds to the start time of the content and the right end of the timeline display unit 53 corresponds to the end time of the content.
- the timeline display of content is to display the content in some manner, but does not mean that the content is reproduced and displayed as a moving image, for example.
- the display of content in the timeline display means that, for example, a figure, a waveform, a symbol, or the like symbolizing the content or a component part of the content is displayed in association with the time axis.
- the timeline display unit 53 includes a scene display area 71.
- each scene of the content is displayed as a rectangle having a width (length) corresponding to the temporal length of each scene. That is, in the scene display area 71, eight rectangles are displayed, and it can be seen that this content is composed of eight scenes.
- the start point and end point of each scene are specified based on the scene change point information included in the metadata read from the metadata database 24, and the rectangle of the scene display area 71 is displayed. ing.
- the rectangles shown in the scene display area 71 are all displayed as white rectangles in the drawing for convenience.
- the rectangles are displayed in the representative colors of the respective scenes.
- the representative color of the scene is specified as, for example, the color corresponding to the largest pixel value among the pixel values of all frames existing in the scene.
- the representative color of the scene may be specified by other methods. In short, a color suitable for the impression of the scene may be set as the representative color.
- the representative images are respectively displayed on the upper side of the scene display area 71 in the figure. That is, nine representative images displayed on the representative image display unit 52 are displayed at positions corresponding to the frame numbers in the content.
- the representative image displayed in the first row and first column of the representative image display unit 52 is an image of a frame included in the first scene of the content, and in the upper part of the scene display area 71, the thumbnail 72-1 is displayed. It is displayed as. Note that a dotted line indicating the position of the frame is attached to the thumbnail 72-1 toward the leftmost rectangle in the drawing of the scene display area 71.
- a dotted line is drawn from the left and right ends of the thumbnail 72-1 in the drawing toward the upper center of the leftmost rectangle in the drawing of the scene display area 71 and a point on the left slightly.
- the temporal position of the frame is shown.
- the representative image displayed in the first row and the first column of the representative image display unit 52 is a frame at a position corresponding to the center of the leftmost rectangle in the drawing of the scene display area 71 on the time axis and a point slightly on the left. It is shown that this is an image. By doing in this way, the user can grasp
- thumbnails are displayed in association with the content displayed on the timeline so that the temporal position of the representative image in the content can be grasped.
- each thumbnail may be generated based on the frame data of the representative image, or may be generated in advance and included in the metadata.
- the thumbnail 72-2 represents the representative image displayed in the second row and third column of the representative image display unit 52, and it can be seen that the thumbnail 72-2 is an image of a frame included in the second scene of the content.
- the thumbnail 72-3 represents the representative image displayed in the third row and the third column of the representative image display unit 52, and it can be seen that the thumbnail 72-3 is an image of a frame included in the third scene of the content.
- thumbnails 72-4 to 72-9 are displayed, and thumbnails representing each of the nine representative images displayed on the representative image display unit 52 are displayed above the scene display area 71. It will be.
- thumbnails are displayed, for example, alternately superimposed.
- the thumbnail 72-1 is displayed so as to overlap the thumbnail 72-2, and a part of the image is hidden.
- the thumbnails displayed in a superimposed manner for example, the thumbnails displayed in a superimposed manner are displayed with a transparency of 50%, and the thumbnails displayed in a superimposed manner are displayed so as to show through.
- buttons 81 to 83 shown in FIG. 2 are configured as GUI components, for example.
- thumbnails 72-1 to 72-9 are displayed on the upper part of the scene display area 71, respectively. That is, initially, when the screen shown in FIG. 2 is displayed without the thumbnails 72-1 to 72-9 being displayed and the button 83 is operated, the thumbnail 72- is displayed above the scene display area 71. 1 to thumbnails 72-9 are respectively displayed.
- the button 81 is a button for reproducing the content and causing the moving image display unit 51 to display the moving image.
- the button 82 is a button for stopping the reproduction of the content.
- the position of the frame currently reproduced is indicated by the slider 91 in the timeline display unit 53.
- FIG. 3 is a diagram showing an example of the screen when a predetermined time elapses when the button 81 is operated to reproduce the content on the screen shown in FIG.
- the slider 91 is, for example, a red rectangular frame displayed superimposed on the scene display area 71, and is displayed so that the length in the horizontal direction in the figure increases with the passage of time.
- the right end portion of the slider 91 represents the current content playback position. As shown in FIG. 3, the right end portion of the slider 91 has moved to the right on the scene display area 71 because a predetermined time has elapsed since the content was reproduced.
- FIG. 4 is a diagram showing an example of a screen when more time has elapsed from the state shown in FIG.
- the right end of the slider 91 is moved further to the right on the scene display area 71, and the position of the frame of the representative image corresponding to the thumbnail 72-4 matches the position of the representative image. I'm doing it.
- the thumbnail 72-4 represents the representative image displayed in the second row and first column of the representative image display unit 52, and the image displayed on the moving image display unit 51 in FIG. It matches the representative image displayed in the second row and the first column.
- the thumbnail 72-4 is enlarged and displayed so as to have a larger display area than other thumbnails.
- the thumbnail 72-4 is displayed so as to overlap the thumbnail 72-3 and the thumbnail 72-5 so that the transparency is 0%. That is, the thumbnails 72-3 and 72-5 are displayed so as not to be seen through.
- the thumbnail representing the representative image is highlighted and displayed.
- the thumbnail display area is enlarged and displayed with the transparency being 0% overlaid so that the thumbnail is highlighted, but the thumbnail display by other methods has been described. The aspect may be changed.
- the contents can be displayed in a timeline in an easy-to-understand manner.
- the timeline display unit 53 each scene of the content and a thumbnail representing the representative image are displayed along the time axis. Therefore, when the user plays the content, what kind of scene can be seen at any time. It becomes possible to grasp in advance.
- thumbnails are sequentially highlighted in the timeline display unit 53 when the position of the right end of the slider 91 is moved. It is possible to evaluate the superiority or inferiority of editing the content with attention.
- step S21 the content input unit 21 receives input of content data.
- step S22 the metadata extraction unit 22 analyzes the content data supplied from the content input unit 21, and extracts metadata from the content. At this time, for example, information related to the scene change point of the content, information related to the time required for timeline display, and information such as a representative image of the content are extracted as metadata.
- the representative image of the content included in the metadata is extracted by a predetermined method.
- Information such as the representative image data and the frame number of the frame corresponding to the representative image is also extracted by the metadata extraction unit.
- step S23 the metadata database 24 stores the metadata extracted in the process of step S22.
- step S24 the output unit 28 reads content metadata from the metadata database 24, and generates display data necessary for timeline display of the content. Thereby, a screen as described above with reference to FIG. 2 is displayed. As described above, initially, the screen shown in FIG. 2 is displayed in a state where the thumbnails 72-1 to 72-9 are not displayed.
- step S25 the output unit 28 determines whether or not thumbnail display is instructed, and waits until it is determined that thumbnail display is instructed.
- step S25 when the button 83 in FIG. 2 is operated, it is determined in step S25 that a thumbnail display has been commanded, and the process proceeds to step S26.
- step S26 the output unit 28 generates display data necessary for displaying thumbnails on the screen corresponding to the display data generated in step S24. Thereby, for example, thumbnails 72-1 to 72-9 are displayed on the upper part of the scene display area 71, respectively.
- the output unit 28 appropriately generates display data for displaying those screens so as to display the screens as shown in FIGS. 3 and 4 in response to GUI operations on the screens. Has been made.
- the search image input unit 26 receives an input of a search image that is a still image.
- the search image is, for example, an image arbitrarily selected by the user, and is input to search for an image similar to the search image from the content input from the content input unit 21.
- FIG. 6 shows an example in which a search is performed on the screen described above with reference to FIGS.
- This figure shows an example in which a search image is searched from the same content as the content shown in FIGS.
- a search image display area 54 is provided, and images 101 to 103 which are search images input via the search image input unit 26 are displayed in the search image display area 54. Yes.
- three representative images are selected from the nine representative images displayed on the representative image display unit 52 as search images.
- the image displayed on the second row and the second column is the image 101
- the first row and the third column are displayed.
- the image displayed on the eye is the image 102
- the image displayed on the first row and the first column among the images displayed on the representative image display unit 52 is the image 103.
- FIG. 7 is a diagram showing an example of a screen on which the search result is displayed by operating the button 84 on the screen shown in FIG.
- the outer frames (display frames) of the images 101 to 103 displayed in the search image display area 54 are displayed in a predetermined color.
- a predetermined color is displayed by a difference in hatching pattern.
- the outer frame of the image 101 is displayed in blue
- the outer frame of the image 102 is displayed in green
- the outer frame of the image 103 is displayed in red.
- a color bar indicating the position of the frame of an image similar to each of the images 101 to 103 is displayed superimposed on a rectangle corresponding to each scene in the scene display area 71.
- the color bar shown in FIG. 7 indicates a plurality of frame positions, and is a bar having a predetermined width in the horizontal direction in the figure. That is, when a still image is searched for in the content of a moving image, normally, a plurality of images with high similarity are detected in succession, so when a frame position of an image with high similarity is colored, a color bar is displayed. It will be done.
- color bars 111-1 to 111-4 indicate the positions of image frames having a high degree of similarity to the image 101, and color bars having the same color (for example, blue) as the color of the outer frame of the image 101.
- the color bars 112-1 to 112-3 represent the positions of the frames of the image having a high degree of similarity with the image 102, and are color bars having the same color (for example, green) as the color of the outer frame of the image 102.
- the color bars 113-1 to 113-3 represent the positions of the frames of the image having a high degree of similarity to the image 103, and are color bars having the same color (for example, red) as the color of the outer frame of the image 103. Yes.
- the user can understand at a glance how long and in what part of the content the part displaying the target image (search image) is present. Furthermore, it is possible to understand at a glance how the respective parts for displaying a plurality of images (for example, the images 101 to 103) are combined in the content.
- the images 101 to 103 are missile images
- the image 101 is before the missile launch
- the image 102 is during the missile launch
- the image 103 is the image after the missile launch.
- images are often repeatedly displayed before, during, and after the launch of the missile.
- the content includes approximately four portions for displaying missile images.
- the first place is the first scene, which corresponds to the leftmost rectangle in the scene display area 71. That is, since the images before, during and after the launch of the missile are displayed in the portions corresponding to the color bar 111-1, the color bar 112-1 and the color bar 113-1, the image of the missile is displayed in this portion. It can be seen that was displayed.
- the second place is the fifth scene, which corresponds to the fifth rectangle from the left of the scene display area 71. That is, since the images before, during, and after the launch of the missile are displayed in the portions corresponding to the color bar 111-2, the color bar 112-2, and the color bar 113-2, the image of the missile is displayed in this portion. It can be seen that was displayed.
- the third place is the seventh scene, corresponding to the seventh rectangle from the left of the scene display area 71. That is, since the image before the missile launch is displayed in the portion corresponding to the color bar 111-3, it can be seen that the missile image is displayed in this portion.
- the fourth place is the eighth scene, which corresponds to the eighth rectangle from the left of the scene display area 71. That is, since the images before, during and after the launch of the missile are displayed in the portions corresponding to the color bar 111-4, the color bar 112-3, and the color bar 113-3, the image of the missile is displayed in this portion. It can be seen that was displayed.
- the user can understand the content at a glance, and can evaluate the editing technique of the editor of the content, for example. become.
- FIG. 7 an example in which an image is searched for one content has been described. However, an image may be searched for a plurality of contents.
- FIG. 8 is a diagram showing an example of a screen displayed by display data generated by the output unit 28 of the content processing apparatus 10 in FIG. 1 when images are searched for a plurality of contents.
- FIG. 8 includes a moving image display unit 151, a timeline display unit 153, and a search image display region 154. In the example of FIG. 8, seven contents are displayed on the timeline display unit 153 in a timeline.
- the timeline display unit 153 includes scene display areas corresponding to the number of contents to be searched.
- the timeline display unit 153 includes scene display areas 171-1 to 171-7.
- each scene of each content is displayed as a rectangle having a width (length) corresponding to the temporal length of each scene.
- a width (length) corresponding to the temporal length of each scene.
- three rectangles are displayed, and it can be seen that this content is composed of three scenes.
- the start point and end point of each scene are specified based on the scene change point information included in the metadata read from the metadata database 24, and are displayed in the scene display area 171-1 to the scene display area 171-7.
- the rectangle is displayed.
- Each rectangle shown in the scene display area 171-1 to the scene display area 171-7 is displayed in, for example, the representative color of the scene (however, in the drawing, all are displayed in white for convenience).
- the representative color of the scene is specified as, for example, the color corresponding to the largest pixel value among the pixel values of all frames existing in the scene. Note that the representative color of the scene may be specified by other methods. In short, a color suitable for the impression of the scene may be set as the representative color.
- a still image display area 175-1 to a still image display area 175-7 are provided on the left side of the scene display area 171-1 to the scene display area 171-7, respectively.
- Each of the images displayed in the still image display area 175-1 to the still image display area 175-7 is, for example, an image of the top frame of each content, a predetermined representative image, or the like.
- each of the character strings described on the upper side of the image displayed in the still image display area 175-1 to the still image display area 175-7 represents, for example, identification information of each content.
- the moving image display unit 151 displays a moving image obtained by reproducing the content selected by the user among the content displayed on the timeline in the timeline display unit 153.
- search image display area 154 a search image whose input has been accepted by the search image input unit 26 is displayed.
- the search image has not been input yet, and the search image is not displayed in the search image display area 154.
- the search image is, for example, an image arbitrarily selected by the user, and is input in order to search for an image similar to the search image from the content displayed on the timeline display unit 153 on the timeline.
- FIG. 9 is a diagram illustrating an example of a screen on which the images 201 to 203 are input as search images and the search results are displayed by operating the button 184 on the screen shown in FIG.
- the outer frames of the images 201 to 203 displayed in the search image display area 154 are displayed in a predetermined color.
- a predetermined color is displayed by a difference in hatching pattern.
- the color bar indicating the position of the image frame similar to each of the image 201 to the image 203 is a rectangle corresponding to each scene in the scene display area 171-1 to the scene display area 171-7. It is displayed superimposed.
- the color bar shown in FIG. 9 indicates a plurality of frame positions as in the case described above with reference to FIG. 7, and is a bar having a predetermined width in the horizontal direction in the drawing. That is, when a still image is searched for in the content of a moving image, normally, a plurality of images with high similarity are detected in succession, so when a frame position of an image with high similarity is colored, a color bar is displayed. It will be done.
- a color bar 211 represents the position of an image frame having a high degree of similarity with the image 201, and is a color bar having the same color as the color of the outer frame of the image 201 (for example, blue).
- a color bar 212 represents the position of an image frame having a high degree of similarity to the image 202, and is a color bar having the same color as the color of the outer frame of the image 202 (for example, green).
- the color bar 213 represents the position of an image frame having a high degree of similarity to the image 203, and is a color bar having the same color as the color of the outer frame of the image 203 (for example, red).
- the user can understand at a glance how long and in what part of the plurality of contents the portion displaying the target image (search image) is present. Furthermore, it is possible to understand at a glance how the respective parts displaying the plurality of images (for example, the images 101 to 103) are combined in the plurality of contents. Then, the search results can be displayed on one screen, and for example, the edited contents of the contents can be compared.
- the moving image display unit 151 displays a moving image obtained by reproducing the content selected by the user from the content displayed on the timeline display unit 153.
- FIG. 10 is a diagram illustrating an example of a screen when a moving image is displayed on the moving image display unit 151. In the example of FIG. 10, it is assumed that the user selects the scene display area 171-3 using a pointing device (not shown) and operates the button 181 to reproduce the content.
- the button 181 configured as a GUI is a button for reproducing the content and displaying the moving image on the moving image display unit 151.
- the button 182 is a button for stopping the content reproduction.
- the periphery of the scene display area 171-3 is highlighted because the scene display area 171-3 is selected.
- the periphery of the scene display area 171-3 is indicated by a dotted line, which indicates that it is highlighted.
- the timeline display unit 153 indicates the position of the frame currently reproduced by the slider.
- the slider 191-3 is displayed.
- the slider 191-3 is, for example, a red rectangular frame displayed superimposed on the scene display area 171-3, and is displayed so that the length in the horizontal direction in the figure increases as time passes.
- the right end portion of the slider 191-3 represents the current content playback position. Since a predetermined time has elapsed since the content was reproduced, the right end portion of the slider 191-3 has moved to the right on the scene display area 171-3.
- the slider 191-3 is displayed. However, when other content is reproduced, the content is displayed on the scene display area. A slider will be displayed.
- the search result can be displayed and the content can be reproduced to display the moving image.
- the search image input unit 26 receives an input of a search image.
- the search image is, for example, an image (still image) arbitrarily selected by the user, and is input to search for an image similar to the search image from the content input from the content input unit 21.
- the images 101 to 103 in FIG. 6 are input as search images.
- step S52 the image feature amount extraction unit 23 analyzes the search image input in step S51.
- step S53 the image feature quantity extraction unit 23 extracts the image feature quantity of the search image as a result of the process in step S51.
- step S54 the search unit 27 uses the image feature amount of the search image extracted by the image feature amount extraction unit 23 in the process of step S53 and the image feature amount stored in the image feature amount database 25 in a predetermined method. Compare with. As a result, the similarity between the image feature amount of the search image and each image feature amount of the still image for one frame constituting the content stored in the image feature amount database 25 is calculated as a numerical value. ing.
- step S55 the search unit 27 specifies, for example, a still image frame having an image feature amount whose similarity to the image feature amount of the search image is equal to or greater than a predetermined threshold.
- step S56 the search unit 27 notifies the output unit 28 of the search result. At this time, the frame number of the still image specified as a result of the process in step S55 is supplied to the output unit 28.
- step S57 the output unit 28 reads the content metadata from the metadata database 24, and based on the frame number of the still image supplied in the process of step S56 and the read metadata, the search result is timed. Display data for line display is generated. Thereby, for example, the screen as described above with reference to FIG. 7 or FIG. 9 is displayed.
- the output unit 28 appropriately generates display data in response to GUI operations on the screen.
- FIG. 12 is a diagram showing an example of a screen that displays the result of searching for a search image for a certain content, similar to the screen described above with reference to FIG. In this example, the content of a baseball broadcast program is retrieved and displayed on a timeline.
- the image 104 and the image 105 are displayed in the search image display area 54.
- two representative images are selected from the nine representative images displayed on the representative image display unit 52 as search images.
- the outer frames of the images 104 and 105 displayed in the search image display area 54 are displayed in a predetermined color.
- the outer frame of the image 104 is displayed in blue, and the outer frame of the image 105 is displayed in green.
- a predetermined color is displayed by a difference in hatching pattern.
- color bars representing the positions of image frames similar to the images 104 and 105 are displayed superimposed on rectangles corresponding to the respective scenes in the scene display area 71.
- Each of the color bars shown in FIG. 12 is a color bar having the same color as the color of the outer frame of the image 104 or the color of the outer frame of the image 105.
- the image 104 is an image when the pitcher throws
- the image 105 is an image in which an outfielder is running (jumping to the ball). Therefore, in the scene display area 71, a color bar representing an image frame similar to the image 104 is displayed, and in a portion where a color bar representing an image frame similar to the image 105 is displayed on the right side, the hit ball is outfield. I can see that I was flying.
- a search is made for the runner image 106 that is one level above, and there is a portion in which color bars of images similar to the image 104, the image 105, and the image 106 appear sequentially in sequence. If you find it, you can see that the batter hit a single hit.
- a search is made for the runner image 107 on the second base, and a color bar of images similar to the image 104, the image 105, and the image 107 appears sequentially in order. Is found, the batter made a two-base hit in that area.
- each rectangle each scene displayed in the scene display area 71 and a thumbnail displayed thereon, the contents can be understood more easily.
- the present invention it is possible to visualize and display the contents of content that cannot be handled by the conventional technology. Therefore, the content of the moving image content can be grasped more easily.
- the present invention it is possible to compare contents contents that cannot be handled by the conventional technology. Therefore, the content of the moving image content can be grasped more easily.
- the content is displayed on the timeline with the horizontal direction in the figure corresponding to the time axis, but the content is displayed with the vertical direction in the figure corresponding to the time axis.
- a timeline may be displayed.
- the series of processes described above can be executed by hardware or can be executed by software.
- a program constituting the software is installed from a network or a recording medium into a computer incorporated in dedicated hardware. Further, by installing various programs, it is installed from a network or a recording medium into a general-purpose personal computer 700 as shown in FIG. 13 that can execute various functions.
- a CPU (Central Processing Unit) 701 executes various processes according to a program stored in a ROM (Read Only Memory) 702 or a program loaded from a storage unit 708 to a RAM (Random Access Memory) 703. To do.
- the RAM 703 also appropriately stores data necessary for the CPU 701 to execute various processes.
- the CPU 701, the ROM 702, and the RAM 703 are connected to each other via a bus 704.
- An input / output interface 705 is also connected to the bus 704.
- the input / output interface 705 is connected to an input unit 706 composed of a keyboard, a mouse, etc., a display composed of an LCD (Liquid Crystal Display), etc., and an output unit 707 composed of a speaker.
- the input / output interface 705 is connected to a storage unit 708 composed of a hard disk and a communication unit 709 composed of a network interface card such as a modem and a LAN card.
- the communication unit 709 performs communication processing via a network including the Internet.
- a drive 710 is connected to the input / output interface 705 as necessary, and a removable medium 711 such as a magnetic disk, an optical disk, a magneto-optical disk, or a semiconductor memory is appropriately mounted. Then, the computer program read from these removable media is installed in the storage unit 708 as necessary.
- a program constituting the software is installed from a network such as the Internet or a recording medium such as a removable medium 711.
- the recording medium shown in FIG. 13 is a magnetic disk (including a floppy disk (registered trademark)) on which a program is recorded, which is distributed to distribute the program to the user separately from the apparatus main body, Removable media consisting of optical disc (including CD-ROM (compact disk-read only memory), DVD (digital versatile disk)), magneto-optical disk (including MD (mini-disk) (registered trademark)), or semiconductor memory It includes not only those configured by 711 but also those configured by a ROM 702 in which a program is recorded, a hard disk included in the storage unit 708, and the like distributed to the user in a state of being incorporated in the apparatus main body in advance.
- a magnetic disk including a floppy disk (registered trademark)
- Removable media consisting of optical disc (including CD-ROM (compact disk-read only memory), DVD (digital versatile disk)), magneto-optical disk (including MD (mini-disk) (registered trademark)), or semiconductor memory It includes not only those configured
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- Databases & Information Systems (AREA)
- Library & Information Science (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Human Computer Interaction (AREA)
- Computer Graphics (AREA)
- Computational Linguistics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
Claims (8)
- 静止画である検索画像の入力を複数受け付ける検索画像入力手段と、
予め定められた方式で前記複数の検索画像の画像特徴量をそれぞれ抽出する画像特徴量抽出手段と、
動画のコンテンツについて、前記コンテンツを構成するフレームの静止画から予め抽出された前記画像特徴量と、前記検索画像のそれぞれの画像特徴量との類似度を算出する類似度算出手段と、
前記算出された類似度を予め定められた閾値と比較することにより、前記検索画像のそれぞれに類似する前記静止画のフレームを特定する特定手段と、
前記特定された前記フレームと、前記コンテンツから予め抽出されたメタデータに基づいて、前記検索画像のそれぞれに類似する前記静止画のフレームの時間的位置を表す情報とともに前記コンテンツを時間軸上に表示してタイムライン表示させるための表示データを生成するタイムライン表示手段と
を備えるコンテンツ検索装置。 - 前記タイムライン表示手段は、
前記複数の検索画像の一覧を表示する検索画像表示部と、
前記コンテンツをタイムライン表示するタイムライン表示部とを有する画面を表示させるための表示データを生成し、
前記検索画像表示部において、前記複数の検索画像のそれぞれが異なる色の表示枠に囲まれて表示され、
前記タイムライン表示部において、前記検索画像のそれぞれに類似する前記静止画のフレームの時間的位置を表す情報が、それぞれ前記表示枠の色と同じ色のカラーバーとして表示される
請求項1に記載のコンテンツ検索装置。 - 前記タイムライン表示手段は、
複数のコンテンツのそれぞれを同一画面上の前記タイムライン表示部に前記タイムライン表示させ、
前記複数のコンテンツのそれぞれにおいて、前記検索画像のそれぞれに類似する前記静止画のフレームの時間的位置を表す情報が、それぞれ前記表示枠の色と同じ色のカラーバーとして表示される
請求項2に記載のコンテンツ検索装置。 - 前記タイムライン表示手段は、
前記タイムライン表示されたコンテンツの再生が指令された場合、前記コンテンツを再生して得られる動画を表示する動画表示部をさらに有する画面を表示させるための表示データを生成する
請求項3に記載のコンテンツ検索装置。 - 前記タイムライン表示手段は、
前記メタデータに含まれる前記コンテンツのシーンチェンジ点の情報に基づいて、前記コンテンツを構成するシーンを特定し、
前記シーンの時間的長さに対応させて前記シーンのそれぞれを表示する図形を、画面の水平または垂直方向を時間軸として表示することで前記コンテンツをタイムライン表示させる
請求項1に記載のコンテンツ検索装置。 - 前記タイムライン表示手段は、
前記シーンを表示する図形を、予め定められた方式により特定された前記シーンの代表色で表示して前記コンテンツをタイムライン表示させる
請求項5に記載のコンテンツ検索装置。 - 検索画像入力手段が、静止画である検索画像の入力を複数受け付け、
画像特徴量抽出手段が、予め定められた方式で前記複数の検索画像の画像特徴量をそれぞれ抽出し、
類似度算出手段が、動画のコンテンツについて、前記コンテンツを構成するフレームの静止画から予め抽出された前記画像特徴量と、前記検索画像のそれぞれの画像特徴量との類似度を算出し、
特定手段が、前記算出された類似度を予め定められた閾値と比較することにより、前記検索画像のそれぞれに類似する前記静止画のフレームを特定し、
タイムライン表示手段が、前記特定された前記フレームと、前記コンテンツから予め抽出されたメタデータに基づいて、前記検索画像のそれぞれに類似する前記静止画のフレームの時間的位置を表す情報とともに前記コンテンツを時間軸上に表示してタイムライン表示させるための表示データを生成するステップ
を含むコンテンツ検索方法。 - コンピュータを、
静止画である検索画像の入力を複数受け付ける検索画像入力手段と、
予め定められた方式で前記複数の検索画像の画像特徴量をそれぞれ抽出する画像特徴量抽出手段と、
動画のコンテンツについて、前記コンテンツを構成するフレームの静止画から予め抽出された前記画像特徴量と、前記検索画像のそれぞれの画像特徴量との類似度を算出する類似度算出手段と、
前記算出された類似度を予め定められた閾値と比較することにより、前記検索画像のそれぞれに類似する前記静止画のフレームを特定する特定手段と、
前記特定された前記フレームと、前記コンテンツから予め抽出されたメタデータに基づいて、前記検索画像のそれぞれに類似する前記静止画のフレームの時間的位置を表す情報とともに前記コンテンツを時間軸上に表示してタイムライン表示させるための表示データを生成するタイムライン表示手段とを備えるコンテンツ検索装置として機能させる
プログラム。
Priority Applications (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2011800173185A CN102822827A (zh) | 2010-04-09 | 2011-03-30 | 内容检索的装置和方法及程序 |
US13/639,012 US8971633B2 (en) | 2010-04-09 | 2011-03-30 | Content retrieval to facilitate recognizing content details of a moving image |
EP11765644.7A EP2560379A4 (en) | 2010-04-09 | 2011-03-30 | CONTENT DEVICE, PROCESS AND PROGRAM |
KR1020127025735A KR20130045251A (ko) | 2010-04-09 | 2011-03-30 | 콘텐츠 검색 장치 및 방법과 프로그램 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2010090610A JP5471749B2 (ja) | 2010-04-09 | 2010-04-09 | コンテンツ検索装置および方法、並びにプログラム |
JP2010-090610 | 2010-04-09 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2011125744A1 true WO2011125744A1 (ja) | 2011-10-13 |
Family
ID=44762686
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/058022 WO2011125744A1 (ja) | 2010-04-09 | 2011-03-30 | コンテンツ検索装置および方法、並びにプログラム |
Country Status (6)
Country | Link |
---|---|
US (1) | US8971633B2 (ja) |
EP (1) | EP2560379A4 (ja) |
JP (1) | JP5471749B2 (ja) |
KR (1) | KR20130045251A (ja) |
CN (1) | CN102822827A (ja) |
WO (1) | WO2011125744A1 (ja) |
Families Citing this family (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102959551B (zh) * | 2011-04-25 | 2017-02-08 | 松下电器(美国)知识产权公司 | 图像处理装置 |
US20130191745A1 (en) * | 2012-01-10 | 2013-07-25 | Zane Vella | Interface for displaying supplemental dynamic timeline content |
WO2012174301A1 (en) | 2011-06-14 | 2012-12-20 | Related Content Database, Inc. | System and method for presenting content with time based metadata |
KR102184987B1 (ko) * | 2013-11-15 | 2020-12-01 | 엘지전자 주식회사 | 영상 표시 장치 및 그 동작 방법 |
US9794599B2 (en) * | 2014-04-10 | 2017-10-17 | Telibrahma Convergent Communications Private Limited | Method and system for auditing multimedia content |
KR102373460B1 (ko) * | 2014-09-15 | 2022-03-11 | 삼성전자주식회사 | 디바이스에 객체를 표시하는 방법 및 그 디바이스 |
JP5818050B1 (ja) * | 2015-01-28 | 2015-11-18 | ビックリック株式会社 | ステータス判定システム |
JP2017016275A (ja) * | 2015-06-29 | 2017-01-19 | ヤマハ株式会社 | 制御方法 |
US9906820B2 (en) * | 2015-07-06 | 2018-02-27 | Korea Advanced Institute Of Science And Technology | Method and system for providing video content based on image |
US10318573B2 (en) * | 2016-06-22 | 2019-06-11 | Oath Inc. | Generic card feature extraction based on card rendering as an image |
JP2018125702A (ja) * | 2017-02-01 | 2018-08-09 | 富士ゼロックス株式会社 | 映像制御システム及びプログラム |
KR20210108691A (ko) * | 2020-02-26 | 2021-09-03 | 한화테크윈 주식회사 | 영상 검색 장치 및 이를 포함하는 네트워크 감시 카메라 시스템 |
JP7428855B2 (ja) * | 2020-03-17 | 2024-02-07 | Lineヤフー株式会社 | 動画解析システム、動画解析装置、動画解析方法、およびプログラム |
CN112182276A (zh) * | 2020-10-12 | 2021-01-05 | 云南财经大学 | 基于图像内容和元数据的天文海量观测数据混合检索方法 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0767073A (ja) | 1993-08-23 | 1995-03-10 | Ricoh Co Ltd | 画像記録再生装置 |
JP2001238154A (ja) | 2000-02-21 | 2001-08-31 | Sharp Corp | 動画表示装置 |
JP2007052564A (ja) * | 2005-08-16 | 2007-03-01 | Fuji Xerox Co Ltd | 情報処理システムおよび情報処理方法 |
JP2007281858A (ja) * | 2006-04-06 | 2007-10-25 | Omron Corp | 動画編集装置 |
JP2009163644A (ja) * | 2008-01-09 | 2009-07-23 | Sony Corp | 映像検索装置、編集装置、映像検索方法およびプログラム |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0778804B2 (ja) | 1992-05-28 | 1995-08-23 | 日本アイ・ビー・エム株式会社 | シーン情報入力システムおよび方法 |
JP3838805B2 (ja) * | 1999-03-02 | 2006-10-25 | 株式会社日立国際電気 | 画像検出方法 |
GB2395852B (en) * | 2002-11-29 | 2006-04-19 | Sony Uk Ltd | Media handling system |
JPWO2005050986A1 (ja) * | 2003-11-19 | 2007-07-12 | 独立行政法人情報通信研究機構 | 映像内容の提示方法及び装置 |
JP2006039753A (ja) * | 2004-07-23 | 2006-02-09 | Canon Inc | 画像処理装置、画像処理方法 |
JP4397869B2 (ja) * | 2005-09-14 | 2010-01-13 | 富士通マイクロエレクトロニクス株式会社 | スミア補正方法及びスミア補正回路 |
JP2007323319A (ja) * | 2006-05-31 | 2007-12-13 | Nippon Telegr & Teleph Corp <Ntt> | 類似検索処理方法及び装置及びプログラム |
JP5034516B2 (ja) * | 2007-01-26 | 2012-09-26 | 富士通モバイルコミュニケーションズ株式会社 | ハイライトシーン検出装置 |
JP4356762B2 (ja) * | 2007-04-12 | 2009-11-04 | ソニー株式会社 | 情報提示装置及び情報提示方法、並びにコンピュータ・プログラム |
US8850318B2 (en) * | 2007-04-23 | 2014-09-30 | Digital Fountain, Inc. | Apparatus and method for low bandwidth play position previewing of video content |
JP5121367B2 (ja) * | 2007-09-25 | 2013-01-16 | 株式会社東芝 | 映像を出力する装置、方法およびシステム |
JP4909854B2 (ja) * | 2007-09-27 | 2012-04-04 | 株式会社東芝 | 電子機器および表示処理方法 |
-
2010
- 2010-04-09 JP JP2010090610A patent/JP5471749B2/ja not_active Expired - Fee Related
-
2011
- 2011-03-30 CN CN2011800173185A patent/CN102822827A/zh active Pending
- 2011-03-30 KR KR1020127025735A patent/KR20130045251A/ko not_active Application Discontinuation
- 2011-03-30 WO PCT/JP2011/058022 patent/WO2011125744A1/ja active Application Filing
- 2011-03-30 US US13/639,012 patent/US8971633B2/en not_active Expired - Fee Related
- 2011-03-30 EP EP11765644.7A patent/EP2560379A4/en not_active Withdrawn
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH0767073A (ja) | 1993-08-23 | 1995-03-10 | Ricoh Co Ltd | 画像記録再生装置 |
JP2001238154A (ja) | 2000-02-21 | 2001-08-31 | Sharp Corp | 動画表示装置 |
JP2007052564A (ja) * | 2005-08-16 | 2007-03-01 | Fuji Xerox Co Ltd | 情報処理システムおよび情報処理方法 |
JP2007281858A (ja) * | 2006-04-06 | 2007-10-25 | Omron Corp | 動画編集装置 |
JP2009163644A (ja) * | 2008-01-09 | 2009-07-23 | Sony Corp | 映像検索装置、編集装置、映像検索方法およびプログラム |
Non-Patent Citations (1)
Title |
---|
See also references of EP2560379A4 |
Also Published As
Publication number | Publication date |
---|---|
CN102822827A (zh) | 2012-12-12 |
KR20130045251A (ko) | 2013-05-03 |
US8971633B2 (en) | 2015-03-03 |
JP5471749B2 (ja) | 2014-04-16 |
EP2560379A1 (en) | 2013-02-20 |
US20130170753A1 (en) | 2013-07-04 |
JP2011223327A (ja) | 2011-11-04 |
EP2560379A4 (en) | 2014-03-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2011125743A1 (ja) | コンテンツ処理装置および方法、並びにプログラム | |
JP5471749B2 (ja) | コンテンツ検索装置および方法、並びにプログラム | |
US9881215B2 (en) | Apparatus and method for identifying a still image contained in moving image contents | |
US8195038B2 (en) | Brief and high-interest video summary generation | |
JP5010292B2 (ja) | 映像属性情報出力装置、映像要約装置、プログラムおよび映像属性情報出力方法 | |
US20080044085A1 (en) | Method and apparatus for playing back video, and computer program product | |
US20100094441A1 (en) | Image selection apparatus, image selection method and program | |
US20200097731A1 (en) | Method for providing key moments in multimedia content and electronic device thereof | |
US20100104145A1 (en) | Electronic apparatus and video display method | |
JP2005210234A (ja) | 映像内容認識装置、録画装置、映像内容認識方法、録画方法、映像内容認識プログラム、および録画プログラム | |
US20080269924A1 (en) | Method of summarizing sports video and apparatus thereof | |
US20030030852A1 (en) | Digital visual recording content indexing and packaging | |
JP2008276340A (ja) | 検索装置 | |
US8300894B2 (en) | Method for decomposition and rendering of video content and user interface for operating the method thereof | |
JPH11220689A (ja) | 映像ソフト処理装置及び同処理プログラム記録記憶媒体 | |
US20080266319A1 (en) | Video processing apparatus and method | |
JP2010081531A (ja) | 映像処理装置及びその方法 | |
JP2005167456A (ja) | Avコンテンツ興趣特徴抽出方法及びavコンテンツ興趣特徴抽出装置 | |
JP3906854B2 (ja) | 動画像の特徴場面検出方法及び装置 | |
JP5600557B2 (ja) | コンテンツ紹介映像作成装置およびそのプログラム | |
WO2022189359A1 (en) | Method and device for generating an audio-video abstract | |
JP4007406B2 (ja) | 動画像の特徴場面検出方法 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201180017318.5 Country of ref document: CN |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11765644 Country of ref document: EP Kind code of ref document: A1 |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011765644 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 8298/CHENP/2012 Country of ref document: IN |
|
ENP | Entry into the national phase |
Ref document number: 20127025735 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13639012 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |