WO2000039707A1 - Systeme personnalise de classement et de saisie video - Google Patents

Systeme personnalise de classement et de saisie video Download PDF

Info

Publication number
WO2000039707A1
WO2000039707A1 PCT/EP1999/010221 EP9910221W WO0039707A1 WO 2000039707 A1 WO2000039707 A1 WO 2000039707A1 EP 9910221 W EP9910221 W EP 9910221W WO 0039707 A1 WO0039707 A1 WO 0039707A1
Authority
WO
WIPO (PCT)
Prior art keywords
story
segment
segments
user
video
Prior art date
Application number
PCT/EP1999/010221
Other languages
English (en)
Inventor
Jan H. Elenbaas
Nevenka Dimitrova
Thomas Mcgee
Mark Simpson
Jacquelyn A. Martino
Mohamed Abdel-Mottaleb
Marjorie Garrett
Carolyn Ramsey
Ranjit Desai
Original Assignee
Koninklijke Philips Electronics N.V.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips Electronics N.V. filed Critical Koninklijke Philips Electronics N.V.
Priority to EP99965522A priority Critical patent/EP1057129A1/fr
Priority to JP2000591534A priority patent/JP2002533841A/ja
Publication of WO2000039707A1 publication Critical patent/WO2000039707A1/fr

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/81Monomedia components thereof
    • H04N21/812Monomedia components thereof involving advertisement data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/735Filtering based on additional data, e.g. user or group profiles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/73Querying
    • G06F16/738Presentation of query results
    • G06F16/739Presentation of query results in form of a video summary, e.g. the video summary being a video sequence, a composite still image or having synthesized frames
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7834Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using audio features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/70Information retrieval; Database structures therefor; File system structures therefor of video data
    • G06F16/78Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/783Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • G06F16/7844Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content using original textual content or text extracted from visual content or transcript of audio data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/4508Management of client data or end-user data
    • H04N21/4532Management of client data or end-user data involving end-user characteristics, e.g. viewer profile, preferences
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/454Content or additional data filtering, e.g. blocking advertisements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/45Management operations performed by the client for facilitating the reception of or the interaction with the content or administrating data related to the end-user or to the client device itself, e.g. learning user preferences for recommending movies, resolving scheduling conflicts
    • H04N21/458Scheduling content for creating a personalised stream, e.g. by combining a locally stored advertisement with an incoming stream; Updating operations, e.g. for OS modules ; time-related management operations
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N7/00Television systems
    • H04N7/16Analogue secrecy systems; Analogue subscription systems
    • H04N7/162Authorising the user terminal, e.g. by paying; Registering the use of a subscription channel, e.g. billing
    • H04N7/163Authorising the user terminal, e.g. by paying; Registering the use of a subscription channel, e.g. billing by receiver means only

Definitions

  • This invention relates to the field of communications and information processing, and in particular to the field of video categorization and retrieval.
  • the continuing integration of computers and television provides for an opportunity for consumers to be provided information of particular interest.
  • many web sites offer news summaries with links to audio- visual and multimedia segments corresponding to current news stories.
  • the sorting and presentation of these news summaries can be customized for each consumer. For example, one consumer may want to see the weather first, followed by world news, then local news, whereas another consumer may only want to see sports stories and investment reports.
  • the advantage of this system is the customization of the news that is being presented to the user; the disadvantage is the need for someone to prepare the summary, and the subsequent need for the consumer to read the summary to determine whether the story is worth viewing.
  • the BNN allows the consumer to enter search words, with which the BNN sorts the story segments by the number of keywords in each story segment that match the search words. Based upon the frequency of occurrences of matching keywords, the user selects stories of interest. Similar search and retrieval techniques are becoming common in the art. For example, conventional text searching techniques can be applied to a computer based television guide, so that a person may search for a particular show title, a particular performer, shows of a particular type, and the like.
  • a disadvantage of the traditional search and retrieval techniques is the need for an explicit search task, and the corresponding selection among alternatives based upon the explicit search. Often, however, a user does not have an explicit search topic in mind. In a typical channel-surfing scenario, a user does not have an explicit search topic.
  • a channel- surfing user randomly samples a variety of channels for any of a number of topics that may be of interest, rather than specifically searching for a particular topic. That is, for example, a user may initiate a random sampling with no particular topic in mind, and select one of the many channels sampled based upon the topic that was being presented on that channel at the time of sampling.
  • a user may be monitoring the television in a "background" mode, while performing another task, such as reading or cooking. When a topic of interest appears, the user redirects his focus of interest to the television, then returns his attention to the other task when a less interesting topic is presented.
  • the user's preferences may include particular broadcast networks, anchor persons, story topics, keywords, and the like. Key frames of each selected news story are sequentially displayed; when the user views a frame of interest, the user can select the news story that is associated with the key frame for detailed viewing.
  • the news stories are stored, and the selection of a news story for detailed viewing effects a playback of the selected story.
  • this invention is particularly well suited for targeted news retrieval, the principles of this invention also allows a user to effect a directed search of other types of broadcasts as well. For example, the user may initiate an automated scan that presents samples of broadcasts that conform to the user's current preferences, akin to directed channel-surfing.
  • FIG. 1 illustrates an example block diagram of a personalized video search system in accordance with this invention.
  • FIG. 2A illustrates an example video stream 200 of a news broadcast.
  • FIG. 2B illustrates the extraction of key frames from a story segment of a video stream in accordance with this invention.
  • FIG. 3 illustrates an example user interface for a video retrieval system in accordance with this invention.
  • FIG. 4 illustrates an example block diagram of a consumer product 400 in accordance with this invention.
  • FIG. 1 illustrates an example block diagram of a personalized video search system in accordance with this invention.
  • the video retrieval system consists of a classification system 100 that classifies each segment of a video stream and a retrieval system 150 that selects and displays segments that match one or more user preferences.
  • the video retrieval system receives a video stream 101 from a broadcast channel selector 105, for example a television tuner or satellite receiver.
  • the video stream may be in digital or analog form, and the broadcast may be any form or media used to communicate the video stream, including point to point communications.
  • the example classification system 100 of FIG. 1 includes a story segment identifier 110, a classifier 120, and a visual characterizer 130.
  • the story segment identifier 110 processes a video stream 101 and identifies discrete segments 111 of the video stream 101.
  • the video stream 101 corresponds to a news broadcast, and includes multiple news stories with interspersed advertisements, or commercials.
  • the story segment identifier 110 partitions the video stream 101 into news story segments 111, either by copying each discrete story segment 111 from the video stream 101 to a storage device 115, or by forming a set of location parameters that identify the beginning and end of each discrete story segment 111 on a copy of the video stream 101.
  • the video stream 101 is stored on a storage device 115 that allows for the replay of segments 111 based on the location of the segments 111 on the medium, such as a video tape recorder, laser disc, DVD, DNR, CD-R/W, computer file system, and the like.
  • the invention is presented as having the story segments 111 stored on the storage device 115. As would be evident to one of ordinary skill in the art, this is equivalent to recording the entire video stream 101 and indexing each story segment 111 relative to the video stream 101.
  • FIG. 2A illustrates an example video stream 200 of a news broadcast.
  • a newsperson, or anchor appears 211 and introduces the first news story segment 221.
  • the anchor reappears 212 to introduce the next story segment 222.
  • the story segment 222 is complete, there is a cut 218 to a commercial 228.
  • the anchor reappears 213 and introduces the next story segment 223. This sequence of anchor-story, interspersed with commercials, repeats until the end of the news broadcast.
  • the repeated appearances 211-214 of the anchor serves to clearly identify the start of each news segment and the end of the prior news segment or commercial. Techniques are commonly available to identify commercials in a video stream, as used for example in devices that mute the sound when a commercial appears.
  • Commercials 228 may also occur within a story segment 222.
  • the cut 218 to a commercial 228 may also include a repeated appearance of the anchor, but the occurrence of the commercial 228 serves to identify the appearance as a cut 218, rather than an introduction to a new story segment.
  • the anchor may appear within the broadcast of the story segments 221- 224, but most broadcasters use one staged location for story introductions, and different staged appearances for dialog shots or repeated appearances after a commercial.
  • the anchor is shown sitting at the news desk for a story introduction, then subsequent images of the newscaster are close ups, without the news desk in the image.
  • the anchor is presented full screen to introduce the story, then on a split screen when speaking with a field reporter.
  • the anchor shot is full facial to introduce a story, and profiled within the story.
  • Each story segment 221-224 of FIG. 2A has an associated audio segment 231-234, and possibly closed caption text 241-244.
  • the audio segments 231-234 are synchronous with the video segments, and may be included within each story segment 221-224. Due to the differing transmission times of audio and text, the closed caption text segments 241-244 do not necessarily consume the same time span as the audio segments 231-234.
  • the story segment identifier 110 may also include a speech recognition device that creates text segments 241-244 corresponding to each audio segment 231-234.
  • the text segments 241-244 include text from other sources as well.
  • a television guide may be available that provides a synopsis of each story, a list of characters, a reviewer's rating, and the like.
  • an on-line guide may be available that provides a list of headlines, a list of newscasters, a list of companies or people contained in the broadcast, and the like.
  • textual annotations indicating the broadcast channel being monitored by the broadcast channel selector 105, such as "ABC", "NBC", “CNN”, etc., as well as the name of each anchor introducing each story.
  • the anchor's name may be automatically determined based on image recognition techniques, or manually determined. Other annotations may include the time of the broadcast, the locale of each story, and so on. In a preferred embodiment of this invention, each of these text formatted information segments will be associated with their corresponding story segment. Teletext formatted data may also be included in text segment 241-244.
  • FIG. 244 of FIG. 2A correspond to the story segments 111, audio segments 112, and text segments 113 from the story segment identifier 110 of FIG. 1, and the video 228, audio 238 and text 248 segments correspond to a commercial.
  • FIG. 2B illustrates the extraction of key frames from a story segment of a video stream in accordance with one aspect of this invention.
  • the story segment 221 includes a number of scenes 251-253.
  • the first scene 251 of story segment 221 corresponds to the image 211 of the anchor introducing the story segment 221.
  • the next scene 252 may be images from a remote camera covering the story, and so on.
  • Each scene consists of frames.
  • the first frame 261, 271, 281 of each scene 251, 252, 253 forms a set of key frames 291, 292, 293 associated with the story segment 221, the key frames forming a pictorial summary of the story segment 221.
  • the key frames 291, 292, 293 of FIG. 2B correspond to the key frames 114 from the story segment identifier 110 of FIG. 1.
  • the first frame of each scene can be identified based upon the differences between frames. As the anchor moves during the introduction of the story, for example, only slight differences will be noted from frame to frame. The region of the image corresponding to the news desk, or the news room backdrop, will not change substantially from frame to frame. When a scene change occurs, for example by switching to a remote camera, the entire image changes substantially.
  • a number of image compression or transform schemes provide for the ability to store or transmit a sequence of images as a sequence of difference frames. If the differences are substantial, the new frames are typically encoded directly as reference frames; subsequent frames are encoded as differences from these reference frames.
  • FIG. 2B illustrates such a scheme by the relative size of each frame F in each scene 251-253.
  • the first frame 261, 271, 281 of each scene 251, 252, 253 are encoded as reference frames, containing a substantial amount of information, or encoded as difference frames containing a substantial number of differences from their prior frames. After the change of scenes, subsequent frames are smaller, reflecting the same overall scene with minor changes caused by the movement of the objects in the frame or changes to the camera angle or magnification.
  • the amount of information contained in each frame is directly related to the changes from one frame to the next.
  • images are transformed using a Discrete Cosine Transformation (DCT), which produces an encoding of each frame having a size that is strongly correlated to the amount of random change from one frame to the next.
  • DCT Discrete Cosine Transformation
  • frames 262, 263, and 264 are shown to be substantially smaller than frame 261, because they contain less information than frame 261, which is the frame corresponding to a scene change.
  • the key frames 291, 292, 293 correspond to the frames containing the most information 261, 271, 281 in the story segment 221.
  • Other techniques of selecting key frames would be evident to one of ordinary skill in the art. For example, one could choose the frame from the center of each scene, or choose the frame having the least difference from all the other frames in the scene, using for example a least squares determination, and the like.
  • manual and semi-automated techniques may also be employed to select key frames, the composite of which form a pictorial summary of each story segment.
  • the classifier 120 characterizes each story segment 111 of FIG. 1. In a preferred embodiment, the classifier 120 effects the characterization automatically, although manual or semi-automated techniques may be used as well.
  • the primary means of characterization in the preferred embodiment is based on the text segments 113 from the story segment identifier 110. If the text segments 113 include annotations such as the broadcast channel and the anchor's name, these annotations are used to identify the story segment in corresponding "broadcaster" and "anchor" categories.
  • keywords such as “victim”, “police”, “crime”, “defendant”, and the like are used to characterize a news story under the topic of "crime”. Keywords such as “democrat”, “republican”, “house”, “senate”, “prime minister”, and the like are used to characterize a news story under the topic of "politics”. Sub categorizations can also be defined, such that “home run” characterizes a story as sub category “baseball” under category “sports”, while “touch down” characterizes a story as sub category “football” under the same category “sports”.
  • the audio segments 112 may be used for categorization.
  • the audio segments 112 may be converted to text and the categorization applied to the text.
  • the audio segments 112 may be analyzed for sounds of laughter, explosions, gunshots, cheers, and the like to determine appropriate characterizations, such as "comedy”, “violence”, and “celebration”.
  • a visual characterizer 130 characterizes story segments 111 based on their visual content.
  • the visual characterizer 130 may be used to identify people appearing in the story segments, based on visual recognition techniques, or to identify topics based on an analysis of the image background information.
  • the visual characterizer 130 may include a library of images of noteworthy people.
  • the visual characterizer 130 identifies images containing a single or predominant figure, and these images are compared to the images in the library.
  • the visual characterizer 130 may also contain a library of context scenes and associated topic categories. For example, an image containing a person aside a map with isobars would characteristically identify the topic as "weather”.
  • image processing techniques can be used to characterize an image as an "indoor” or “outdoor” image, a "city”, “country”, or “sea” locale, and so on.
  • These visual characterizations 131 are provided to the classifier 120 for adding, modifying, or supplementing the categorizations formed from the text 113 and audio 112 segments associated with each story segment 111.
  • the appearance of smoke in a story segment 111 may be used to refine a characterization of a siren sound in the audio segment 112 as "fire", rather than "police”.
  • the visual characterizer 130 may also be used to prioritize key frames.
  • a newscast may have dozens or hundreds of key frames based upon a selection of each new scene.
  • the number of key frames is reduced by selecting those images likely to contain more information than others.
  • Certain image contents are indicative of images having significant content. For example, a person's name is often displayed below the image of the person when the person is first introduced during a newscast. This composite image of a person and text will, in general, convey significant information regarding the story segment 111. Similarly a close-up of a person or small group of people will generally be more informative than a distant scene, or a scene of a large group of people.
  • key frames are prioritized by such image content analysis, as well as by other cues, such as the chronology of scenes. In general, the more important scenes are displayed earlier in the story segment 111 than less important scenes.
  • the prioritization of key frames is also used to create a visual table of contents for the story segments 111, as well as for a visual table of contents for the video stream 101, by selecting a given number frames in priority order.
  • the classification system 100 provides the set of characterizations, or classification 121, of each story segment 111 from the classifier 120, and the set of key frames
  • the classification 121 may be provided in a variety of forms. Predefined categories such as “broadcaster”, “anchor”, “time”, “locale”, and “topic” are provided in the preferred embodiment, with certain categories, such as “locale” and “topic” allowing for multiple entries. Another method of classification that is used in conjunction with the predefined categories is a histogram of select keywords, or a list of people or organizations mentioned in the story segment 111.
  • the classification 121 used in the classification system 100 should be consistent or compatible with, albeit not necessarily identical to, the filtering system used in the filter 160 of the retrieval system 150.
  • a classification translator can be appended between the classification system 100 and retrieval system 150 to convert the classification 121, or a portion of the classification 121, to a form that is compatible with the filtering system used in the filter 160.
  • This translation may be automatic, manual, or semi-automated.
  • the filter 160 of the retrieval system 150 identifies the story segments 111 that conform to a set of user preferences 191, based on the classification 121 of each of the story segments 111.
  • the user is provided a profiler 190 that encodes a set of user input into preferences 191 that are compatible with the filtering system of the filter 160 and compatible with the classification 121. For example, if the classification 121 includes an identification of broadcast channels or anchors, the profiler 190 will provide the user the option of specifying particular channels or anchors for inclusion or exclusion by the filter 160.
  • the profiler 190 includes both "constant” as well as “temporal” preferences, allowing the user to easily modify those preferences that are dependent upon the user's current state of mind while maintaining a set of overall preferences.
  • the temporal set for example, would be a choice of topics such as "sports" and "weather”.
  • the constant set for example, would be a list of anchors to exclude regardless of whether the anchor was addressing the current topic of interest.
  • the constant set may include topics such as "baseball” or "stock market”, which are to be included regardless of the temporal selections.
  • the profiler 190 allows for combinations of criteria using conjunctions, disjunctions, and the like. For example, the user may specify a constant interest in all "stock market" stories that contain one or more words that match a specified list of company names.
  • the filter 160 identifies each of the story segments 111 with a classification 121 that matches the user preferences 191.
  • the degree of matching, or tightness of the filter is controllable by the user.
  • a user may request all story segments 111 that match any one of the user's preferences 191; in another extreme, the user may request all story segments 111 that match all of the user's preferences 191.
  • the user may request all story segments 111 that match at least two out of three topic areas, and also contain at least one of a set of keywords, and so on.
  • the user may also have negative preferences 191, such as those topics or keywords that the user does not want, for example "sports" but not "hockey".
  • the filter 160 identifies each of the story segments 111 satisfying the user's preferences 191 as filtered segments 161.
  • the filter 160 contains a sorter that ranks each story in dependence upon the degree of matching between the classification 121 and the user preferences 191, using for example a count of the number of keywords of each topic in each classification 121 of the story segments 111.
  • the ranking herein is presented as a unidimensional, scalar quantity, although techniques for multidimensional ranking, or vector ranking, are common in the art.
  • the ranking 162 may be heavily weighted by the user's preferred anchor, or preferred broadcast channel; this ranking 162 may also be weighted by the time of each newscast, in preference to the most recent story.
  • the user has the option to adjust the weighting factors. For example, the user may make a negative selection absolute: if the segment contains the negated topic or keyword, it is assigned the lowest rating, regardless of other matching preferences. Any number of common techniques can be used to effect such prioritization, including the use of artificial intelligence techniques such as knowledge based systems, fuzzy logic systems, expert systems, learning systems and the like.
  • the filter 160 selects story segments 111 based on this ranking 162, and provides the ranking 162 of each of these selected, or filtered, segments 161 to the presenter 170 of the retrieval system 150.
  • the filter 160 also identifies the occurrences of similar stories in multiple story segments, to identify popular stories, commonly called "top stories". This identification is determined by a similarity of classifications 121 among story segments 111, independent of the user's preferences 191. The similarity measure may be based upon the same topic classifications being applied to different story segments 111, upon the degree of correlation between the histograms of keywords, and so on. Based upon the number of occurrences of similar stories, the filter 160 identifies the most popular current stories among the story segments 111, independent of the user's preferences 191. Alternatively, the filter 160 identifies the most popular current stories having at least some commonality with the preferences 191. From these most popular current stories, the filter chooses one or more story segments 111 for presentation by the presenter 170, based upon the user's preferences 191 for broadcast channel, anchor person, and so on.
  • the presenter 170 presents the key frames 114 of the filtered story segments 161 on a display 175.
  • the set of key frames associated with each story segment 111 provides a pictorial summary of each story segment 111.
  • the presenter 170 presents the pictorial summary 171 of those story segments 161 which correspond to the user preferences 191.
  • the number of key frames displayed for each story segment 161 is determined by the aforementioned prioritization schemes based on image content, chronology, associated text, and the like.
  • the presentation of the pictorial summary may be accompanied by the playing of portions of the audio segments that are associated with the story segment 111.
  • the portion of the audio segment may be the first audio segment of each story segment, corresponding to the introduction of the story segment by the anchor.
  • a summary of the text segment may also be displayed coincident with the display of the pictorial summary 171.
  • the user selects the filtered story segment for full playback by a player 180 in the retrieval system 150.
  • the user may effect the selection by pointing to the displayed key frames of the story of interest, using for example a mouse, or by voice command, gesture, keyboard input, and the like.
  • the player 180 Upon receipt of the user selection 176 the player 180 displays the selected story segment 181 on the display 175.
  • FIG. 3 illustrates an example user interface for the retrieval system 150.
  • the display 175 contains panes 310 for displaying filtered story segments key frames 171.
  • the display 175 includes four panes 310a, 310b, 310c and 310d, although fewer or more panes can be selected via the presenter controls 350.
  • the presenter sequentially presents each of the key frames 171 in the panes 310.
  • each of the key frames 171 corresponding to one story segment 161 are presented sequentially in one of the panes 310a, 310b, 310c, or 310d. That is, in FIG.
  • the key frames of four story segments 161 are displayed simultaneously, each pane providing the pictorial summary for each of the story segments 161.
  • the user has the option of determining the duration of each key frame 171, and whether the key frames 171 from a story segment 161 are repeated for a given time duration before the set of key frames 171 from another story segment 161 are presented in that pane. After all the key frames 114 of all the filtered story segments 161 are presented, the cycle is repeated, thereby providing a continuous slide show of the key frames of story segments that conform to the user's preferences.
  • Alternative display methods can be employed. For example, four segments from a story segment 161 may be displayed in all four of the panes 310a-310d simultaneously.
  • one pane may be defined as a primary pane, which is configured to contain the highest priority scene of the story segment 161 while the other panes sequentially display lower priority scenes.
  • presenter controls 350 are provided to facilitate the customization of the presentation and selection of key frames 171.
  • the presenter 170 can use the ranking 162 to determine the frequency or duration of each presented set of key frames 171. That is, for example, the presenter 170 may present the key frames 114 of filtered segments 161 at a repetition rate that is proportional to the degree of correspondence between the filtered segments 161 and user preferences 191. Similarly, if a large number of filtered segments 161 are provided by the filter 160, the presenter 170 may present the key frames 114 of the segments 161 that have a high correspondence with the user preferences 191 at every cycle, but may present the key frames 114 of the segments that have a low correspondence with the user preferences 191 at fewer than every cycle.
  • the presenter controls 350 also allow the user to control the interaction between the presenter 170 and the player 180.
  • the user can simultaneously view a selected story segment 181 in one pane 310 while key frames 171 from other story segments continue to be displayed in the other panes.
  • the selected story segment 181 may be displayed on the entire area of the display 175.
  • play control functions in 350 for conventional playback functions such as volume control, repeat, fast forward, reverse, and the like. Because the story segments 111 are partitioned into scenes in the story segment identifier, the playback functions 350 may include such options as next scene, prior scene, and so on.
  • buttons 320 are provided to allow the user to set preferences 191 in select categories.
  • the “media” button 320a provides the user options regarding the broadcast channels, anchor persons, and the like.
  • the “time” button 320b provides the user options regarding time settings, such as how far back in time the filter 160 should consider story segments.
  • the “topics” button 320c allows the user to choose among topics, such as sports, art, finance, crime, etc.
  • the “locale” button 320d allows the user to specify geographic areas of interest.
  • the “top stories” button 320e allows the user to specify filter parameters that are to applied to the aforementioned identification of popular story segments.
  • the "keywords” button 320f allows the user to identify specific keywords of interest. Other categories and options may also be provided, as would be evident to one of ordinary skill in the art.
  • the user interface of FIG. 3 also allows for selection of presentation 330 and player 340 modes.
  • the presentor 170 can be set to present key frames of story segments selected by the user's preference settings, or key frames of "top" story segments.
  • the player 180 can be set to operate in a browse mode, corresponding to the operation discussed above, wherein the user browses the key frames and selects story segments of interest; or in a play thru mode, wherein the player 180 presents each of the filtered story segments 161 in succession; and in a scan mode, wherein the player 180 presents the first scene of each filtered story segment 161 in succession.
  • the presentation can be multidimensional, wherein, for example, the degree of correlation of a segment 111 to the user's preferences 191 identifies a depth, and the key frames are presented in a multidimensional perspective view using this depth to determine how far away from the user the key frames appear.
  • different categories 320 of user preferences can be associated with different planes of view, and the key frames of each segment having strong correlation with the user preferences in each category are displayed in each corresponding plane.
  • the principles presented herein will be recognized by one of ordinary skill in the art to be applicable to other retrieval tasks as well.
  • the principles of the invention presented herein can be used for directed channel-surfing.
  • a channel- surfing user searches for a program of interest by randomly or systematically sampling a number of broadcast channels until one of the broadcast programs strikes the user's interest.
  • the classification system 100 and retrieval system 150 in an on-line mode, a more efficient search for programs of interest can be effected, albeit with some processing delay.
  • the story segment identifier 110 provides text segments 113, audio segments 112, and key frames 114 corresponding to the current non-commercial portions of the broadcast channel.
  • the classifier 120 classifies these portions using the techniques presented above.
  • the filter 160 identifies those portions that conform to the user's preferences 191, and the presenter 170 presents the set of key frames 171 from each of the filtered portions 161.
  • the broadcast channel selector 105 is tuned to the channel corresponding to the selected key frames 171
  • the story segment identifier 110, storage device 115 and player 180 are placed in a bypass mode to present the video stream 101 of the selected channel to the display 175.
  • FIG. 4 illustrates an example consumer product 400 in accordance with this invention.
  • the product 400 may be a home computer or a television; it may be a video recording device such as a VCR, CD-R/W, or DVR device; and so on.
  • the example product 400 records potentially interesting story segments 111 for presentation and selection by a user.
  • the story segments 111 are extracted or indexed from a video stream 101 by the classification system 100, as discussed above with regard to FIG. 1.
  • the video stream 101 is selected from a multichannel input 401, such as a cable or antenna input, via a selector 420 and tuner 410.
  • the selector 420 is a programmable multi-event channel selector, such as found in conventional VCR devices.
  • the user programs the selector 420 to tune the tuner 410 to a particular channel of interest at each particular event time for a specified duration. For example, a user may program the time and duration of morning news on one channel, the evening news on another channel, and late night news on yet another channel.
  • the stories 111 are segmented and stored on the recorder 430 via the classification system 100, which also classifies each segment 111 and extracts relevant key frames 171 for display on the input/output device 440, as discussed above.
  • the recorder 430 is a continuous-loop recorder, or continuous circular buffer recorder, which automatically erases the oldest segments 111 as it records each of the newest segments 111 , so as to continually provide as many recent segments 111 as it recording media allows.
  • the user accesses the system via the input/output device 440 and is presented the key frames of the most recent segments 111 that match the user's preferences; thereafter, the user selects segments 181 for display based on the presented key frames 171.
  • the retrieval system 150 may be configured to provide selective erasure, via 451, rather than the oldest-erasure scheme discussed above.
  • the retrieval system 150 identifies the segments 111 that are on the recording media that have the least correlation with the user's preferences. Instead of replacing the oldest segments with the newest segments, the segments of least potential interest to the user are replaced by the newest segments.
  • the retrieval system 150 also terminates the recording of the newest segment when it determines, based on the classification of the newest segment by the classification system 100, that the newest segment is of no interest to the user, based on the user preferences.
  • the product 400 optionally provides for the selection of channels by the selector 420 via a prefilter 425.
  • the prefilter 425 effects a filtering of the segments 111 by controlling the selection of channels 401 via the selector 420 and tuner 410.
  • ancillary text information is commonly available that describes the programs that are to be presented on each of the channels of the multichannel input 401.
  • this ancillary information, or program guide may be a part of the multichannel input 401, or via a separate program guide connection 402.
  • the prefilter 425 identifies the programs in the program guide 402 that have a strong correlation with the user preferences 191, and programs the selector 420 to select these programs for recording, classification, and retrieval, as discussed above.
  • the capabilities and parameters of this invention may be adjusted depending upon the capabilities of each particular embodiment.
  • the product 400 may be a portable palm-top viewing device for commuters who have little time to watch live newscasts. The commuter connects the product 400 to a source of multichannel input 401 overnight to record stories 111 of potential interest; then, while commuting (as a passenger) uses the product 400 to retrieve stories of interest 181 from these recorded stories 111.
  • resources are limited, and the parameters of each component are adjusted accordingly.
  • the number of key frames 114 associated with each segment 111 may be substantially reduced, the prefilter 425 or filter 160 may be substantially more selective, and so on.
  • the classification 100 and retrieval systems 150 of FIG. 1 may be provided as standalone devices that dynamically adjusts their parameters based upon the components to which they are attached.
  • the classification system 100 may be a very large and versatile system that is used for classifying story segments for a variety of users, and different models of retrieval systems 150, each having different levels of complexity and cost, are provided to the users for retrieving selected story segments.
  • the key frames 114 have been presented herein as singular images, although a key frame could equivalently be a sequence of images, such as a short video clip, and the presentation of the key frames would be a presentation of each of these video clips.
  • the components of the classification system 100 and retrieval system 150 may be implemented in hardware, software, or a combination of both. The components may include tools and techniques common to the art of classification and retrieval, including expert systems, knowledge based systems, and the like.
  • the presentor 170 and filter 160 may include a randomization factor, that augments the presentation of key frames 114 of segments 161 having a high correspondence with the user preferences 191 with key frames 114 of randomly selected segments, regardless of their correspondence with the preferences 191.
  • the source of the video stream 101 may be digital or analog, and the story segments 111 may be stored in digital or analog form, independent of the source of the video stream 101.
  • the techniques presented herein may also be used for the classification, retrieval, and presentation of video information from sources such as public and private networks, including the Internet and the World Wide Web, as well.
  • the association between sets of key frames 114 and story segments 111 may be via embedded HTML commands containing web site addresses, and the retrieval of a selected story segment 181 is via the selection of a corresponding web site.
  • the broadcast channel selector 105 may be an integral part of the story segment identifier 110, or it may be absent if the classification and retrieval system is being used to retrieve story segments from a single source video stream, or a previously recorded video stream 101.
  • the story segment identifier 110 may process multiple broadcast channels simultaneously using parallel processors.
  • the filter 160 and profiler 190 may be integrated as a single selector device.
  • the key frames 114 may be stored on, or indexed from, the recorder 115, and the presenter 170 functionality provided by the player 180.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Databases & Information Systems (AREA)
  • Theoretical Computer Science (AREA)
  • Signal Processing (AREA)
  • Library & Information Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Computer Security & Cryptography (AREA)
  • Business, Economics & Management (AREA)
  • Marketing (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Television Signal Processing For Recording (AREA)

Abstract

Cette invention concerne un système de saisie vidéo permettant à l'utilisateur de choisir rapidement et facilement, puis de recevoir des programmes vidéo d'intérêt. A partir de thèmes ordonnés, ce système de classement vidéo propose des échantillons d'histoires qui répondent aux préférences du moment de l'utilisateur. Ces préférences peuvent porter sur une chaîne de télédiffusion particulière, des personnes, des thèmes, des mots clés, etc. Pour chaque histoire sélectionnée, des trames marquantes sont affichées séquentiellement. Lorsque passe une trame qui l'intéresse, le téléspectateur choisit l'histoire dont est tirée cette trame pour la visionner plus en détail. Cette invention convient particulièrement bien pour la saisie de nouvelles ciblées. Selon un mode de réalisation préféré, on stocke des séquences de nouvelles, lesquelles sont sélectionnées et diffusées en fonction de trames tirées de ces séquences. De par ses caractéristiques, cette invention permet à l'utilisateur d'effectuer des recherches directes dans d'autres programmes. Celui-ci peut par exemple lancer une recherche automatique qui fait apparaître des échantillons de programmes conformes à ses préférences, comme s'il naviguait directement sur les canaux.
PCT/EP1999/010221 1998-12-23 1999-12-15 Systeme personnalise de classement et de saisie video WO2000039707A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP99965522A EP1057129A1 (fr) 1998-12-23 1999-12-15 Systeme personnalise de classement et de saisie video
JP2000591534A JP2002533841A (ja) 1998-12-23 1999-12-15 個人用ビデオ分類及び検索システム

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US22027798A 1998-12-23 1998-12-23
US09/220,277 1998-12-23

Publications (1)

Publication Number Publication Date
WO2000039707A1 true WO2000039707A1 (fr) 2000-07-06

Family

ID=22822888

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP1999/010221 WO2000039707A1 (fr) 1998-12-23 1999-12-15 Systeme personnalise de classement et de saisie video

Country Status (5)

Country Link
EP (1) EP1057129A1 (fr)
JP (1) JP2002533841A (fr)
KR (1) KR100711948B1 (fr)
CN (1) CN1116649C (fr)
WO (1) WO2000039707A1 (fr)

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1191460A1 (fr) * 2000-09-21 2002-03-27 Swisscom AG Méthode et système de classification de fichiers audio et vidéo numériques
JP2002169819A (ja) * 2000-12-01 2002-06-14 Kanebo Ltd 動画情報配信方法およびそれに用いる動画情報配信システム
WO2002051140A2 (fr) * 2000-12-18 2002-06-27 Koninklijke Philips Electronics N.V. Appareil et procede de classification de programmes faisant appel a des indices observes dans les informations de transcription
WO2003010965A1 (fr) * 2001-07-26 2003-02-06 Koninklijke Philips Electronics N.V. Procede de facturation d'annonceurs a partir d'une commutation d'annonces publicitaires entre des canaux de television
EP1291790A2 (fr) * 2001-08-15 2003-03-12 Siemens Corporate Research, Inc. Groupement et classement automatique de contenu textuel
WO2003046761A2 (fr) * 2001-11-28 2003-06-05 Koninklijke Philips Electronics N.V. Systeme et procede de recherche d'informations associees a des sujets cibles
WO2003050718A2 (fr) * 2001-12-11 2003-06-19 Koninklijke Philips Electronics N.V. Systeme et procede d'extraction d'informations relatives a des personnes dans des programmes video
WO2005001715A1 (fr) * 2003-06-30 2005-01-06 Koninklijke Philips Electronics, N.V. Systeme et procede permettant de produire un condense multimedia de flux multimedias
US6920281B1 (en) 2000-06-30 2005-07-19 Koninklijke Philips Electronics N.V. Video memory manager for use in a video recorder and method of operation
JP2006033659A (ja) * 2004-07-21 2006-02-02 Sony Corp コンテンツ記録再生装置、コンテンツ記録再生方法及びそのプログラム
WO2006103633A1 (fr) 2005-03-31 2006-10-05 Koninklijke Philips Electronics, N.V. Synthese de recits d'actualites composites
WO2008007279A2 (fr) * 2006-07-04 2008-01-17 Koninklijke Philips Electronics N.V. Procédé de substitution de contenu
WO2008007274A3 (fr) * 2006-07-04 2008-05-29 Koninkl Philips Electronics Nv Procédé de substitution de contenu
CN101894129A (zh) * 2010-05-31 2010-11-24 中国科学技术大学 基于在线视频分享网站结构及视频描述文本信息的视频主题发现的方法
WO2011039773A2 (fr) * 2009-09-14 2011-04-07 Tata Consultancy Services Ltd. Système d'analyse de nouvelles télévisées pour canaux de diffusion multilingue
CN102184235A (zh) * 2011-05-13 2011-09-14 广州星海传媒有限公司 一种基于机顶盒的数字电视节目推荐方法及系统
EP2345251A4 (fr) * 2008-10-31 2012-04-11 Hewlett Packard Development Co Organisation de données vidéo
US8209623B2 (en) 2003-12-05 2012-06-26 Sony Deutschland Gmbh Visualization and control techniques for multimedia digital content
US8671337B2 (en) 2007-03-27 2014-03-11 Sharp Laboratories Of America, Inc. Methods, systems and devices for multimedia-content presentation
CN106550254A (zh) * 2015-09-23 2017-03-29 北京丰源星际传媒科技有限公司 影院放映电影贴片广告的监播方法及系统
USRE47908E1 (en) 1991-12-23 2020-03-17 Blanding Hovenweep, Llc Ergonomic man-machine interface incorporating adaptive pattern recognition based control system
US20200159759A1 (en) * 2018-11-20 2020-05-21 Comcast Cable Communication, Llc Systems and methods for indexing a content asset
USRE48056E1 (en) 1991-12-23 2020-06-16 Blanding Hovenweep, Llc Ergonomic man-machine interface incorporating adaptive pattern recognition based control system
CN113360709A (zh) * 2021-05-28 2021-09-07 维沃移动通信(杭州)有限公司 短视频侵权风险的检测方法、装置和电子设备

Families Citing this family (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
BRPI0212375B1 (pt) * 2001-09-07 2016-05-24 Intergraph Hardware Tech Co método para estabilizar uma imagem
KR100452085B1 (ko) * 2004-01-14 2004-10-12 엔에이치엔(주) 카테고리 별 키워드의 입력 순위를 제공하기 위한 검색서비스 시스템 및 그 방법
KR100683501B1 (ko) * 2005-02-23 2007-02-15 인천대학교 산학협력단 신경망 기법을 이용한 뉴스 비디오의 앵커 화면 추출 장치및 그 방법
JP2006279429A (ja) * 2005-03-29 2006-10-12 Kddi Corp 映像コンテンツの推薦システム
JP4919879B2 (ja) * 2007-06-07 2012-04-18 ソニー株式会社 情報処理装置および方法、並びにプログラム
US8588583B2 (en) * 2007-08-22 2013-11-19 Adobe Systems Incorporated Systems and methods for interactive video frame selection
US7836093B2 (en) * 2007-12-11 2010-11-16 Eastman Kodak Company Image record trend identification for user profiles
JP5709359B2 (ja) * 2009-04-07 2015-04-30 ソニー株式会社 通信装置、通信方法、及び通信システム
CN101887445B (zh) * 2009-05-12 2012-10-10 大相科技股份有限公司 处理动态图像的方法、系统与加值处理的方法
KR101043630B1 (ko) 2009-07-31 2011-06-22 고려대학교 산학협력단 뉴스 제공 방법 및 시스템
CN103299600B (zh) * 2011-01-04 2016-08-10 汤姆逊许可公司 用于传输直播媒体内容的装置和方法
CN106202421B (zh) * 2012-02-02 2020-01-31 联想(北京)有限公司 一种获得视频的方法、装置及播放视频的方法、装置
CN103258555A (zh) * 2012-02-16 2013-08-21 岳振华 一种视频编辑方法及一种视频编辑装置
CN102799637A (zh) * 2012-06-27 2012-11-28 北京邮电大学 一种电视节目内部自动生成主要人物摘要的方法
CN103200463A (zh) * 2013-03-27 2013-07-10 天脉聚源(北京)传媒科技有限公司 一种视频摘要生成方法和装置
KR101475479B1 (ko) * 2013-06-27 2014-12-23 삼성중공업 주식회사 선박용 영상 데이터 처리 장치 및 방법
CN108959323B (zh) * 2017-05-25 2021-12-07 腾讯科技(深圳)有限公司 视频分类方法和装置
CN108241729A (zh) * 2017-09-28 2018-07-03 新华智云科技有限公司 筛选视频的方法及设备
CN109241303B (zh) * 2018-06-29 2022-10-21 Oppo广东移动通信有限公司 视频创建方法及相关产品
CN113704597A (zh) * 2020-05-21 2021-11-26 阿波罗智联(北京)科技有限公司 内容推荐方法、装置和设备

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH05342263A (ja) * 1992-06-08 1993-12-24 Nippon Telegr & Teleph Corp <Ntt> ビデオデータ検索支援方法
JPH10111872A (ja) * 1996-10-08 1998-04-28 Nippon Telegr & Teleph Corp <Ntt> 動画配信装置及び方法
JP2001503895A (ja) * 1996-11-15 2001-03-21 サーノフ コーポレイション ビデオ情報を効果的に表示、保存、およびアクセスするための方法および装置
JP3426960B2 (ja) * 1998-04-23 2003-07-14 日本電信電話株式会社 映像編集方法及び装置及びそのプログラムを記録した記録媒体

Non-Patent Citations (8)

* Cited by examiner, † Cited by third party
Title
ANDREW MERLINO; DARYL MOREY; MARK MAYBURY: "MITRE Corporation, Bedford MA, Broadcast News Navigation using Story Segmentation", ACM MULTIMEDIA CONFERENCE PROCEEDINGS, 1997, pages 381 - 389
ARIKI Y ET AL: "A TV NEWS RETRIEVAL SYSTEM WITH INTERACTIVE QUERY FUNCTION", PROCEEDINGS OF THE IFCIS INTERNATIONAL CONFERENCE ON COOPERATIVE INFORMATION SYSTEMS, COOPIS, 24 June 1997 (1997-06-24), XP000197416 *
CHRISTEL, M., STEVENS, S., KANADE, T. ET AL: "Techniques for the Creation and Exploration of Digital Video Libraries", MULTIMEDIA TOOLS AND APPLICATIONS, vol. 2, 1996, Boston, Kluwer Academic Publishers, XP002134742 *
HAUPTMANN A G ET AL: "Artificial intelligence techniques in the interface to a digital video library", HUMAN FACTORS IN COMPUTING SYSTEMS. CHI 97 EXTENDED ABSTRACTS, PROCEEDINGS OF 1997 ANNUAL INTERNATIONAL CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, ATLANTA, GA, USA, 22-27 MARCH 1997, 1997, New York, NY, USA, ACM, USA, pages 2 - 3, XP002134683, ISSN: SBN 0-8979-926-2 *
M. CHRISTEL ET AL.: "Techniques for the Creation and Exploration of Digital Video Libraries", MULTIMEDIA TOOLS AND APPLICATIONS, vol. 2, 1996
SMITH J R ET AL: "VISUALLY SEARCHING THE WEB FOR CONTENT", IEEE MULTIMEDIA,US,IEEE COMPUTER SOCIETY, vol. 4, no. 3, 1 July 1997 (1997-07-01), pages 12 - 20, XP000702058, ISSN: 1070-986X *
SMITH M A ET AL: "VIDEO SKIMMING AND CHARACTERIZATION THROUGH THE COMBINATION OF IMAGE AND LANGUAGE UNDERSTANDING TECHNIQUES", PROCEEDINGS OF THE IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION,US,LOS ALAMITOS, IEEE COMP. SOC. PRESS, vol. CONF. 16, 1997, pages 775 - 781, XP000776576, ISBN: 0-7803-4236-4 *
Y ARIKI ET AL.: "A TV News Retrieval System with Interactive Query Function", PROCEEDINGS OF THE IFCIS INTERNATIONAL CONFERENCE ON COOPERATIVE INFORMATION SYSTEMS. COOPIS, 24 June 1997 (1997-06-24)

Cited By (37)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
USRE47908E1 (en) 1991-12-23 2020-03-17 Blanding Hovenweep, Llc Ergonomic man-machine interface incorporating adaptive pattern recognition based control system
USRE48056E1 (en) 1991-12-23 2020-06-16 Blanding Hovenweep, Llc Ergonomic man-machine interface incorporating adaptive pattern recognition based control system
USRE49387E1 (en) 1991-12-23 2023-01-24 Blanding Hovenweep, Llc Ergonomic man-machine interface incorporating adaptive pattern recognition based control system
US6920281B1 (en) 2000-06-30 2005-07-19 Koninklijke Philips Electronics N.V. Video memory manager for use in a video recorder and method of operation
EP1191460A1 (fr) * 2000-09-21 2002-03-27 Swisscom AG Méthode et système de classification de fichiers audio et vidéo numériques
JP2002169819A (ja) * 2000-12-01 2002-06-14 Kanebo Ltd 動画情報配信方法およびそれに用いる動画情報配信システム
WO2002051140A2 (fr) * 2000-12-18 2002-06-27 Koninklijke Philips Electronics N.V. Appareil et procede de classification de programmes faisant appel a des indices observes dans les informations de transcription
WO2002051140A3 (fr) * 2000-12-18 2002-10-10 Koninkl Philips Electronics Nv Appareil et procede de classification de programmes faisant appel a des indices observes dans les informations de transcription
US7210157B2 (en) 2000-12-18 2007-04-24 Koninklijke Philips Electronics N.V. Apparatus and method of program classification using observed cues in the transcript information
WO2003010965A1 (fr) * 2001-07-26 2003-02-06 Koninklijke Philips Electronics N.V. Procede de facturation d'annonceurs a partir d'une commutation d'annonces publicitaires entre des canaux de television
EP1291790A3 (fr) * 2001-08-15 2006-01-18 Siemens Corporate Research, Inc. Groupement et classement automatique de contenu textuel
US7096179B2 (en) 2001-08-15 2006-08-22 Siemens Corporate Research, Inc. Text-based automatic content classification and grouping
EP1291790A2 (fr) * 2001-08-15 2003-03-12 Siemens Corporate Research, Inc. Groupement et classement automatique de contenu textuel
WO2003046761A3 (fr) * 2001-11-28 2004-02-12 Koninkl Philips Electronics Nv Systeme et procede de recherche d'informations associees a des sujets cibles
WO2003046761A2 (fr) * 2001-11-28 2003-06-05 Koninklijke Philips Electronics N.V. Systeme et procede de recherche d'informations associees a des sujets cibles
WO2003050718A3 (fr) * 2001-12-11 2004-05-06 Koninkl Philips Electronics Nv Systeme et procede d'extraction d'informations relatives a des personnes dans des programmes video
WO2003050718A2 (fr) * 2001-12-11 2003-06-19 Koninklijke Philips Electronics N.V. Systeme et procede d'extraction d'informations relatives a des personnes dans des programmes video
WO2005001715A1 (fr) * 2003-06-30 2005-01-06 Koninklijke Philips Electronics, N.V. Systeme et procede permettant de produire un condense multimedia de flux multimedias
US8209623B2 (en) 2003-12-05 2012-06-26 Sony Deutschland Gmbh Visualization and control techniques for multimedia digital content
JP2006033659A (ja) * 2004-07-21 2006-02-02 Sony Corp コンテンツ記録再生装置、コンテンツ記録再生方法及びそのプログラム
JP4586446B2 (ja) * 2004-07-21 2010-11-24 ソニー株式会社 コンテンツ記録再生装置、コンテンツ記録再生方法及びそのプログラム
WO2006103633A1 (fr) 2005-03-31 2006-10-05 Koninklijke Philips Electronics, N.V. Synthese de recits d'actualites composites
WO2008007279A3 (fr) * 2006-07-04 2008-05-08 Koninkl Philips Electronics Nv Procédé de substitution de contenu
WO2008007274A3 (fr) * 2006-07-04 2008-05-29 Koninkl Philips Electronics Nv Procédé de substitution de contenu
US8930984B2 (en) 2006-07-04 2015-01-06 Koninklijke Philips N.V. Method of content substitution
WO2008007279A2 (fr) * 2006-07-04 2008-01-17 Koninklijke Philips Electronics N.V. Procédé de substitution de contenu
US8671337B2 (en) 2007-03-27 2014-03-11 Sharp Laboratories Of America, Inc. Methods, systems and devices for multimedia-content presentation
EP2345251A4 (fr) * 2008-10-31 2012-04-11 Hewlett Packard Development Co Organisation de données vidéo
WO2011039773A2 (fr) * 2009-09-14 2011-04-07 Tata Consultancy Services Ltd. Système d'analyse de nouvelles télévisées pour canaux de diffusion multilingue
WO2011039773A3 (fr) * 2009-09-14 2011-06-16 Tata Consultancy Services Ltd. Système d'analyse de nouvelles télévisées pour canaux de diffusion multilingue
CN101894129A (zh) * 2010-05-31 2010-11-24 中国科学技术大学 基于在线视频分享网站结构及视频描述文本信息的视频主题发现的方法
CN102184235A (zh) * 2011-05-13 2011-09-14 广州星海传媒有限公司 一种基于机顶盒的数字电视节目推荐方法及系统
CN106550254A (zh) * 2015-09-23 2017-03-29 北京丰源星际传媒科技有限公司 影院放映电影贴片广告的监播方法及系统
CN106550254B (zh) * 2015-09-23 2020-02-18 北京丰源星际传媒科技有限公司 影院放映电影贴片广告的监播方法及系统
US20200159759A1 (en) * 2018-11-20 2020-05-21 Comcast Cable Communication, Llc Systems and methods for indexing a content asset
CN113360709A (zh) * 2021-05-28 2021-09-07 维沃移动通信(杭州)有限公司 短视频侵权风险的检测方法、装置和电子设备
CN113360709B (zh) * 2021-05-28 2023-02-17 维沃移动通信(杭州)有限公司 短视频侵权风险的检测方法、装置和电子设备

Also Published As

Publication number Publication date
CN1298522A (zh) 2001-06-06
KR100711948B1 (ko) 2007-05-02
JP2002533841A (ja) 2002-10-08
CN1116649C (zh) 2003-07-30
EP1057129A1 (fr) 2000-12-06
KR20010041194A (ko) 2001-05-15

Similar Documents

Publication Publication Date Title
US20050028194A1 (en) Personalized news retrieval system
KR100711948B1 (ko) 개인화된 비디오 분류 및 검색 시스템
KR100794152B1 (ko) 오디오/데이터/시각 정보 선택을 위한 방법 및 장치
KR100915847B1 (ko) 스트리밍 비디오 북마크들
KR101150748B1 (ko) 멀티미디어 스트림들의 멀티미디어 요약을 생성하기 위한시스템 및 방법
US7434247B2 (en) System and method for determining the desirability of video programming events using keyword matching
US20030093794A1 (en) Method and system for personal information retrieval, update and presentation
KR100965457B1 (ko) 퍼스널 프로파일에 기초한 콘텐츠의 증가
US20010049826A1 (en) Method of searching video channels by content
US20030093580A1 (en) Method and system for information alerts
US10171860B2 (en) System and method for creating, receiving and using interactive information
JP2008148077A (ja) 動画再生装置
EP1114554A1 (fr) Procede de selection automatique de canaux video
Smeaton et al. TV news story segmentation, personalisation and recommendation
Dimitrova et al. PNRS: personalized news retrieval system

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 99805318.X

Country of ref document: CN

AK Designated states

Kind code of ref document: A1

Designated state(s): CN JP KR

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE

WWE Wipo information: entry into national phase

Ref document number: 1999965522

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 1020007009264

Country of ref document: KR

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWP Wipo information: published in national office

Ref document number: 1999965522

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 1020007009264

Country of ref document: KR

WWG Wipo information: grant in national office

Ref document number: 1020007009264

Country of ref document: KR