US20060075361A1 - Video browser based on character relation - Google Patents

Video browser based on character relation Download PDF

Info

Publication number
US20060075361A1
US20060075361A1 US11284657 US28465705A US2006075361A1 US 20060075361 A1 US20060075361 A1 US 20060075361A1 US 11284657 US11284657 US 11284657 US 28465705 A US28465705 A US 28465705A US 2006075361 A1 US2006075361 A1 US 2006075361A1
Authority
US
Grant status
Application
Patent type
Prior art keywords
segment
segment group
highlights
relation
video
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US11284657
Inventor
Jung Song
Jin Lee
Original Assignee
Song Jung M
Lee Jin S
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/30Information retrieval; Database structures therefor ; File system structures therefor
    • G06F17/30781Information retrieval; Database structures therefor ; File system structures therefor of video data
    • G06F17/30817Information retrieval; Database structures therefor ; File system structures therefor of video data using information manually generated or using information not derived from the video content, e.g. time and location information, usage information, user ratings
    • GPHYSICS
    • G06COMPUTING; CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/30Information retrieval; Database structures therefor ; File system structures therefor
    • G06F17/30781Information retrieval; Database structures therefor ; File system structures therefor of video data
    • G06F17/30846Browsing of video data
    • G06F17/30852Browsing the internal structure of a single video sequence
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/102Programmed access in sequence to addressed parts of tracks of operating record carriers
    • G11B27/105Programmed access in sequence to addressed parts of tracks of operating record carriers of operating discs
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/11Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information not detectable on the record carrier
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/19Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
    • G11B27/28Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
    • G11B27/32Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier
    • G11B27/326Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on separate auxiliary tracks of the same or an auxiliary record carrier used signal is a video-frame or a video-field (P.I.P.)
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • G11B27/34Indicating arrangements

Abstract

A video data structure and a video browser for video browsing are disclosed. The data structure includes segment group information indicating that a first segment group represents highlights sharing at least two objects in a program's content. The segment group information is used to display the highlights represented by the first segment group. The present invention can be implemented to allow users to easily understand relations between characters and changes in relations between characters in a movie or a drama by displaying significant events in the corresponding relations and by displaying character relations in a tree structure.

Description

    CROSS REFERENCE TO RELATED APPLICATIONS
  • The present application is a continuation of co-pending application Ser. No. 09/645,613, filed Aug. 25, 2000, the entirety of which is incorporated by reference herein. This application claims the benefit of Korea Patent Application No. 35688/1999, filed Aug. 26, 1999.
  • BACKGROUND
  • 1. Field of the Invention
  • The present invention relates to a video browser, and more particularly to a video browser for browsing a video based on relations between characters.
  • 2. Background of the Related Art
  • Typically, users simply view movies and/or dramas as broadcasted through a TV or played at a movie theater. However, a user may wish to view a particular movie or drama at a particular time, or wish to view only a particular section of a movie or a drama. Accordingly, various techniques which enable selective watching of a movie/drama or sections of a movie/drama have been suggested.
  • In the related art, for example, various video data may be represented or classified into format portion, index portion, media portion, segment portion, target portion, and/or representation portion. Also, data on various characters or objects, such as a name of an object, position on the screen, or numeric data with relation to a segment of the video data in which the object appears, may be represented by the target and representation chunk. Accordingly, a user can select an object through a table and reproduce for display a particular segment where the object is shown in the video.
  • In another related art, various additional data of a video data are obtained before, during or after the production of the video data. Thereafter, an additional information table of the obtained data is composed and provided to users. Namely, the additional data table may include a position where an actor appears, a position where a character of the actor appears, and a position where stage properties appear, such that a scene can be reproduced as selected by a user through the additional data table. For example, if a user selects a stage property, information on the selected stage property, such as the manufacturer and price, may be displayed on a screen, and the user may be able to connect with the manufacturer or a seller of the stage property through a network connection.
  • In still another related art, recording information on each segment of a video in a video map has been suggested. That is, information, such as the degree of violence, the degree of adult content, the degree of importance of content, characters positions, and the degree of difficulty in understanding, may be indicated for each segment of a video in the video map. Thus, the user may set a degree of preference for one or more items of the video map, and only segments of the video meeting the set degree of preference would be reproduced, thereby limiting a display of particular contents to unauthorized viewers.
  • Similarly, other techniques in the related art as described above provide items simply arranged without any relation to the objects appearing in the movie or drama, based upon the selection of the user. However, the content of a movie or drama generally builds around relations between characters, places and events. For example, relations between characters may not change from beginning to the end of the story or may continuously vary. Moreover, since one or more characters relate to a specific character in the movie or drama, the browsing method in the related art substantially fails to provide an accurate understanding of the story of the movie or drama to the user.
  • Therefore, techniques in the related arts have disadvantages in that it is difficult to understand a video centering on relations among characters according to the development of events, changes of relations, and relations among characters and places as events develop.
  • SUMMARY
  • Accordingly, an object of the present invention is to solve at least the problems and disadvantages of the related art.
  • Another object of the present invention is to provide a video browser in which event segments showing changes in relations between characters are summarized and displayed based on a video data structure which includes characters in a video that significantly act on development of an event, and includes a constant and variable relations between characters.
  • A still another object of the present invention is to provide a video browser in which main events that significantly act on variable relations between characters are connected with corresponding variable relations based on a video data structure which includes characters in a video that significantly act on development of an event, and includes a constant and variable relations between characters.
  • A further object of the present invention is to provide a video browser in which a key frame of event segments which show a corresponding relation between characters is displayed based on a video data structure which includes characters in a video that significantly act on development of an event, and includes a constant and variable relations between characters.
  • A further object of the present invention is to provide a video browser in which a constant relation and changes in relations between characters are displayed depending on preset relation types based on a video data structure which includes characters in a video that significantly act on development of an event, and includes a constant and variable relations between characters.
  • Additional advantages, objects, and features of the invention will be set forth in part in the description which follows and in part will become apparent to those having ordinary skill in the art upon examination of the following or may be learned from practice of the invention. The objects and advantages of the invention may be realized and attained as particularly pointed out in the appended claims.
  • To achieve the objects and in accordance with the purposes of the invention, as embodied and broadly described herein, a video browsing system for browsing a video based on a data structure in which a constant and variable relations between characters of a video are connected with characters and corresponding event segments, comprises (a) browsing and displaying a video segment showing a constant relation between characters and variable relations between characters based on an event segment showing a corresponding constant relation and variable relation, and (b) displaying a video segment of a corresponding event segment in response to a user's selection.
  • The video browser of the present invention is characterized in that event segments showing a constant relation between characters and changes in the relations is summarized and displayed. Also, main events that significantly act on variable relations are connected with corresponding variable relations and characters.
  • Moreover, the video browser of the present invention is characterized in that constant relation and variable relation between the characters are browsed and displayed as a constant relation and a variable relation between characters selected by the user. The constant relation and variable relation between characters may also be browsed and displayed as a constant relation and a variable relation between every characters related with a selected character.
  • In still another video browser of the present invention, the constant relation and variable relation between characters are displayed in a tree structure. The video browser of the present invention may display a constant relation and variable relation between characters depending on preset relation types. The relation types may be set in semantic types such as “family relation,” “business relation,” and “social relation.”
  • Furthermore, the video browser of the present invention is characterized in that event segments showing constant and variable relations between characters are displayed as key frames. Finally, each key frame may show a corresponding relation between characters.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention will be described in detail with reference to the following drawings in which like reference numerals refer to like elements wherein:
  • FIG. 1 shows an example of a video data structure based on character relation according to the present invention;
  • FIG. 2 shows an example of a video data structure in which events showing character relations are summarized in accordance to the present invention;
  • FIG. 3 shows a video browser based on the video data structure of FIG. 2;
  • FIG. 4 shows a video data structure of an event segment showing variable relations between characters;
  • FIG. 5 shows a video browser based on the data structure of FIG. 4 in accordance with the present invention;
  • FIG. 6 shows an example of a video browser according to the present invention; and
  • FIG. 7 shows another example of a video browser according to the present invention.
  • DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
  • Reference will now be made in detail to the preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings.
  • FIG. 1 shows a video data structure based on character relations in accordance with the present invention for a video browser based on content. A video browser based on content is disclosed in U.S. Pat. No. 6,602,297, entitled “Motional Video Drowsing Data Structure and Browsing Method Therefor,” and is fully incorporated herein.
  • Referring to FIG. 1, a visual description scheme (DS) 101 is divided into a visualization DS 102, a syntactic structure DS 103, and a semantic structure DS 104. The visualization DS 102 is organized into a highlight view DS 105 for displaying a summarized data as a video highlight, and a key frame view DS 106 for displaying summarized data based on a key frame. Because, a video plot can be summarized briefly or with greater amounts of detail, the highlight view DS 105 enables a display of multi-levels of summarized data with a level 107, a view type 108 and a time DS 109. Similarly, the key frame view DS 106 enables a display of multi-levels of summarized data with a level 110 and a time DS 111.
  • Namely, the highlight view DS 105 is organized into a level 107 which has multiple levels of highlight data based, upon a degree of detail in summarizing a video. In such case, summary data in each level may include a corresponding period in the time DS 109 for use in highlight. The key frame view DS 106 is also organized into a level 110 which has multiple levels of summarized data based upon a degree of detail. In such case, summary data in each level may include a corresponding period in the time DS 111 for use as a key frame.
  • The syntactic structure DS 103 is used for displaying the actual video and includes actual video segments to be displayed. The syntactic structure DS 103 is organized into actual video segments in segment DSs 112, and corresponding temporal positions of video segments in the time DS 113.
  • The semantic structure DS 104 includes additional information describing a video, and is organized into an event DS 114 including event information, an object DS 117 including object information, and an event/object relation graph DS 119 including information describing relations between objects and places, and corresponding events to the relations. Namely, the event DS 114 describes events, and the object DS 117 describes objects such as characters and places. The event/object relation graph DS 119 describes a constant relation or changes in a relation between characters, a relation between an object and a place, or a relation between an object and an event.
  • Here, a constant relation means either a relation between characters that cannot change throughout a video, such as a parent to child relation, or a relation which is most representative of the relations between characters.
  • When an event is selected to display a video segment corresponding to the event, the event DS 114 of the semantic structure DS 104 is used. The event DS 114 is divided into a Reference to Segment 115 including information necessary for displaying a segment of a video corresponding to the selected event and an annotation DS 116 including information which connects events with actual positions of the events in a video and information for explaining events in a video. Particularly, the event DS 114 is used in displaying a selected event and connects the selected event with a position of the event in a video data through the Reference to Segment 115, and the contents of a selected event is annotated through the annotation DS 116.
  • The object DS 117 is organized into an object annotation DS 118 including information for describing objects such as characters or places. That is, the object DS 117 is used in displaying characters, and detailed information of the characters can be summarized or annotated through the annotation DS 118.
  • The event/object relation graph DS 119 includes information on a constant relation and variable relation between characters. The event/object relation graph DS 119 is organized into an entity relation 120 with a return which allows a display of character relations in a tree structure in which a relation may have subordinate relations. According to the present invention, constant relations between characters are placed in an upper level of the tree while changes of relations between the characters are placed in a lower level of the tree.
  • The entity relation 120 is also divided into a relation 121, a Reference to Object 124, and a Reference to Event 125. The relation 121 is organized into a name 122 including information on the titles of relations, and a type 123 including information on the nature of relations. For example, a nature of relation may be ‘family’ and a title of relation may be ‘spouse.’ The Reference to Object 124 connects related characters with each other and the Reference to Event 125 connects events which shows particular relations.
  • In the above video data structure, the notation above each data such as {0,1}, {0,*}, or {1,*} indicates the number of data for the corresponding data. For example, the notation of {0,1} for the visualization DS 102 indicates that the visual DS 101 can have zero or one visualization DS. On the other hand, the notation of {0,*} for the segment DS 112 indicates that the syntactic structure DS 103 may have from zero to any number of segment DS.
  • FIG. 2 shows a method of using a data structure of FIG. 1 to easily understand and browse a video based on relations between characters. Referring to FIG. 2, a representative relation (constant relation) between ‘character 1’ and ‘character 2’ is placed at the top of the tree structure, and variable relations (relation 1˜relation 4) between ‘character 1’ and ‘character 2’ are placed at the bottom of the tree structure. Also, main events which are significant in a change of relation in a variable relation are placed between variable relations. The main events are summarized video data and are displayed by highlight data.
  • Here, the events are connected through the Reference to Event 125 in the entity relation 120 and the video segments are connected through the Reference to Segment 115 in the event DS 114. Thereafter, an actual video segment Is displayed by connecting the time DS 113 of the segment DS 112 with the time DS 109 of the level 107 in the highlight DS 105 of the visual DS 102.
  • FIG. 3 shows an example screen of a video browser based on the video data structure of FIG. 2. Referring to FIG. 3, ‘character n’ of a video is displayed on a character screen 301, and a video is displayed on a main screen 302 according to a selection received through a user interface 303.
  • For example, if a user selects ‘character 1’ and ‘character 3’ from the character screen 301, main events which are significant in bringing about a change in the variable relations between ‘character 1’ and ‘character 3’ can be browsed based on the data structures of FIGS. 1 and 2. As a result, contents corresponding to the main events are summarized and displayed on the main screen 302.
  • FIG. 4 shows another method of using a video data structure including information on variable relations of characters and main events significant in the variable relations.
  • Referring to FIG. 4, a representative relation (constant relation) between ‘character 1’ and ‘character 2’ is placed at a top of a tree structure, and variable relations (relation 1˜relation 4) between ‘character 1’ and ‘character 2’ are placed at the bottom of the tree structure. Main events (event 1˜event 3) which are significant in bringing about a change in the variable relations are placed between the variable relations. The method of FIG. 4 is realized in such a manner that a main event segment which brings a new relation is connected with the variable relations through the Reference to Event 125 of the entity relation 120.
  • FIG. 5 shows a video browser in accordance with the present invention based on the data structure of FIG. 4. Referring to FIG. 5, characters in a video are displayed on a character screen 501, and connection between variable relations and events are displayed on a character relation-variable event screen 502. Also, an event segment corresponding to a variable relation can be reproduced and displayed on a main screen 503.
  • For example, if a viewer selects ‘character 1’ and ‘character 3,’ variable relations between ‘character 1’ and ‘character 3,’ and main events which are significant in bringing about a change in the variable relation are respectively displayed on the screen 502. At this time, a relation or event may be displayed by an key frame or annotation. Also, a video segment corresponding to an event such as ‘event 2’ selected from the screen 502 is reproduced and displayed on the main screen 503.
  • FIG. 6 shows another example screen of a video browser according to the present invention. Referring to FIG. 6, main characters of a video are displayed on a character screen 602. Also, characters having relations with a character selected from the character screen 601, and a constant relation and variable relations between the selected character and related characters are displayed on a relation screen 602.
  • At this time, a constant relation between the selected character and a related character is displayed on the top of a relation tree structure while variable relations between the selected character and the related character is displayed on the bottom of the tree structure.
  • Furthermore, a key frame of events significant in both the constant relation and variable relations is displayed on a main scene screen 603. Here, key frames of event segment corresponding to a type(s) of event may be displayed on the main scene screen according to a user selection through a selection screen 604. Namely, key frames of main event segments which directly show a selected relation or key frames of secondary event segments which indirectly show the selected relation, or both can be displayed according to the user selection. For example, if a selected relation is a husband and wife, a main event segment may be a video segment of a marriage ceremony while a secondary event segment may be a video segment of third parties discussing the marriage ceremony. An event segment corresponding to an event, for example a key frame, selected from the main scene screen 603 can thus be reproduced and displayed on the main screen 605.
  • For example, when a user selects ‘character 1’ from among the characters in the character screen 601, other characters ‘character 2’˜‘character 4’ related with ‘character 1’ are displayed on the relation screen 602. If ‘relation 2’ with ‘character 2’ is selected from the relation screen 602, event segments corresponding to ‘relation 2’ with ‘character 2’ is displayed on the main scene screen 603 as key frames. Also, a period of a video corresponding to ‘event 6’ selected from the main scene screen 603 is reproduced and displayed on the main screen 605.
  • At this time, the video browser of FIG. 6 can be implemented by setting a relation type in the data structure of FIG. 1. In other words, events connected with a relation may vary depending on whether the type 123 of the relation 121 is the direct or indirect type as discussed above. Also, the events are connected through the Reference to Event 125 of the entity relation 120 and event segments are connected through the Reference to Segment 115 of the event DS 114. Thereafter, an actual video segment is displayed by connecting the time DS 113 of the segment DS 112 with the time DS 111 of the level 110 in the key frame view DS 106 of the visual DS 102.
  • FIG. 7 shows still another example screen of a video browser according to the present invention. Referring to FIG. 7, characters of a video are displayed on a character screen 701; and character having relations with a selected character, and a constant relation and a variable relation between the selected character and related characters are displayed on a relation screen 702. Here, the constant and variable relations may categorized into different natures of relations and relations corresponding to a particular nature(s) of relation selected through a selection screen 703 may be displayed in the relation screen. The nature of relation may be a family relation, a business relation, or a social relation.
  • Thus, a key frame of events significant in the selected nature(s) of relation is displayed on a main scene screen 704. A video segment corresponding to the selected relation and event is reproduced and displayed, on a main screen 705, For example, when a user selects ‘character 1’ from characters displayed on the character screen 701, characters (‘character 2’˜‘character 4’) ‘aving a social relation with “character 1” may be displayed and an event segment of ‘event 6’ corresponding to ‘social relation 2’ with ‘character 2’ can be reproduced and displayed on the main screen 705.
  • As in FIG. 6, a constant relation, between a selected character and a related character is displayed on the top of a relation tree structure while variable relations between the selected character and the related character is displayed on the bottom of the tree structure.
  • The video browser of FIG. 7 can be realized by setting a relation type 123 of the relation data 121 in the data structure of FIG. 1. In other words, events related with the relation data 121 may vary depending on whether the relation type 123 of the relation data 121 is a ‘family relation,’ a ‘business relation,’ or a ‘social relation.’ As a result, browsing of FIG. 7 can be performed. At this time, the events are mutually connected through the reference to event 125 of the real relation 120. Segments are mutually connected through the reference to segment 115 of the event DS 114. Then, the time DS 113 of the segment DS 112 is connected with the time DS 111 of the level 107 displayed in the key frame DS 106 of the visual DS 102.
  • As aforementioned, the video browser based on a character relation according to the present invention has the following advantages. The event periods of the variable relation between the characters are summarized and displayed, and the main events concerned in the variable relation between the characters are connected between the variable relations and then displayed.
  • Furthermore, the data structure for browsing a video is based on a relation between characters and a variable relation between them. A key frame of an event period showing a corresponding relation between the characters clearly or directly, or a key frame of an event period showing it with hint or indirectly can be displayed based on the data structure. Also, the relations between the characters are divided into a family relation, a social relation, and an emotion relation to be selectively provided to a user, so that video browsing can be performed based on the relations between the characters.
  • Accordingly, contents of the video can be browsed based on characters that significantly act on development of a story of a movie or drama, the relation between the characters, and the variable relation between them. After all, the user can easily browse the video based on characters, event and relation which are concerned in development of a story and important for contents of a real video. Also, the user can understand and browse a desired video in various types based on the video data structure which represents characters and their relation.
  • Furthermore, the present invention may be applied to a VOD system in the broadcasting field so that the user can view a desired part. This could lead to an advantage that the user can reproduce and view the desired part within short time and to effective utilization of the network source. Also, the present invention may be applied to a video player for home use and a broadcasting video player so that the video browsing environment can be provided to easily browse a desired part of a movie or drama stored in the video player.
  • The foregoing embodiments are merely exemplary and are not to be construed as limiting the present invention. The present teachings can be readily applied to other types of apparatuses. The description of the present invention is intended to be illustrative, and not to limit the scope of the claims. Many alternatives, modifications, and variations will be apparent to those skilled in the art.

Claims (24)

  1. 1. A method for processing multimedia data, the method comprising:
    receiving a content description for a multimedia program, the content description including segment group information identifying a first segment group that includes a plurality of segments from the multimedia program, the segment group information indicating that the first segment group represents highlights sharing at least two objects in the content of the multimedia program; and
    using the segment group information to display the highlights represented by the first segment group.
  2. 2. The method of claim 1, wherein the first segment group represents highlights sharing at least two characters in the content of the multimedia program.
  3. 3. The method of claim 1, wherein the content description specifies a duration for a running time of the highlights represented by the first segment group.
  4. 4. The method of claim 1, wherein the segment group information identifies a second segment group as an alternative group that represents highlights from the content of the multimedia program.
  5. 5. The method of claim 4, wherein the segment group information specifies different durations for the first and second segment groups.
  6. 6. The method of claim 4, wherein the first and second groups represent highlights with different levels of detail.
  7. 7. The method of claim 1, wherein the content description includes segment information.
  8. 8. The method of claim 7, wherein the segment information specifies a respective starting point for each segment in the first segment group.
  9. 9. A system for processing multimedia data, the system comprising:
    a display device; and
    data processing apparatus configured to perform operations comprising:
    receiving a content description for a multimedia program, the content description including segment group information identifying a first segment group that includes a plurality of segments from the multimedia program, the segment group information indicating that the first segment group represents highlights sharing at least two objects in the content of the multimedia program; and
    using the segment group information to display the highlights represented by the first segment group on the display device.
  10. 10. The system of claim 9, wherein the first segment group represents highlights sharing at least two characters in the content of the multimedia program.
  11. 11. The system of claim 9, wherein the content description specifies a duration for a running time of the highlights represented by the first segment group.
  12. 12. The system of claim 9, wherein the segment group information identifies a second segment group as an alternative group that represents highlights from the content of the multimedia program.
  13. 13. The system of claim 12, wherein the segment group information specifies different durations for the highlights represented by the first and second segment groups.
  14. 14. The system of claim 12, wherein the first and second groups represent highlights with different levels of detail.
  15. 15. The system of claim 9, wherein the content description includes segment information.
  16. 16. The system of claim 15, wherein the segment information specifies a respective starting point for each segment in the first segment group.
  17. 17. A computer program product for processing multimedia data, the computer program comprising instructions to cause data processing apparatus to perform operations comprising:
    receiving a content description for a multimedia program, the content description including segment group information identifying a first segment group that includes a plurality of segments from the multimedia program, the segment group information indicating that the first segment group represents highlights sharing at least two objects in the content of the multimedia program; and
    using the segment group information to display the highlights represented by the first segment group.
  18. 18. The computer program product of claim 17, wherein the first segment group represents highlights sharing at least two characters in the content of the multimedia program.
  19. 19. The computer program product of claim 17, wherein the content description specifies a duration for a running time of the highlights represented by the first segment group.
  20. 20. The computer program product of claim 17, wherein the segment group information identifies a second segment group as an alternative group that represents highlights from the content of the multimedia program.
  21. 21. The computer program product of claim 20, wherein the segment group information specifies different durations for the first and second segment groups.
  22. 22. The computer program product of claim 20, wherein the first and second groups represent highlights with different levels of detail.
  23. 23. The computer program product of claim 17, wherein the content description includes segment information.
  24. 24. The computer program product of claim 23, wherein the segment information specifies a respective starting point for each segment in the first segment group.
US11284657 1999-08-26 2005-11-22 Video browser based on character relation Abandoned US20060075361A1 (en)

Priority Applications (4)

Application Number Priority Date Filing Date Title
KR19990035688A KR100370247B1 (en) 1999-08-26 1999-08-26 Video browser based on character relation
KR35688/1999 1999-08-26
US09645613 US7509581B1 (en) 1999-08-26 2000-08-25 Video browser based on character relation
US11284657 US20060075361A1 (en) 1999-08-26 2005-11-22 Video browser based on character relation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11284657 US20060075361A1 (en) 1999-08-26 2005-11-22 Video browser based on character relation

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US09645613 Continuation US7509581B1 (en) 1999-08-26 2000-08-25 Video browser based on character relation

Publications (1)

Publication Number Publication Date
US20060075361A1 true true US20060075361A1 (en) 2006-04-06

Family

ID=36127136

Family Applications (2)

Application Number Title Priority Date Filing Date
US09645613 Active 2021-01-17 US7509581B1 (en) 1999-08-26 2000-08-25 Video browser based on character relation
US11284657 Abandoned US20060075361A1 (en) 1999-08-26 2005-11-22 Video browser based on character relation

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US09645613 Active 2021-01-17 US7509581B1 (en) 1999-08-26 2000-08-25 Video browser based on character relation

Country Status (3)

Country Link
US (2) US7509581B1 (en)
KR (1) KR100370247B1 (en)
WO (1) WO2001015016A1 (en)

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20050064150A (en) * 2003-12-23 2005-06-29 엘지전자 주식회사 Method for managing and reproducing a menu information of high density optical disc
KR20050066265A (en) * 2003-12-26 2005-06-30 엘지전자 주식회사 Method for managing and reproducing a menu information of high density optical disc
KR20050066264A (en) * 2003-12-26 2005-06-30 엘지전자 주식회사 Method for managing and reproducing a menu information of high density optical disc
WO2006025544A1 (en) * 2004-09-01 2006-03-09 Fujitsu Ten Limited Broadcasting receiver
US7797328B2 (en) * 2004-12-21 2010-09-14 Thomas Lane Styles System and method of searching for story-based media
US7912841B2 (en) * 2006-09-13 2011-03-22 I. Know Nv. Data processing based on data linking elements
US8584044B2 (en) * 2007-11-16 2013-11-12 Microsoft Corporation Localized thumbnail preview of related content during spatial browsing
US9304992B2 (en) * 2012-07-11 2016-04-05 Cellco Partnership Story element indexing and uses thereof
USD773488S1 (en) * 2012-09-04 2016-12-06 Huawei Technologies Co., Ltd. Display screen with graphical user interface for viewing and installing applications in an electronic mall
USD770486S1 (en) 2014-01-24 2016-11-01 Microsoft Corporation Display screen with graphical user interface
USD749111S1 (en) 2014-01-24 2016-02-09 Microsoft Corporation Display screen with graphical user interface
USD764525S1 (en) 2014-08-28 2016-08-23 Microsoft Corporation Display screen with graphical user interface
USD764524S1 (en) 2014-08-28 2016-08-23 Microsoft Corporation Display screen with graphical user interface
USD794647S1 (en) * 2014-11-21 2017-08-15 General Electric Company Display screen or portion thereof with graphical user interface
KR20160090593A (en) * 2015-01-22 2016-08-01 한국전자통신연구원 Apparatus and method for providing primitive visual knowledge
WO2016123788A1 (en) * 2015-02-06 2016-08-11 Empire Technology Development Llc Rhythm based multimedia generator
USD792902S1 (en) * 2016-04-22 2017-07-25 Cnh Industrial America Llc Display screen or portion thereof with graphical user interface

Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5574845A (en) * 1994-11-29 1996-11-12 Siemens Corporate Research, Inc. Method and apparatus video data management
US5812945A (en) * 1995-12-22 1998-09-22 Pitney Bowes Inc. Metered payment cellular telephone communication system
US5956026A (en) * 1997-12-19 1999-09-21 Sharp Laboratories Of America, Inc. Method for hierarchical summarization and browsing of digital video
US5982364A (en) * 1993-03-12 1999-11-09 Quantel, Ltd. Video processing system including frame position indicators
US6169547B1 (en) * 1993-02-23 2001-01-02 Hitachi, Ltd. Method for displaying an icon of media data
US6236395B1 (en) * 1999-02-01 2001-05-22 Sharp Laboratories Of America, Inc. Audiovisual information management system
US6278446B1 (en) * 1998-02-23 2001-08-21 Siemens Corporate Research, Inc. System for interactive organization and browsing of video
US6323911B1 (en) * 1995-10-02 2001-11-27 Starsight Telecast, Inc. System and method for using television schedule information
US6400996B1 (en) * 1999-02-01 2002-06-04 Steven M. Hoffberg Adaptive pattern recognition based control system and method
US6469711B2 (en) * 1996-07-29 2002-10-22 Avid Technology, Inc. Graphical user interface for a video editing system
US6492998B1 (en) * 1998-12-05 2002-12-10 Lg Electronics Inc. Contents-based video story browsing system
US6546188B1 (en) * 1998-01-16 2003-04-08 Sony Corporation Editing system and editing method
US6602297B1 (en) * 1998-12-05 2003-08-05 Lg Electronics Inc. Motional video browsing data structure and browsing method therefor
US6608633B1 (en) * 1998-10-30 2003-08-19 Sony Corporation Visual display of categorical information through visual factors such as scale and location
US6611277B1 (en) * 1999-10-29 2003-08-26 Sony Corporation Information processing apparatus and method as well as medium
US6629097B1 (en) * 1999-04-28 2003-09-30 Douglas K. Keith Displaying implicit associations among items in loosely-structured data sets
US6678689B2 (en) * 1999-12-30 2004-01-13 Lg Electronics Inc. Multimedia structure and method for browsing multimedia with defined priority of multimedia segments and semantic elements
US20070157225A1 (en) * 1996-09-27 2007-07-05 Matsushita Electric Industrial Co., Ltd. Method and apparatus for receiving and displaying coupon information

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5821945A (en) * 1995-02-03 1998-10-13 The Trustees Of Princeton University Method and apparatus for video browsing based on content and structure

Patent Citations (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6169547B1 (en) * 1993-02-23 2001-01-02 Hitachi, Ltd. Method for displaying an icon of media data
US5982364A (en) * 1993-03-12 1999-11-09 Quantel, Ltd. Video processing system including frame position indicators
US5574845A (en) * 1994-11-29 1996-11-12 Siemens Corporate Research, Inc. Method and apparatus video data management
US6323911B1 (en) * 1995-10-02 2001-11-27 Starsight Telecast, Inc. System and method for using television schedule information
US5812945A (en) * 1995-12-22 1998-09-22 Pitney Bowes Inc. Metered payment cellular telephone communication system
US6469711B2 (en) * 1996-07-29 2002-10-22 Avid Technology, Inc. Graphical user interface for a video editing system
US20070157225A1 (en) * 1996-09-27 2007-07-05 Matsushita Electric Industrial Co., Ltd. Method and apparatus for receiving and displaying coupon information
US5956026A (en) * 1997-12-19 1999-09-21 Sharp Laboratories Of America, Inc. Method for hierarchical summarization and browsing of digital video
US6546188B1 (en) * 1998-01-16 2003-04-08 Sony Corporation Editing system and editing method
US6278446B1 (en) * 1998-02-23 2001-08-21 Siemens Corporate Research, Inc. System for interactive organization and browsing of video
US6608633B1 (en) * 1998-10-30 2003-08-19 Sony Corporation Visual display of categorical information through visual factors such as scale and location
US6492998B1 (en) * 1998-12-05 2002-12-10 Lg Electronics Inc. Contents-based video story browsing system
US6602297B1 (en) * 1998-12-05 2003-08-05 Lg Electronics Inc. Motional video browsing data structure and browsing method therefor
US6400996B1 (en) * 1999-02-01 2002-06-04 Steven M. Hoffberg Adaptive pattern recognition based control system and method
US6236395B1 (en) * 1999-02-01 2001-05-22 Sharp Laboratories Of America, Inc. Audiovisual information management system
US6629097B1 (en) * 1999-04-28 2003-09-30 Douglas K. Keith Displaying implicit associations among items in loosely-structured data sets
US6611277B1 (en) * 1999-10-29 2003-08-26 Sony Corporation Information processing apparatus and method as well as medium
US6678689B2 (en) * 1999-12-30 2004-01-13 Lg Electronics Inc. Multimedia structure and method for browsing multimedia with defined priority of multimedia segments and semantic elements

Also Published As

Publication number Publication date Type
WO2001015016A1 (en) 2001-03-01 application
US7509581B1 (en) 2009-03-24 grant
KR20010019342A (en) 2001-03-15 application
KR100370247B1 (en) 2003-01-29 grant

Similar Documents

Publication Publication Date Title
US6577346B1 (en) Recognizing a pattern in a video segment to identify the video segment
US8832742B2 (en) Systems and methods for acquiring, categorizing and delivering media in interactive media guidance applications
US6442755B1 (en) Electronic program guide using markup language
US20090019488A1 (en) System and method for providing personal content recommendations
US20100114857A1 (en) User interface with available multimedia content from multiple multimedia websites
US20100299701A1 (en) Media content retrieval system and personal virtual channel
US20080086747A1 (en) Systems and methods for acquiring, categorizing and delivering media in interactive media guidance applications
US20120042334A1 (en) Display apparatus and display method thereof
US20070220554A1 (en) Recommended recording and downloading guides
US20080104127A1 (en) Presenting media guidance search results based on relevancy
US6236395B1 (en) Audiovisual information management system
US20080104058A1 (en) Presenting media guidance search results based on relevancy
US20020097265A1 (en) User interface for collecting viewer ratings of media content and facilitating adaption of content recommenders
US20050125823A1 (en) Promotional philosophy for a video-on-demand-related interactive display within an interactive television application
US20080104057A1 (en) Presenting media guidance search results based on relevancy
US20090307658A1 (en) Methods and apparatus for rendering user interfaces and display information on remote client devices
US20020012526A1 (en) Digital video reproduction method, digital video reproducing apparatus and digital video recording and reproducing apparatus
US20080172689A1 (en) Authorable content rating system
US20060215993A1 (en) Broadcast program recorder
US20050120369A1 (en) System and method for filtering content
US20010020981A1 (en) Method of generating synthetic key frame and video browsing system using the same
US20060090182A1 (en) Method and system for multimedia advertising
US20030122861A1 (en) Method, interface and apparatus for video browsing
US20120078953A1 (en) Browsing hierarchies with social recommendations
US20060026638A1 (en) Maintaining a graphical user interface state that is based on a selected type of content