US20020013783A1 - Generating semantic descriptions for content data from component semantic descriptions stored remotely from the content data - Google Patents

Generating semantic descriptions for content data from component semantic descriptions stored remotely from the content data Download PDF

Info

Publication number
US20020013783A1
US20020013783A1 US09/809,578 US80957801A US2002013783A1 US 20020013783 A1 US20020013783 A1 US 20020013783A1 US 80957801 A US80957801 A US 80957801A US 2002013783 A1 US2002013783 A1 US 2002013783A1
Authority
US
United States
Prior art keywords
semantic
descriptions
description
component
recited
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US09/809,578
Inventor
Hawley Rising
Ali Tabatabai
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sony Corp
Sony Electronics Inc
Original Assignee
Sony Corp
Sony Electronics Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sony Corp, Sony Electronics Inc filed Critical Sony Corp
Priority to US09/809,578 priority Critical patent/US20020013783A1/en
Assigned to SONY ELECTRONICS, INC., SONY CORPORATION OF JAPAN reassignment SONY ELECTRONICS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RISING, HAWLEY, III
Publication of US20020013783A1 publication Critical patent/US20020013783A1/en
Assigned to SONY ELECTRONICS, INC., SONY CORPORATION reassignment SONY ELECTRONICS, INC. CORRECTIVE ASSIGNMENT TO CORRECT THE NAME OF THE ASSIGNEE. FILED ON 03-14-2001, RECORDED ON REEL 011624 FRAME 0929 ASSIGNOR HEREBY CONFIRMS THE ASSIGNMENT OF THE ENTIRE INTEREST. Assignors: RISING, HAWLEY III
Assigned to SONY CORPORATION, SONY ELECTRONICS INC. reassignment SONY CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: TABATABAI, ALI
Assigned to SONY CORPORATION reassignment SONY CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: SONY CORPORATION OF JAPAN
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/40Information retrieval; Database structures therefor; File system structures therefor of multimedia data, e.g. slideshows comprising image and additional audio data
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/31Indexing; Data structures therefor; Storage structures
    • G06F16/316Indexing structures
    • G06F16/319Inverted lists
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/36Creation of semantic tools, e.g. ontology or thesauri
    • G06F16/367Ontology

Definitions

  • the present invention generally relates to the field of semantic descriptions for content data. More particularly, the present invention relates to the field of methods and systems for implementing powerful and flexible semantic description tools to describe the underlying meaning of the content data.
  • the MPEG-7 “Multimedia Content Description Interface” standard which is being developed by the Moving Pictures Expert Group (MPEG) focuses, unlike the preceding MPEG standards (e.g., MPEG-1, MPEG-2, and MPEG-4), on representing information about the content data, not the content data itself.
  • MPEG-7 “Multimedia Content Description Interface” standard which is being developed by the Moving Pictures Expert Group (MPEG) focuses, unlike the preceding MPEG standards (e.g., MPEG-1, MPEG-2, and MPEG-4), on representing information about the content data, not the content data itself.
  • MPEG-7 MPEG-7 “Multimedia Content Description Interface” standard which is being developed by the Moving Pictures Expert Group (MPEG) focuses, unlike the preceding MPEG standards (e.g., MPEG-1, MPEG-2, and MPEG-4), on representing information about the content data, not the content data itself.
  • MPEG-7 seeks to provide a simple, flexible, interoperable solution to the problems of indexing, searching, and retrieving content data.
  • MPEG-7 aims to standardize a core set of Descriptors that can be used to describe the various features of the content data; pre-defined structures of Descriptors and their relationships, called Description Schemes; a language to define Description Schemes and Descriptors, called the Description Definition Language (DDL); and coded representations of descriptions to enable efficient storage and fast access.
  • the DDL is being based on XML Schema.
  • the MPEG-7 descriptions (a set of instantiated Description Schemes) are linked to the content data itself to allow fast and efficient searching for material of a users interest.
  • MPEG-7 intends to describe content data regardless of storage, coding, display, transmission, medium, or technology.
  • MPEG-7 addresses a wide variety of media types including: still pictures, graphics, 3D models, audio, speech, video, and any combination thereof (e.g., multimedia presentations, scenarios, etc.).
  • Examples of content data within the MPEG-7 standard include an MPEG-4 data stream; a video tape; a CD containing music, sound, or speech; a picture printed on paper, and an interactive multimedia installation on the Web (i.e., the Internet).
  • the MPEG-7 standard includes different types of Descriptors and Description Schemes. Some Descriptors and Description Schemes describe what is in the content data in terms of syntactic structure, color histogram, shape of an object, texture, motion, pitch, rhythm, etc.
  • semantic Description Schemes describe the underlying meaning or understanding of the content data.
  • a goal, advertisement, and Madonna are examples of a semantic description (an instantiated semantic Description Scheme).
  • Other examples of semantic descriptions includes a storyline for a movie (i.e., content data), a description of a scene in the movie, a description of an image, a description of a piece of music, etc.
  • the semantic description is based on the underlying meaning of the content data.
  • the semantic description is expressed with words.
  • computer systems or other computational systems are not able to usefully manipulate (e.g., create, exchange, retrieve, etc.) semantic descriptions expressed with only words.
  • a computer system or other computational system can usefully manipulate semantic descriptions having structure. For example, it is not sufficient to describe the movie Zorro as having the entities Zorro, Zorro's girlfriend, a bad guy, a first sword fight, a second sword fight, etc. Relationships between these entities are needed, hence providing the structure.
  • categorical structure tools such as a category, a graph, a functor, and a natural transformation were needed to create any type or kind of semantic description.
  • the semantic entity tools and the categorical structure tools facilitate creation of a semantic mosaic description for content data.
  • the semantic mosaic description is formed from multiple semantic descriptions. These semantic descriptions are integrated with each other such that each semantic description is modified at a local level within localized regions without substantially changing each semantic description outside these localized regions.
  • the semantic mosaic description facilitates navigation or browsing through the multiple semantic descriptions and the content data.
  • the semantic entity tools and the categorical structure tools facilitate creation of a semantic description for content data using multiple component semantic descriptions stored remotely from the content data.
  • Reference information is associated with the content data, whereas the reference information includes the identity of the component semantic descriptions needed to form the semantic description, the location of these component semantic descriptions, and the manner of processing these component semantic descriptions to form the semantic description.
  • the component semantic descriptions identified in the reference information are retrieved (e.g., from a location on a network, a control dictionary, etc.). Then, the semantic description is formed in the manner specified in the reference information using the component semantic descriptions.
  • the semantic description does not have to be stored in a discrete location, saving storage resources and promoting re-use of component semantic descriptions.
  • FIG. 1 illustrates an exemplary computer system in which the present invention can be practiced.
  • FIG. 2 illustrates semantic entity tools and categorical structure tools in accordance with an embodiment of the present invention.
  • FIG. 4 illustrates numerous mental spaces in accordance with an embodiment of the present invention, showing creation of a blend mental space by integrating or blending input mental space1 and input mental space2.
  • FIG. 5A illustrates a morphism in accordance with an embodiment of the present invention.
  • FIG. 5B illustrates a first functor, a second functor, a first category, and a second category in accordance with an embodiment of the present invention.
  • FIG. 5D illustrates a graph in accordance with an embodiment of the present invention.
  • FIG. 6 illustrates a semantic descriptions and a semantic description1 in accordance with an embodiment of the present invention.
  • FIG. 7 illustrates a semantic mosaic description based on the semantic description1 of FIG. 6 and the semantic description2 of FIG. 6.
  • FIG. 8 illustrates formation of a semantic description for content data using multiple component semantic descriptions stored in locations on a network in accordance with an embodiment of the present invention.
  • FIG. 9 illustrates formation of a semantic description for content data using multiple component semantic descriptions stored in a control dictionary in accordance with an embodiment of the present invention.
  • FIG. 10 illustrates a flow chart showing a method of forming a semantic description for content data using multiple component semantic descriptions stored remotely from the content data in accordance with an embodiment of the present invention.
  • FIG. 1 An exemplary computer system 100 is shown in FIG. 1.
  • FIG. 1 illustrates an exemplary computer system 100 on which embodiments of the present invention may be practiced. It is appreciated that the computer system 100 of FIG. 1 is exemplary only and that the present invention can operate within a number of different computer systems including general-purpose computer systems and embedded computer systems.
  • Computer system 100 includes an address/data bus 110 for communicating information, a central processor 101 coupled with bus 110 for processing information and instructions, a volatile memory 102 (e.g., random access memory RAM) coupled with the bus 110 for storing information and instructions for the central processor 101 and a non-volatile memory 103 (e.g., read only memory ROM) coupled with the bus 110 for storing static information and instructions for the processor 101 .
  • Exemplary computer system 100 also includes a data storage device 104 (“disk subsystem”) such as a magnetic or optical disk and disk drive coupled with the bus 110 for storing information and instructions.
  • Data storage device 104 can include one or more removable magnetic or optical storage media (e.g., diskettes, tapes) which are computer readable memories.
  • Memory units of computer system 100 include volatile memory 102 , non-volatile memory 103 and data storage device 104 .
  • Exemplary computer system 100 can further include an optional signal generating device 108 (e.g., a network interface card “NIC”) coupled to the bus 110 for interfacing with other computer systems.
  • an optional alphanumeric input device 106 including alphanumeric and function keys coupled to the bus 110 for communicating information and command selections to the central processor 101 .
  • Exemplary computer system 100 also includes an optional cursor control or directing device 107 coupled to the bus 110 for communicating user input information and command selections to the central processor 101 .
  • An optional display device 105 can also be coupled to the bus 110 for displaying information to the computer user.
  • Display device 105 may be a liquid crystal device, other flat panel display, cathode ray tube, or other display device suitable for creating graphic images and alphanumeric characters recognizable to the user.
  • Cursor control device 107 allows the user to dynamically signal the two-dimensional movement of a visible symbol (cursor) on a display screen of display device 105 .
  • cursor control device 107 are known in the art including a trackball, mouse, touch pad, joystick or special keys on alphanumeric input device 106 capable of signaling movement of a given direction or manner of displacement.
  • a cursor can be directed and/or activated via input from alphanumeric input device 106 using special keys and keysequence commands.
  • the present invention is applicable to the MPEG-7 standard or to any other application which uses semantic descriptions.
  • FIG. 2 illustrates semantic entity tools 210 and categorical structure tools 220 in accordance with an embodiment of the present invention.
  • emphasis and focus is placed on determining what is necessary and needed to create any type or kind of semantic description for content data in various applications such as MPEG-7.
  • numerous semantic description tools 210 and 220 are selected. These semantic description tools 210 and 220 provide sufficient flexibility and power to create any type or kind of semantic description.
  • Numerous semantic entity tools 211 - 217 and numerous categorical structure tools 221 - 227 were identified as necessary and needed to create any type or kind of semantic description.
  • Semantic entity tools 211 - 217 are tools that represent entities in a semantic description.
  • Categorical structure tools 221 - 227 are tools that represent categorical structures of semantic entities 211 - 217 and relations among these categorical structures.
  • the semantic entity tools 211 - 217 and the categorical structure tools 221 - 227 are implemented as Description Schemes.
  • the semantic entity tools 210 include core semantic entities 211 - 214 , constructed semantic entities 216 - 217 , and a context semantic entity 215 .
  • the core semantic entities 211 - 214 include a semantic object 211 , a semantic state 212 , a semantic event 213 , and a semantic episode 214 .
  • the context semantic entity 215 includes a frame 215 .
  • the constructed semantic entities 216 - 217 include a mental space 216 and a descriptive structure 217 .
  • each constructed semantic entity can include a core semantic entity 211 , a context semantic entity 215 , and relationships among these.
  • the categorical structure tools 220 include a relation 221 , a morphism 222 , a graph 223 , a category 224 , a functor 225 , a natural transformation 226 , and a characteristic function 227 .
  • a semantic object or object 211 in a semantic description is derived from a physical object or an abstraction of the physical object.
  • the semantic object 211 describes a physical or abstract object semantically.
  • Physical objects have spatial contiguity, and temporal duration. Physical objects are described in various ways.
  • the change in a semantic object 211 over time, or the particular circumstances, or type of a generic semantic object 211 are described with reference to attributes, which are qualities of the semantic object 211 . The collection of these qualities changes over time, and can be called the semantic state or state 212 of the semantic object 211 .
  • semantic objects 211 have semantic states 212 .
  • Physical objects are frequently divisible.
  • the subsets of the material of a physical object can be physical objects in their own right. These subsets can be referred as physical subobjects.
  • semantic objects 211 can have semantic subobjects.
  • the collection of semantic subobjects of a given semantic object 211 , or the collection of semantic subobjects of a collection of semantic objects 211 admits a partial order, by inclusion.
  • semantic states 212 are frequently complex, containing more than a single attribute, they can have subcollections. These subcollections can be semantic states 212 if these subcollections have semantic meaning. Thus, semantic states 212 can have semantic substates.
  • a change in semantic state 212 is a semantic event or event 213 . Since, as was remarked above, semantic states 212 are complex, a semantic event 213 may likewise be complex, since the semantic event 213 may indicate the change in a large number of attributes. Consequently, if such a set of attributes admits a subset with semantic meaning, and that subset can change independently from the rest, a semantic event 213 can have semantic subevents.
  • a semantic description formed with a semantic object 211 may or may not describe semantic subobjects, semantic states 212 , semantic substates, semantic events 213 , or semantic subevents. More importantly, the semantic description may contain relationships other than inclusion of parts.
  • a semantic episode or episode 214 denotes an inclusive semantic description of what transpires over a period of time, from some (possibly implied) starting time to (also possibly implied) ending time, with a duration greater than zero.
  • a semantic episode 214 can be a temporal designation with semantic meaning. If there are time periods of shorter duration between the start of the semantic episode 214 and the end of the semantic episode 214 , which have semantic meaning, these may be called semantic subepisodes.
  • the semantic description includes relationships.
  • One relationship that has already been seen and holds for all of the above identified semantic entities 211 - 214 is that of inclusion, in the manner of a semantic subobject, semantic subevent, semantic subepisode, or semantic substate.
  • the lists of relationships between such semantic entities 211 - 214 can be quite long.
  • the formal definitions of two mathematical concepts, which will facilitate them, are the definition of a relation 221 and the definition of a morphism or mapping 222 , which are illustrated in FIG. 2.
  • a relation on a group of mathematical objects is a subset of the formal cartesian product of the mathematical objects.
  • a binary relation is a subset of the set of ordered pairs of mathematical objects.
  • Inclusion is a partial order.
  • containment is a partial order (i.e., when one mathematical object is contained in another). Containment and inclusion are not the same: One would hardly say that a fish is part of a fish tank, but it is likely to be found there.
  • a morphism or mapping 222 is an assignment consisting of ordered pairs from a set called the domain and a set called the codomain. It can have more distinction than that, for instance, a function is a mapping where the codomain is the real (or complex) numbers, and for each element a of the domain, there is exactly one element b of the codomain.
  • a relationship between mathematical objects is either a relation 221 or a morphism/mapping 222 .
  • relations 221 can be expressed as compositions of mappings, (a and b map to (a,b) which maps via the characteristic function 227 of the subset mentioned above to either true or false.
  • a generalization of the characteristic function 227 maps to a discrete set, and is called a subobject classifier.
  • a relationship is a morphism or mapping 222 .
  • semantic objects 211 , semantic events 212 , semantic states 213 , and semantic episodes 214 can all have relationships. It is also possible to have relationships between these semantic entities 211 - 214 , the most obvious being between semantic objects 211 , semantic events 213 , and semantic states 212 , but semantic episodes 214 may sometimes be effectively described by relationships as well.
  • semantic events 213 are described as a change in semantic state 212 , a semantic state 212 being a collection of attributes for a semantic object 211 .
  • a relationship is a morphism or mapping 221 . Mappings may be parametrized.
  • a change in the parameters of a mapping between two of the above identified semantic entities 211 - 214 fits well as a semantic event 213 .
  • semantic entities 211 - 214 of the above categories it is possible for semantic entities 211 - 214 of the above categories to be described by a complex set of mappings. This set is also a relationship.
  • a change in the relationship between members of the above identified semantic entities 211 - 214 is a semantic event 213 . That change may as easily be a change in the mapping that describes the relationship, as a change in the parameters of that mapping (It is possible to write this all in a way that makes every semantic event 213 a change in parameters, by using a function space and indexing it over an appropriate parameter set).
  • Each “mental space”, then, is an extended description containing entities, relationships, and frames, and several “mental spaces” may be active at once, in order to properly define all the entities in the semantic description. These “mental spaces” enter into relationships with each other. Since these “mental spaces” borrow structure and entities from each other, there are mappings necessarily between such “mental spaces”. The whole composite forms a backdrop to the expressed description, and completes the process of attaching semantic meaning to the entities involved in the speech.
  • frames 215 may be prototypical semantic descriptions themselves, or they may be sets of rules, definitions, and descriptive structures.
  • Descriptive structures 217 are abstractions of semantic objects 211 , semantic episodes 214 , semantic states 212 , and relationships (which are either relations 221 or morphisms/mappings 222 as described above) to graphs 223 , with or without extra properties.
  • Mental spaces 216 are collections of semantic objects 211 , relationships (which are either relations 221 or morphisms/mappings 222 as described above), and frames 215 , together with mappings which embed descriptive structures 217 from semantic descriptions or from other mental spaces.
  • FIG. 3 illustrates numerous mental spaces 310 , 320 , and 330 in accordance with an embodiment of the present invention, showing creation of a new mental space 330 by recruiting frames 360 and 362 and borrowing descriptive structure from other mental spaces 310 and 320 .
  • the mapping 340 indicates that new mental space 330 borrows descriptive structure from mental space1 310 .
  • the mapping 345 indicates that new mental space 330 borrows descriptive structure from mental space2 320 .
  • the recruitment arrow 355 indicates that the new mental space 330 recruits the frame 362 from the set of frames 360 - 363 .
  • the recruitment arrow 350 indicates that the new mental space 330 recruits the frame 360 from the set of frames 360 - 363 .
  • FIG. 4 illustrates numerous mental spaces 410 - 440 in accordance with an embodiment of the present invention, showing creation of a blend mental space 440 by integrating or blending input mental space1 420 and input mental space2 430 .
  • the generic mental space 410 has structures that are found in both the input mental space1 420 and input mental space2 430 .
  • the blend mental space 440 integrates borrowed descriptive structures from the input mental space1 420 and input mental space2 430 to form new structures.
  • the structure required to represent the complex nature of semantic description for content data may need to be as complex.
  • the mapping which projects structure from one mental space to another mental space is properly known as-metaphor, or analogy.
  • metaphors are used daily without realization of its use. The expression “getting close to the deadline”, borrows spatial structure to talk about time.
  • Semantic objects 211 are descriptions of real objects, or of composites or abstractions of these real objects. They contain semantic states 212 . Semantic objects 211 may have semantic subobjects. The semantic states 212 may have semantic substates. Semantic states 212 are collections of attributes. Semantic states 212 may be attached to semantic objects 211 , relationships (which are either relations 221 or morphisms/mappings 222 ), and semantic episodes 214 . By extension, they may be attribute collections of mental spaces 216 . Semantic states 212 may have semantic substates. -Semantic events 213 are changes in semantic states 212 .
  • a semantic event 213 may be a change in any of the constituents of a description of a semantic object 211 , a semantic episode 214 , or a relationship (including what represents the mental spaces 216 ). Since semantic states 212 may have semantic substates, semantic events 213 may have semantic subevents.
  • semantic episodes 214 are semantically significant time spans. They may coincide with the behavior of semantic objects 211 , with the occurrence of semantic events 213 , with changes in relationships, or changes in the mental spaces 216 used to provide context to the semantic objects 211 , semantic events 213 , and relationships. If semantically significant time spans are properly contained in a semantic episode 214 , these semantically significant time spans are semantic subepisodes. Frames 215 are preassumed or predefined sets of rules for interpreting or describing a set of semantic objects 211 . As such, they may be prototypical descriptions themselves, or they may be sets of rules, definitions, and descriptive structures 217 .
  • Descriptive structures 217 are abstractions of semantic objects 211 , semantic episodes 214 , semantic states 212 , and relationships to graphs 223 , with or without extra properties.
  • Mental spaces 216 are collections of semantic objects 211 , relationships (which are either relations 221 or morphisms/mappings 222 ), and frames 215 , together with mappings which embed descriptive structures 217 from semantic descriptions or from other mental spaces.
  • categorical structure tools 220 such as a category 224 , a graph 223 , a functor 225 , and a natural transformation 226 were needed to create any type or kind of semantic description for content data.
  • semantic description of content data is therefore characterized by structure.
  • the relationships between semantic objects 211 form structure.
  • the mapping of semantic objects 211 , semantic states 212 , and semantic events 213 into a semantic episode 214 is structure.
  • the mappings that make up the underlying mental spaces 216 are structure. It is possible to represent semantic states 212 as maps from the entities they describe to spaces of attribute values.
  • FIG. 5A illustrates a morphism 510 in accordance with an embodiment of the present invention.
  • the morphism 510 is a directed arrow from entity 525 to entity 520 .
  • Any of the entities 520 and 525 can be a semantic object 211 , a semantic event 213 , a relationship, a semantic states 212 , a semantic episode 214 , a frame 215 , a descriptive structure 217 , a mental space 216 , or any other entity.
  • a graph 223 has a set of morphisms between mathematical objects and a set of mathematical objects, with the morphisms as edges and the mathematical objects as vertices or nodes.
  • FIG. 5D illustrates a graph 570 in accordance with an embodiment of the present invention, showing the edges 571 and the nodes 572 .
  • FIG. 5E illustrates a graph morphism 593 (F) between graph1 591 and graph2 592 in accordance with an embodiment of the present invention.
  • the graph morphism 593 (F) is a pair of mappings: a mapping of an edge between two nodes and a mapping of the two nodes.
  • graphs 223 when graphs 223 obey categorical constraints (i.e., they respect identity and composition on the mathematical objects),-the graphs 223 are categories 224 .
  • Graphs 223 can also be regarded as mathematical objects in their own right (i.e., making mathematical objects out of maps). This was done above when a change in the semantic state 212 of a relationship was allowed to be a semantic event 213 .
  • FIG. 5B illustrates a first functor 530 , a second functor 531 , a first category 536 , and a second category 535 in accordance with an embodiment of the present invention.
  • the functor1 530 and the functor2 531 are directed from the categoryl 536 to the category2 535 .
  • FIG. 5C illustrates a natural transformation 580 in accordance with an embodiment of the present invention.
  • the functor1 581 is directed from the category1 583 to the category2 584 .
  • the functor 2 582 is directed from the category 3 585 to the category4 586 .
  • the natural transformation 580 is directed from functor1 581 to functor2 582 .
  • a map from part of a semantic description into a semantic description is defined. This can be done by relying on a characteristic function 227 to the part in question, composed with a map to the target mental space 216 . All of this is categorical structure. Moreover, the spaces from which structure is generated are required to have such characteristic functions 227 . In addition, a large number of relationships are required. Lastly, it is possible to form product spaces in which to create these relationships. In sum, the categorical structures described above enable creation of any type or kind of semantic description for content data.
  • the semantic entity tools 210 (FIG. 2) and the categorical structure tools 220 (FIG. 2) facilitate creation of a semantic mosaic description for content data.
  • the semantic mosaic description is formed from multiple semantic descriptions. These semantic descriptions are integrated with each other such that each semantic description is modified at a local level within localized regions without substantially changing each semantic description outside these localized regions.
  • the semantic mosaic description facilitates navigation or browsing through the multiple semantic descriptions and the content data.
  • FIG. 6 illustrates a semantic description1 610 and a semantic description2 660 in accordance with an embodiment of the present invention.
  • the semantic description1 610 and the semantic description2 660 were formed using the semantic entity tools 210 (FIG. 2) and the categorical structure tools 220 (FIG. 2).
  • a semantic mosaic description is formed from the semantic description1 610 and the semantic description2 660 .
  • the semantic description1 610 and the semantic description2 660 are not substantially changed outside of the localized regions 615 and 665 when they form the semantic mosaic description. It should be understood that any number of semantic descriptions can be integrated into a semantic mosaic description.
  • FIG. 7 illustrates a semantic mosaic description 750 based on the semarntic description1 610 of FIG. 6 and the semantic description2 of FIG. 6.
  • the semantic mosaic description 750 provides several benefits.
  • the semantic mosaic description 750 enables additional semantic information to be added to a semantic description where necessary or needed without affecting the entire semantic description.
  • the semantic mosaic description 750 can represent a complete semantic description which is formed from multiple partial semantic descriptions.
  • the semantic mosaic description can facilitate navigating or browsing through the semantic description1 610 and 660 as is done with content data such as audio-visual material. If the localized regions 615 and 665 (FIG. 6) have common elements, the transitions within the semantic mosaic description 750 are smooth. More importantly, as a whole the semantic mosaic description 750 may or may not semantically describe something, but within regions of the semantic mosaic description 750 , something is semantically described.
  • the semantic entity tools 210 and the categorical structure tools 220 (FIG. 2) facilitate creation of a semantic description for content data using multiple component semantic descriptions stored remotely from the content data.
  • Reference information can be associated with the content data, whereas the reference information includes the identity of the component semantic descriptions needed to form the semantic description, the location of these component semantic descriptions, and the manner of processing these component semantic descriptions to form the semantic description.
  • the component semantic descriptions identified in the reference information are retrieved (e.g., from a location on a network, a control dictionary, etc.). Then, the semantic description is formed in the manner specified in the reference information using the component semantic descriptions.
  • the semantic description does not have to be stored in a discrete location, saving storage resources and promoting re-use of component semantic descriptions.
  • FIG. 8 illustrates formation of a semantic description 840 for content data 805 using multiple component semantic descriptions stored in locations on a network 850 in accordance with an embodiment of the present invention.
  • a plurality of component semantic descriptions 830 A- 830 E are distributively stored in a plurality of locations on a network 850 .
  • the plurality of component semantic descriptions 830 A- 830 E are stored remotely from the content data 805 .
  • the network 850 can be the Internet 850 or any other type of network.
  • a semantic description 840 is formed from copies of one or more of the component semantic descriptions 830 A- 830 B.
  • a complicated semantic description can be formed quickly and easily by referencing, adding new semantic information to, modifying, combining, or extracting partial semantic descriptions from the component semantic descriptions 830 A- 830 E.
  • the semantic description for an elaborate wedding can be formed by using the distributively stored component semantic descriptions of a basic wedding, a fancy wedding gown, a stretch limousine, an expensive wedding cake, etc. These component semantic descriptions are modified and combined to form the semantic description for the elaborate wedding.
  • partial semantic descriptions can be extracted from the component semantic descriptions and then combined and/or modified with other component semantic descriptions.
  • the semantic description 840 can be generated when needed, reducing the demand for storage resources and encouraging re-use of component semantic descriptions 830 A- 830 E.
  • the content data 805 includes reference information 810 .
  • the computer system 820 or any other computational system such as a MPEG-7 device utilizes the reference information 810 to generate the semantic description 840 for the content data 805 .
  • the reference information 810 includes the identity of the component semantic descriptions 830 A- 830 B needed to form the semantic description 840 , the location of these component semantic descriptions 830 A- 830 B, and the manner of processing these component semantic descriptions 830 A- 830 B to form the semantic description 840 . It should be understood that the reference information 810 can have any other type of information.
  • each component semantic description 830 A- 830 E is assigned a uniform resource identifier (URI) to facilitate access to the component semantic descriptions 830 A- 830 E.
  • URI uniform resource identifier
  • the reference information 810 has the URI for the component semantic descriptions 830 A 830 E needed to form the semantic description 840 .
  • the computer system 820 or any other computational system such as a MPEG-7 device utilizes the URI(s) to retrieve the corresponding component semantic descriptions 830 A- 830 B, as illustrated in FIG. 8.
  • each component semantic description 830 A- 830 E has information pertaining to its use. This information can indicate whether the component semantic description can be subsumed (i.e., can be embedded in another semantic description without changing its intended meaning). Moreover, this information can indicate whether the component semantic description can be subdivided (i.e., admits subdivisions which make the extraction of subsets of its semantic information natural). In addition, this information can indicate whether the component semantic description can be transformed. Furthermore, this information can indicate whether the component semantic description is transitive (i.e., functions as a subset if embedded in another semantic description).
  • FIG. 9 illustrates formation of a semantic description 840 for content data 805 using multiple component semantic descriptions stored in a control dictionary 860 in accordance with an embodiment of the present invention.
  • the discussion of FIG. 8 is applicable to FIG. 9.
  • the plurality of component semantic descriptions 83 OA- 830 E are distributively stored in a control dictionary 860 rather than in a plurality of locations on a network.
  • semantic descriptions pertaining to mathematics can be generated from component semantic descriptions retrieved from a control dictionary 860 emphasizing mathematical terms.
  • An index value associated with the control dictionary 860 can be utilized to access the component semantic descriptions stored in the control dictionary 860 . It is possible to have a plurality of control dictionaries and different types of control dictionaries.
  • FIG. 10 illustrates a flow chart showing a method 1000 of forming a semantic description for content data using multiple component semantic descriptions stored remotely from the content data in accordance with an embodiment of the present invention. Reference is made to FIGS. 8 and 9.
  • step 1005 the method 1000 in accordance with an embodiment of the present invention begins.
  • numerous component semantic descriptions 830 A- 830 E are distributively stored. Specifically, the numerous component semantic descriptions 830 A- 830 E are stored remotely from the content data.
  • the component semantic descriptions 830 A- 830 E can be stored in locations on a network 850 .
  • the component semantic descriptions 830 A- 830 E can be stored in one or more control dictionaries 860 .
  • the component semantic descriptions 830 A- 830 E can have generic semantic information or specific semantic information.
  • reference information 810 (configured as described above) is associated with the content data 805 .
  • This association can take place in a real-time environment or in a non real-time environment, whereas a real-time environment means that the reference information 810 is generated at the same time as the content data 805 is being captured.
  • step 1020 it is determined whether to generate the specific semantic description 840 from one or more component semantic descriptions 830 A- 830 E.
  • the computer system 820 or any other computational system such as a MPEG-7 device may receive a request for the specific semantic description 840 for the content data 805 in order to display, search, index, filter, or otherwise process the content data 805 .
  • the method 1000 ends if the specific semantic description 840 is not needed.
  • the computer system 820 or any other computational system retrieves the component semantic descriptions 830 A- 830 B identified by the reference information 810 from a network 850 or from a control dictionary 860 .
  • the computer system 820 or any other computational system such as a MPEG-7 device generates the specific semantic description 840 using the retrieved component semantic descriptions 830 A- 830 B and the reference information 810 which indicates the manner of processing these component semantic descriptions 830 A- 830 B to form the specific semantic description 840 .
  • the reference information 810 indicates the manner of referencing, adding new semantic information to, modifying, combining, or extracting partial semantic descriptions from the component semantic descriptions 830 A- 830 B.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Multimedia (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Animal Behavior & Ethology (AREA)
  • Computational Linguistics (AREA)
  • Machine Translation (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

The semantic entity tools and the categorical structure tools facilitate creation of a semantic description for content data using multiple component semantic descriptions stored remotely from the content data. Reference information is associated with the content data. When the semantic description is desired, the component semantic descriptions identified in the reference information are retrieved (e.g., from a location on a network, a control dictionary, etc.). Then, the semantic description is formed in the manner specified in the reference information using the component semantic descriptions. Thus, the semantic description does not have to be stored in a discrete location, saving storage resources and promoting re-use of component semantic descriptions.

Description

    RELATED U.S. APPLICATION
  • This patent application claims the benefit of U.S. Provisional Application No. 60/189,202, filed on Mar. 14, 2000, entitled “Report On The Importance Of Structure In Semantic Descriptions”, by Hawley K. Rising III, and Ali Tabatabai. This patent application claims the benefit of U.S. Provisional Application No. 60/189,626, filed on Mar. 14, 2000, entitled “Contribution On The Distribution Of Semantic Information”, by Hawley K. Rising III, and Ali Tabatabai. This patent application claims the benefit of U.S. Provisional Application No. 60/191,312, filed on Mar. 21, 2000, entitled “Report On The Importance Of Structure In Semantic Descriptions Using Semantic Mosaics”, by Hawley K. Rising III, and Ali Tabatabai.[0001]
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention [0002]
  • The present invention generally relates to the field of semantic descriptions for content data. More particularly, the present invention relates to the field of methods and systems for implementing powerful and flexible semantic description tools to describe the underlying meaning of the content data. [0003]
  • 2. Related Art [0004]
  • The MPEG-7 “Multimedia Content Description Interface” standard which is being developed by the Moving Pictures Expert Group (MPEG) focuses, unlike the preceding MPEG standards (e.g., MPEG-1, MPEG-2, and MPEG-4), on representing information about the content data, not the content data itself. The goal of the MPEG-7 standard is to provide a rich set of standardized tools to describe content data. In particular, MPEG-7 seeks to provide a simple, flexible, interoperable solution to the problems of indexing, searching, and retrieving content data. More specifically, MPEG-7 aims to standardize a core set of Descriptors that can be used to describe the various features of the content data; pre-defined structures of Descriptors and their relationships, called Description Schemes; a language to define Description Schemes and Descriptors, called the Description Definition Language (DDL); and coded representations of descriptions to enable efficient storage and fast access. The DDL is being based on XML Schema. Moreover, the MPEG-7 descriptions (a set of instantiated Description Schemes) are linked to the content data itself to allow fast and efficient searching for material of a users interest. [0005]
  • Continuing, MPEG-7 intends to describe content data regardless of storage, coding, display, transmission, medium, or technology. MPEG-7 addresses a wide variety of media types including: still pictures, graphics, 3D models, audio, speech, video, and any combination thereof (e.g., multimedia presentations, scenarios, etc.). Examples of content data within the MPEG-7 standard include an MPEG-4 data stream; a video tape; a CD containing music, sound, or speech; a picture printed on paper, and an interactive multimedia installation on the Web (i.e., the Internet). [0006]
  • The MPEG-7 standard includes different types of Descriptors and Description Schemes. Some Descriptors and Description Schemes describe what is in the content data in terms of syntactic structure, color histogram, shape of an object, texture, motion, pitch, rhythm, etc. [0007]
  • On the other hand, semantic Description Schemes describe the underlying meaning or understanding of the content data. In particular, a goal, advertisement, and Madonna are examples of a semantic description (an instantiated semantic Description Scheme). Other examples of semantic descriptions includes a storyline for a movie (i.e., content data), a description of a scene in the movie, a description of an image, a description of a piece of music, etc. [0008]
  • Again, the semantic description is based on the underlying meaning of the content data. Typically, the semantic description is expressed with words. Unfortunately, computer systems or other computational systems are not able to usefully manipulate (e.g., create, exchange, retrieve, etc.) semantic descriptions expressed with only words. However, if structure is incorporated into the semantic descriptions, a computer system or other computational system can usefully manipulate semantic descriptions having structure. For example, it is not sufficient to describe the movie Zorro as having the entities Zorro, Zorro's girlfriend, a bad guy, a first sword fight, a second sword fight, etc. Relationships between these entities are needed, hence providing the structure. [0009]
  • Numerous proposals have been made to limit the types of structure to be incorporated into the semantic descriptions of the MPEG-7 standard. In particular, these proposals advocate creating specific, static semantic description schemes having only certain types of structure. Moreover, these proposals further encourage setting-up and running experiments to verify these specific, static semantic description schemes. [0010]
  • There are several problems with these proposals. First, these experiments can conclude that these specific, static semantic description schemes function well during these experiments because of the conditions of the experiments. Yet, these specific, static semantic description schemes can still fail when applied to new descriptive situations. For example, if these specific, static semantic description schemes can be applied to describe a soccer game, there is no way of knowing whether these specific, static semantic description schemes can be applied to describe a human birth. Secondly, these experiments do not indicate or help to determine the range of semantic descriptions that are impossible to implement or no longer capable of being implemented with these specific, static semantic description schemes because of the limitation on the types of structure incorporated. [0011]
  • SUMMARY OF THE INVENTION
  • Instead of focusing on specific, static semantic description schemes, emphasis and focus is placed on determining what is necessary and needed to create any type or kind of semantic description for content data in various applications such as MPEG-7. In particular, numerous semantic description tools are selected. These semantic description tools provide sufficient flexibility and power to create any type or kind of semantic description. Numerous semantic entity tools and numerous categorical structure tools were identified as necessary and needed to create any type or kind of semantic description. Semantic entity tools are tools that represent entities in a semantic description. Categorical structure tools are tools that represent categorical structures of semantic entities and relations among these categorical structures. [0012]
  • The process of developing semantic descriptions was analyzed using principles from cognitive science. This analysis showed that the process of developing semantic descriptions typically did not involve transferring or communicating entire semantic descriptions from one person to another person. Instead, each person developed his/her own semantic description based on prior experiences which were recalled, modified, combined, and extracted in various ways. From this observation, it was determined that semantic entity tools which had operational properties resembling these cognitive operations were needed to create any type or kind of semantic description. [0013]
  • Moreover, the principles of category theory were examined to determine whether categorical structures (structures observing the principles of category theory) could provide sufficient flexible structure to create any type or kind of semantic description. This examination revealed that the semantic entity tools could be mapped onto categorical structures such as a graph. Hence, categorical structure tools such a category, a graph, a functor, and a natural transformation were needed to create any type or kind of semantic description. [0014]
  • In another embodiment of the present invention, the semantic entity tools and the categorical structure tools facilitate creation of a semantic mosaic description for content data. The semantic mosaic description is formed from multiple semantic descriptions. These semantic descriptions are integrated with each other such that each semantic description is modified at a local level within localized regions without substantially changing each semantic description outside these localized regions. In particular, the semantic mosaic description facilitates navigation or browsing through the multiple semantic descriptions and the content data. [0015]
  • In yet another embodiment of the present invention, the semantic entity tools and the categorical structure tools facilitate creation of a semantic description for content data using multiple component semantic descriptions stored remotely from the content data. Reference information is associated with the content data, whereas the reference information includes the identity of the component semantic descriptions needed to form the semantic description, the location of these component semantic descriptions, and the manner of processing these component semantic descriptions to form the semantic description. When the semantic description is desired, the component semantic descriptions identified in the reference information are retrieved (e.g., from a location on a network, a control dictionary, etc.). Then, the semantic description is formed in the manner specified in the reference information using the component semantic descriptions. Thus, the semantic description does not have to be stored in a discrete location, saving storage resources and promoting re-use of component semantic descriptions.[0016]
  • These and other advantages of the present invention will no doubt become apparent to those of ordinary skill in the art after having read the following detailed description of the preferred embodiments which are illustrated in the drawing figures. [0017]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings, which are incorporated in and form a part of this specification, illustrate embodiments of the invention and, together with the description, serve to explain the principles of the present invention. [0018]
  • FIG. 1 illustrates an exemplary computer system in which the present invention can be practiced. [0019]
  • FIG. 2 illustrates semantic entity tools and categorical structure tools in accordance with an embodiment of the present invention. [0020]
  • FIG. 3 illustrates numerous mental spaces in accordance with an embodiment of the present invention, showing creation of a new mental space by recruiting frames and borrowing structure from other mental spaces. [0021]
  • FIG. 4 illustrates numerous mental spaces in accordance with an embodiment of the present invention, showing creation of a blend mental space by integrating or blending input mental space1 and input mental space2. [0022]
  • FIG. 5A illustrates a morphism in accordance with an embodiment of the present invention. [0023]
  • FIG. 5B illustrates a first functor, a second functor, a first category, and a second category in accordance with an embodiment of the present invention. [0024]
  • FIG. 5C illustrates a natural transformation in accordance with an embodiment of the present invention. [0025]
  • FIG. 5D illustrates a graph in accordance with an embodiment of the present invention. [0026]
  • FIG. 5E illustrates a graph morphism in accordance with an embodiment of the present invention. [0027]
  • FIG. 6 illustrates a semantic descriptions and a semantic description1 in accordance with an embodiment of the present invention. [0028]
  • FIG. 7 illustrates a semantic mosaic description based on the semantic description1 of FIG. 6 and the semantic description2 of FIG. 6. [0029]
  • FIG. 8 illustrates formation of a semantic description for content data using multiple component semantic descriptions stored in locations on a network in accordance with an embodiment of the present invention. [0030]
  • FIG. 9 illustrates formation of a semantic description for content data using multiple component semantic descriptions stored in a control dictionary in accordance with an embodiment of the present invention. [0031]
  • FIG. 10 illustrates a flow chart showing a method of forming a semantic description for content data using multiple component semantic descriptions stored remotely from the content data in accordance with an embodiment of the present invention.[0032]
  • The drawings referred to in this description should not be understood as being drawn to scale except if specifically noted. [0033]
  • DETAILED DESCRIPTION OF THE INVENTION
  • Reference will now be made in detail to the preferred embodiments of the present invention, examples of which are illustrated in the accompanying drawings. While the invention will be described in conjunction with the preferred embodiments, it will be understood that they are not intended to limit the invention to these embodiments. On the contrary, the invention is intended to cover alternatives, modifications and equivalents, which may be included within the spirit and scope of the invention as defined by the appended claims. Furthermore, in the following detailed description of the present invention, numerous specific details are set forth in order to provide a thorough understanding of the present invention. However, it will be recognized by one of ordinary skill in the art that the present invention may be practiced without these specific details. In other instances, well known methods, procedures, components, and circuits have not been described in detail as not to unnecessarily obscure aspects of the present invention. [0034]
  • Notation and Nomenclature
  • Some portions of the detailed descriptions which follow are presented in terms of procedures, logic blocks, processing, and other symbolic representations of operations on data bits within a computer memory. These descriptions and representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. In the present application, a procedure, logic block, process, etc., is conceived to be a self-consistent sequence of steps or instructions leading to a desired result. The steps are those requiring physical manipulations of physical quantities. Usually, though not necessarily, these quantities take the form of electrical or magnetic signals capable of being stored, transferred, combined, compared, and otherwise manipulated in a computer system. It has proved convenient at times, principally for reasons of common usage, to refer to these signals as bits, values, elements, symbols, characters, terms, numbers, or the like. [0035]
  • It should be borne in mind, however, that all of these and similar terms are to be associated with the appropriate physical quantities and are merely convenient labels applied to these quantities. Unless specifically stated otherwise as apparent from the following discussions, it is appreciated that throughout the present invention, a variety of terms are discussed that refer to the actions and processes of an electronic system or a computer system, or other electronic computing device/system. The computer system or similar electronic computing device manipulates and transforms data represented as physical (electronic) quantities within the computer system's registers and memories into other data similarly represented as physical quantities within the computer system memories or registers or other such information storage, transmission, or display devices. The present invention is also well suited to the use of other computer systems such as, for example, optical, mechanical, or quantum computers. [0036]
  • Exemplary Computer System Environment
  • Aspects of the present invention are discussed in terms of steps executed on a computer system or any other computational system. Although a variety of different computer systems can be used with the present invention, an [0037] exemplary computer system 100 is shown in FIG. 1.
  • With reference to FIG. 1, portions of the present invention are comprised of computer-readable and computer executable instructions which reside, for example, in computer-usable media of an electronic system such as the exemplary computer system. FIG. 1 illustrates an [0038] exemplary computer system 100 on which embodiments of the present invention may be practiced. It is appreciated that the computer system 100 of FIG. 1 is exemplary only and that the present invention can operate within a number of different computer systems including general-purpose computer systems and embedded computer systems.
  • [0039] Computer system 100 includes an address/data bus 110 for communicating information, a central processor 101 coupled with bus 110 for processing information and instructions, a volatile memory 102 (e.g., random access memory RAM) coupled with the bus 110 for storing information and instructions for the central processor 101 and a non-volatile memory 103 (e.g., read only memory ROM) coupled with the bus 110 for storing static information and instructions for the processor 101. Exemplary computer system 100 also includes a data storage device 104 (“disk subsystem”) such as a magnetic or optical disk and disk drive coupled with the bus 110 for storing information and instructions. Data storage device 104 can include one or more removable magnetic or optical storage media (e.g., diskettes, tapes) which are computer readable memories. Memory units of computer system 100 include volatile memory 102, non-volatile memory 103 and data storage device 104.
  • [0040] Exemplary computer system 100 can further include an optional signal generating device 108 (e.g., a network interface card “NIC”) coupled to the bus 110 for interfacing with other computer systems. Also included in exemplary computer system 100 of FIG. 1 is an optional alphanumeric input device 106 including alphanumeric and function keys coupled to the bus 110 for communicating information and command selections to the central processor 101. Exemplary computer system 100 also includes an optional cursor control or directing device 107 coupled to the bus 110 for communicating user input information and command selections to the central processor 101. An optional display device 105 can also be coupled to the bus 110 for displaying information to the computer user. Display device 105 may be a liquid crystal device, other flat panel display, cathode ray tube, or other display device suitable for creating graphic images and alphanumeric characters recognizable to the user. Cursor control device 107 allows the user to dynamically signal the two-dimensional movement of a visible symbol (cursor) on a display screen of display device 105. Many implementations of cursor control device 107 are known in the art including a trackball, mouse, touch pad, joystick or special keys on alphanumeric input device 106 capable of signaling movement of a given direction or manner of displacement. Alternatively, it will be appreciated that a cursor can be directed and/or activated via input from alphanumeric input device 106 using special keys and keysequence commands.
  • Category Theory and Cognitive Science in the Design of Semantic Descriptions for Content Data
  • The present invention is applicable to the MPEG-7 standard or to any other application which uses semantic descriptions. [0041]
  • FIG. 2 illustrates [0042] semantic entity tools 210 and categorical structure tools 220 in accordance with an embodiment of the present invention. Instead of focusing on specific, static semantic description schemes, emphasis and focus is placed on determining what is necessary and needed to create any type or kind of semantic description for content data in various applications such as MPEG-7. In particular, numerous semantic description tools 210 and 220 are selected. These semantic description tools 210 and 220 provide sufficient flexibility and power to create any type or kind of semantic description. Numerous semantic entity tools 211-217 and numerous categorical structure tools 221-227 were identified as necessary and needed to create any type or kind of semantic description. Semantic entity tools 211-217 are tools that represent entities in a semantic description. Categorical structure tools 221-227 are tools that represent categorical structures of semantic entities 211-217 and relations among these categorical structures. In an embodiment, the semantic entity tools 211-217 and the categorical structure tools 221-227 are implemented as Description Schemes.
  • The [0043] semantic entity tools 210 include core semantic entities 211-214, constructed semantic entities 216-217, and a context semantic entity 215. The core semantic entities 211-214 include a semantic object 211, a semantic state 212, a semantic event 213, and a semantic episode 214. The context semantic entity 215 includes a frame 215. Moreover, the constructed semantic entities 216-217 include a mental space 216 and a descriptive structure 217. In an embodiment, each constructed semantic entity can include a core semantic entity 211, a context semantic entity 215, and relationships among these.
  • The [0044] categorical structure tools 220 include a relation 221, a morphism 222, a graph 223, a category 224, a functor 225, a natural transformation 226, and a characteristic function 227.
  • Referring again to FIG. 1, a semantic object or object [0045] 211 in a semantic description is derived from a physical object or an abstraction of the physical object. In particular, the semantic object 211 describes a physical or abstract object semantically. Physical objects have spatial contiguity, and temporal duration. Physical objects are described in various ways. Moreover, the change in a semantic object 211 over time, or the particular circumstances, or type of a generic semantic object 211, are described with reference to attributes, which are qualities of the semantic object 211. The collection of these qualities changes over time, and can be called the semantic state or state 212 of the semantic object 211. Thus, semantic objects 211 have semantic states 212.
  • Physical objects are frequently divisible. The subsets of the material of a physical object can be physical objects in their own right. These subsets can be referred as physical subobjects. Thus, semantic objects [0046] 211 can have semantic subobjects. The collection of semantic subobjects of a given semantic object 211, or the collection of semantic subobjects of a collection of semantic objects 211, admits a partial order, by inclusion.
  • Likewise, since [0047] semantic states 212 are frequently complex, containing more than a single attribute, they can have subcollections. These subcollections can be semantic states 212 if these subcollections have semantic meaning. Thus, semantic states 212 can have semantic substates.
  • A change in [0048] semantic state 212 is a semantic event or event 213. Since, as was remarked above, semantic states 212 are complex, a semantic event 213 may likewise be complex, since the semantic event 213 may indicate the change in a large number of attributes. Consequently, if such a set of attributes admits a subset with semantic meaning, and that subset can change independently from the rest, a semantic event 213 can have semantic subevents.
  • Thus, a semantic description formed with a semantic object [0049] 211 may or may not describe semantic subobjects, semantic states 212, semantic substates, semantic events 213, or semantic subevents. More importantly, the semantic description may contain relationships other than inclusion of parts.
  • A semantic episode or [0050] episode 214 denotes an inclusive semantic description of what transpires over a period of time, from some (possibly implied) starting time to (also possibly implied) ending time, with a duration greater than zero. A semantic episode 214 can be a temporal designation with semantic meaning. If there are time periods of shorter duration between the start of the semantic episode 214 and the end of the semantic episode 214, which have semantic meaning, these may be called semantic subepisodes.
  • The semantic description includes relationships. One relationship that has already been seen and holds for all of the above identified semantic entities [0051] 211-214 is that of inclusion, in the manner of a semantic subobject, semantic subevent, semantic subepisode, or semantic substate. The lists of relationships between such semantic entities 211-214 can be quite long. The formal definitions of two mathematical concepts, which will facilitate them, are the definition of a relation 221 and the definition of a morphism or mapping 222, which are illustrated in FIG. 2.
  • A relation on a group of mathematical objects is a subset of the formal cartesian product of the mathematical objects. For instance, a binary relation is a subset of the set of ordered pairs of mathematical objects. A partial order is a subset such that if (a,b) and (b,c) are in the set, so is (a,c), and if (a,b) and (b,a) are in the set then a=b. Inclusion is a partial order. Moreover, containment is a partial order (i.e., when one mathematical object is contained in another). Containment and inclusion are not the same: One would hardly say that a fish is part of a fish tank, but it is likely to be found there. [0052]
  • A morphism or [0053] mapping 222 is an assignment consisting of ordered pairs from a set called the domain and a set called the codomain. It can have more distinction than that, for instance, a function is a mapping where the codomain is the real (or complex) numbers, and for each element a of the domain, there is exactly one element b of the codomain.
  • Thus, a relationship between mathematical objects is either a [0054] relation 221 or a morphism/mapping 222. Since relations 221 can be expressed as compositions of mappings, (a and b map to (a,b) which maps via the characteristic function 227 of the subset mentioned above to either true or false. A generalization of the characteristic function 227 maps to a discrete set, and is called a subobject classifier.), a relationship is a morphism or mapping 222. There are several kinds of relationships. Inclusion was mentioned above. Moreover, containment, similarity, example of, and relative position are also relationships.
  • Since inclusion is a relationship on all of the categories of semantic entities [0055] 211-214 identified above, semantic objects 211, semantic events 212, semantic states 213, and semantic episodes 214 can all have relationships. It is also possible to have relationships between these semantic entities 211-214, the most obvious being between semantic objects 211, semantic events 213, and semantic states 212, but semantic episodes 214 may sometimes be effectively described by relationships as well. As noted above, semantic events 213 are described as a change in semantic state 212, a semantic state 212 being a collection of attributes for a semantic object 211. Furthermore, a relationship is a morphism or mapping 221. Mappings may be parametrized. Thus, a change in the parameters of a mapping between two of the above identified semantic entities 211-214 fits well as a semantic event 213. In fact, it is possible for semantic entities 211-214 of the above categories to be described by a complex set of mappings. This set is also a relationship. A change in the relationship between members of the above identified semantic entities 211-214 is a semantic event 213. That change may as easily be a change in the mapping that describes the relationship, as a change in the parameters of that mapping (It is possible to write this all in a way that makes every semantic event 213 a change in parameters, by using a function space and indexing it over an appropriate parameter set).
  • The process of developing semantic descriptions was analyzed using principles from cognitive science such as “input mental spaces”, mappings between “mental spaces”, and “blend mental spaces”. Cognitive science provides schemes for interpreting semantic content in language. The understanding of “mental spaces” and their mappings is apropos to creation of semantic descriptions for content data. In particular, mappings, precedences, and contexts that really imbue semantic descriptions with meaning depend on the rules governing perception and interpretation. This can be described by a “mental space”, mappings between “mental spaces”, and integration of part or all of a set of “mental spaces” into a new “mental space”. The interpretation of speech, which is, after all, the prototype for semantic description of content data, requires the construction of a set of “mental spaces” which provide context for the communication. These “mental spaces” are built by importing a lot of information not included in the speech (which is interpreted as semantic description). The maps by which this is done include recruiting “frames”, which are predefined constructs for interpretation, projecting structure from one semantic description to another, and integrating or abstracting imported material from more than one other semantic description. This process is not limited to descriptive speech per se. [0056]
  • Each “mental space”, then, is an extended description containing entities, relationships, and frames, and several “mental spaces” may be active at once, in order to properly define all the entities in the semantic description. These “mental spaces” enter into relationships with each other. Since these “mental spaces” borrow structure and entities from each other, there are mappings necessarily between such “mental spaces”. The whole composite forms a backdrop to the expressed description, and completes the process of attaching semantic meaning to the entities involved in the speech. [0057]
  • This analysis shows that the process of developing semantic descriptions typically does not involve transferring or communicating entire semantic descriptions from one person to another person. Instead, each person develops his/her own semantic description based on prior experiences which are recalled, modified, combined, extracted, and mapped in various ways. From this observation, it was determined that semantic entity tools which had operational properties resembling these cognitive operations were needed to create any type or kind of semantic description. As illustrated in FIG. 2, the mental space [0058] 216, the descriptive structure 217, and the frame 215 are semantic entity tools originating from cognitive concepts. Frames 215 are preassumed or predefined sets of rules for interpreting or describing a set of semantic objects 211. As such, frames 215 may be prototypical semantic descriptions themselves, or they may be sets of rules, definitions, and descriptive structures. Descriptive structures 217 are abstractions of semantic objects 211, semantic episodes 214, semantic states 212, and relationships (which are either relations 221 or morphisms/mappings 222 as described above) to graphs 223, with or without extra properties. Mental spaces 216 are collections of semantic objects 211, relationships (which are either relations 221 or morphisms/mappings 222 as described above), and frames 215, together with mappings which embed descriptive structures 217 from semantic descriptions or from other mental spaces.
  • FIG. 3 illustrates numerous [0059] mental spaces 310, 320, and 330 in accordance with an embodiment of the present invention, showing creation of a new mental space 330 by recruiting frames 360 and 362 and borrowing descriptive structure from other mental spaces 310 and 320. In particular, the mapping 340 indicates that new mental space 330 borrows descriptive structure from mental space1 310. The mapping 345 indicates that new mental space 330 borrows descriptive structure from mental space2 320. Moreover, the recruitment arrow 355 indicates that the new mental space 330 recruits the frame 362 from the set of frames 360-363. In addition, the recruitment arrow 350 indicates that the new mental space 330 recruits the frame 360 from the set of frames 360-363.
  • FIG. 4 illustrates numerous mental spaces [0060] 410-440 in accordance with an embodiment of the present invention, showing creation of a blend mental space 440 by integrating or blending input mental space1 420 and input mental space2 430. The generic mental space 410 has structures that are found in both the input mental space1 420 and input mental space2 430. The blend mental space 440 integrates borrowed descriptive structures from the input mental space1 420 and input mental space2 430 to form new structures.
  • Thus, the structure required to represent the complex nature of semantic description for content data may need to be as complex. At first glance, one might be tempted to limit the structure in some way, so that the semantic description would be less complex. Necessarily, this is done at the price of decreasing the type of semantic descriptions that can be constructed, and it may not be obvious how. For instance, the mapping which projects structure from one mental space to another mental space is properly known as-metaphor, or analogy. One is tempted to throw this out, given that one only wants a semantic description of content data (e.g., audiovisual material). However, metaphors are used daily without realization of its use. The expression “getting close to the deadline”, borrows spatial structure to talk about time. In a world where this has been formalized in mathematics and physics, it may not seem to be an analogy, but it is. It is also quite imperceptible. The point is that unless all semantic descriptions for content data are to be written out in formal well formed propositions, or a language which properly restricts them is to be created, it would be difficult, if not impossible, and quite possibly undesirable to restrict semantic descriptions for content data as advocated by those proposing the specific, static semantic description schemes. [0061]
  • In reviewing the [0062] semantic entity tools 210 in FIG. 2, the importance of structure is evident. Semantic objects 211 are descriptions of real objects, or of composites or abstractions of these real objects. They contain semantic states 212. Semantic objects 211 may have semantic subobjects. The semantic states 212 may have semantic substates. Semantic states 212 are collections of attributes. Semantic states 212 may be attached to semantic objects 211, relationships (which are either relations 221 or morphisms/mappings 222), and semantic episodes 214. By extension, they may be attribute collections of mental spaces 216. Semantic states 212 may have semantic substates. -Semantic events 213 are changes in semantic states 212. As such, a semantic event 213 may be a change in any of the constituents of a description of a semantic object 211, a semantic episode 214, or a relationship (including what represents the mental spaces 216). Since semantic states 212 may have semantic substates, semantic events 213 may have semantic subevents.
  • Continuing with FIG. 2, [0063] semantic episodes 214 are semantically significant time spans. They may coincide with the behavior of semantic objects 211, with the occurrence of semantic events 213, with changes in relationships, or changes in the mental spaces 216 used to provide context to the semantic objects 211, semantic events 213, and relationships. If semantically significant time spans are properly contained in a semantic episode 214, these semantically significant time spans are semantic subepisodes. Frames 215 are preassumed or predefined sets of rules for interpreting or describing a set of semantic objects 211. As such, they may be prototypical descriptions themselves, or they may be sets of rules, definitions, and descriptive structures 217. Descriptive structures 217 are abstractions of semantic objects 211, semantic episodes 214, semantic states 212, and relationships to graphs 223, with or without extra properties. Mental spaces 216 are collections of semantic objects 211, relationships (which are either relations 221 or morphisms/mappings 222), and frames 215, together with mappings which embed descriptive structures 217 from semantic descriptions or from other mental spaces.
  • Furthermore, the principles of category theory were examined to determine whether categorical structures (structures observing the principles of category theory) could provide sufficient flexible structure to create any type or kind of semantic description for content data. This examination revealed that the [0064] semantic entity tools 210 could be mapped onto categorical structures such as a graph 223. Hence, categorical structure tools 220 such a category 224, a graph 223, a functor 225, and a natural transformation 226 were needed to create any type or kind of semantic description for content data.
  • As is evident from the discussion above, a semantic description of content data (e.g., audiovisual material) is therefore characterized by structure. The relationships between semantic objects [0065] 211 form structure. The mapping of semantic objects 211, semantic states 212, and semantic events 213 into a semantic episode 214 is structure. The mappings that make up the underlying mental spaces 216 are structure. It is possible to represent semantic states 212 as maps from the entities they describe to spaces of attribute values.
  • As shown in FIG. 2, the [0066] categorical structure tools 220 take many forms. Morphisms 222 are directed arrows between mathematical objects or entities. The above identified relationships between semantic objects 211, semantic states 212, and semantic episodes 214 have been described by maps such as these (morphisms). FIG. 5A illustrates a morphism 510 in accordance with an embodiment of the present invention. The morphism 510 is a directed arrow from entity 525 to entity 520. Any of the entities 520 and 525 can be a semantic object 211, a semantic event 213, a relationship, a semantic states 212, a semantic episode 214, a frame 215, a descriptive structure 217, a mental space 216, or any other entity.
  • With reference to FIG. 2, a [0067] graph 223 has a set of morphisms between mathematical objects and a set of mathematical objects, with the morphisms as edges and the mathematical objects as vertices or nodes. FIG. 5D illustrates a graph 570 in accordance with an embodiment of the present invention, showing the edges 571 and the nodes 572. FIG. 5E illustrates a graph morphism 593 (F) between graph1 591 and graph2 592 in accordance with an embodiment of the present invention. As illustrated in FIG. 5E, the graph morphism 593 (F) is a pair of mappings: a mapping of an edge between two nodes and a mapping of the two nodes. Moreover, the graph morphism 593 (F) has the property that s(F(e))=F(s(e)) and t(F(e))=F(t(e)).
  • Referring to FIG. 2, when [0068] graphs 223 obey categorical constraints (i.e., they respect identity and composition on the mathematical objects),-the graphs 223 are categories 224. Graphs 223 can also be regarded as mathematical objects in their own right (i.e., making mathematical objects out of maps). This was done above when a change in the semantic state 212 of a relationship was allowed to be a semantic event 213.
  • With reference to FIG. 2 again, when the [0069] morphisms 222 between categories 224 obey categorical constraints (i.e., the identity maps to the identity, the morphisms respect composition), the morphisms 222 are called functors 225. FIG. 5B illustrates a first functor 530, a second functor 531, a first category 536, and a second category 535 in accordance with an embodiment of the present invention. The functor1 530 and the functor2 531 are directed from the categoryl 536 to the category2 535.
  • In FIG. 2, if the [0070] functors 225 map according to categorical constraints, the functors 225 are called natural transformations 226. This is the end of the line:
  • Regarding [0071] functors 225 as objects and natural transformations 226 as morphisms 222 produces a category 224, allowing use of the categorical structures tools 220 described above. FIG. 5C illustrates a natural transformation 580 in accordance with an embodiment of the present invention. The functor1 581 is directed from the category1 583 to the category2 584. The functor2 582 is directed from the category3 585 to the category4 586. The natural transformation 580 is directed from functor1 581 to functor2 582.
  • With reference to FIG. 2 again, a map from part of a semantic description into a semantic description is defined. This can be done by relying on a [0072] characteristic function 227 to the part in question, composed with a map to the target mental space 216. All of this is categorical structure. Moreover, the spaces from which structure is generated are required to have such characteristic functions 227. In addition, a large number of relationships are required. Lastly, it is possible to form product spaces in which to create these relationships. In sum, the categorical structures described above enable creation of any type or kind of semantic description for content data.
  • Semantic Mosaic Description
  • In an embodiment of the present invention, the semantic entity tools [0073] 210 (FIG. 2) and the categorical structure tools 220 (FIG. 2) facilitate creation of a semantic mosaic description for content data. The semantic mosaic description is formed from multiple semantic descriptions. These semantic descriptions are integrated with each other such that each semantic description is modified at a local level within localized regions without substantially changing each semantic description outside these localized regions. In particular, the semantic mosaic description facilitates navigation or browsing through the multiple semantic descriptions and the content data.
  • FIG. 6 illustrates a [0074] semantic description1 610 and a semantic description2 660 in accordance with an embodiment of the present invention. The semantic description1 610 and the semantic description2 660 were formed using the semantic entity tools 210 (FIG. 2) and the categorical structure tools 220 (FIG. 2). By integrating or blending (as described with respect to mental spaces in FIG. 4) the localized region 615 of the semantic description1 610 and the localized region 665 of the semantic description2 660, a semantic mosaic description is formed from the semantic description1 610 and the semantic description2 660. More importantly, the semantic description1 610 and the semantic description2 660 are not substantially changed outside of the localized regions 615 and 665 when they form the semantic mosaic description. It should be understood that any number of semantic descriptions can be integrated into a semantic mosaic description.
  • FIG. 7 illustrates a semantic [0075] mosaic description 750 based on the semarntic description1 610 of FIG. 6 and the semantic description2 of FIG. 6. The semantic mosaic description 750 provides several benefits. First, the semantic mosaic description 750 enables additional semantic information to be added to a semantic description where necessary or needed without affecting the entire semantic description. Moreover, the semantic mosaic description 750 can represent a complete semantic description which is formed from multiple partial semantic descriptions. Additionally, the semantic mosaic description can facilitate navigating or browsing through the semantic description1 610 and 660 as is done with content data such as audio-visual material. If the localized regions 615 and 665 (FIG. 6) have common elements, the transitions within the semantic mosaic description 750 are smooth. More importantly, as a whole the semantic mosaic description 750 may or may not semantically describe something, but within regions of the semantic mosaic description 750, something is semantically described.
  • Distributed Semantic Description
  • In an embodiment of the present invention, the semantic entity tools [0076] 210 (FIG. 2) and the categorical structure tools 220 (FIG. 2) facilitate creation of a semantic description for content data using multiple component semantic descriptions stored remotely from the content data. Reference information can be associated with the content data, whereas the reference information includes the identity of the component semantic descriptions needed to form the semantic description, the location of these component semantic descriptions, and the manner of processing these component semantic descriptions to form the semantic description. When the semantic description is desired, the component semantic descriptions identified in the reference information are retrieved (e.g., from a location on a network, a control dictionary, etc.). Then, the semantic description is formed in the manner specified in the reference information using the component semantic descriptions. Thus, the semantic description does not have to be stored in a discrete location, saving storage resources and promoting re-use of component semantic descriptions.
  • FIG. 8 illustrates formation of a [0077] semantic description 840 for content data 805 using multiple component semantic descriptions stored in locations on a network 850 in accordance with an embodiment of the present invention. As illustrated in FIG. 8, a plurality of component semantic descriptions 830A-830E are distributively stored in a plurality of locations on a network 850. In particular, the plurality of component semantic descriptions 830A-830E are stored remotely from the content data 805. The network 850 can be the Internet 850 or any other type of network. A semantic description 840 is formed from copies of one or more of the component semantic descriptions 830A-830B.
  • A complicated semantic description can be formed quickly and easily by referencing, adding new semantic information to, modifying, combining, or extracting partial semantic descriptions from the component [0078] semantic descriptions 830A-830E. For example, the semantic description for an elaborate wedding can be formed by using the distributively stored component semantic descriptions of a basic wedding, a fancy wedding gown, a stretch limousine, an expensive wedding cake, etc. These component semantic descriptions are modified and combined to form the semantic description for the elaborate wedding. Additionally, partial semantic descriptions can be extracted from the component semantic descriptions and then combined and/or modified with other component semantic descriptions. Moreover, the semantic description 840 can be generated when needed, reducing the demand for storage resources and encouraging re-use of component semantic descriptions 830A-830E.
  • Re-use of component [0079] semantic descriptions 830A-830E leads to standardization of semantic descriptions. Thus, applications such as the MPEG-7 standard are better able to handle and process the semantic descriptions.
  • In FIG. 8, the [0080] content data 805 includes reference information 810. The computer system 820 or any other computational system such as a MPEG-7 device utilizes the reference information 810 to generate the semantic description 840 for the content data 805. In particular, the reference information 810 includes the identity of the component semantic descriptions 830A-830B needed to form the semantic description 840, the location of these component semantic descriptions 830A-830B, and the manner of processing these component semantic descriptions 830A-830B to form the semantic description 840. It should be understood that the reference information 810 can have any other type of information.
  • Since the plurality of component [0081] semantic descriptions 830A-830E are distributively stored in a plurality of locations on a network 850, each component semantic description 830A-830E is assigned a uniform resource identifier (URI) to facilitate access to the component semantic descriptions 830A-830E. In practice, the reference information 810 has the URI for the component semantic descriptions 830A830E needed to form the semantic description 840. The computer system 820 or any other computational system such as a MPEG-7 device utilizes the URI(s) to retrieve the corresponding component semantic descriptions 830A-830B, as illustrated in FIG. 8.
  • In an embodiment, each component [0082] semantic description 830A-830E has information pertaining to its use. This information can indicate whether the component semantic description can be subsumed (i.e., can be embedded in another semantic description without changing its intended meaning). Moreover, this information can indicate whether the component semantic description can be subdivided (i.e., admits subdivisions which make the extraction of subsets of its semantic information natural). In addition, this information can indicate whether the component semantic description can be transformed. Furthermore, this information can indicate whether the component semantic description is transitive (i.e., functions as a subset if embedded in another semantic description).
  • FIG. 9 illustrates formation of a [0083] semantic description 840 for content data 805 using multiple component semantic descriptions stored in a control dictionary 860 in accordance with an embodiment of the present invention. The discussion of FIG. 8 is applicable to FIG. 9. Moreover, the plurality of component semantic descriptions 83OA-830E are distributively stored in a control dictionary 860 rather than in a plurality of locations on a network. For example, semantic descriptions pertaining to mathematics can be generated from component semantic descriptions retrieved from a control dictionary 860 emphasizing mathematical terms. An index value associated with the control dictionary 860 can be utilized to access the component semantic descriptions stored in the control dictionary 860. It is possible to have a plurality of control dictionaries and different types of control dictionaries.
  • FIG. 10 illustrates a flow chart showing a method [0084] 1000 of forming a semantic description for content data using multiple component semantic descriptions stored remotely from the content data in accordance with an embodiment of the present invention. Reference is made to FIGS. 8 and 9.
  • At [0085] step 1005, the method 1000 in accordance with an embodiment of the present invention begins.
  • Continuing at [0086] step 1010, numerous component semantic descriptions 830A-830E are distributively stored. Specifically, the numerous component semantic descriptions 830A-830E are stored remotely from the content data. The component semantic descriptions 830A-830E can be stored in locations on a network 850. Alternatively, the component semantic descriptions 830A-830E can be stored in one or more control dictionaries 860. In addition, the component semantic descriptions 830A-830E can have generic semantic information or specific semantic information.
  • Furthermore at [0087] step 1015, reference information 810 (configured as described above) is associated with the content data 805. This association can take place in a real-time environment or in a non real-time environment, whereas a real-time environment means that the reference information 810 is generated at the same time as the content data 805 is being captured.
  • At [0088] step 1020, it is determined whether to generate the specific semantic description 840 from one or more component semantic descriptions 830A-830E. For example, the computer system 820 or any other computational system such as a MPEG-7 device may receive a request for the specific semantic description 840 for the content data 805 in order to display, search, index, filter, or otherwise process the content data 805. At step 1035, the method 1000 ends if the specific semantic description 840 is not needed.
  • Otherwise, at [0089] step 1025, the computer system 820 or any other computational system such as a MPEG-7 device retrieves the component semantic descriptions 830A-830B identified by the reference information 810 from a network 850 or from a control dictionary 860.
  • At [0090] step 1030, the computer system 820 or any other computational system such as a MPEG-7 device generates the specific semantic description 840 using the retrieved component semantic descriptions 830A-830B and the reference information 810 which indicates the manner of processing these component semantic descriptions 830A-830B to form the specific semantic description 840. In particular, the reference information 810 indicates the manner of referencing, adding new semantic information to, modifying, combining, or extracting partial semantic descriptions from the component semantic descriptions 830A-830B.
  • The foregoing descriptions of specific embodiments of the present invention have been presented for purposes of illustration and description. They are not intended to be exhaustive or to limit the invention to the precise forms disclosed, and obviously many modifications and variations are possible in light of the above teaching. The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, to thereby enable others skilled in the art to best utilize the invention and various embodiments with various modifications as are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the Claims appended hereto and their equivalents. [0091]

Claims (35)

What is claimed is:
1. A method of forming a semantic description for content data, comprising the steps of:
a) retrieving one or more of a plurality of component semantic descriptions stored remotely from said content data according to reference information associated with said content data; and
b) generating said semantic description using said one or more component semantic descriptions and said reference information.
2. A method as recited in claim 1 wherein said step b) includes modifying said one or more component semantic descriptions to generate said semantic description.
3. A method as recited in claim 1 wherein said step b) includes extracting a partial semantic description from said one or more component semantic descriptions to generate said semantic description.
4. A method as recited in claim 1 wherein said step b) includes combining said one or more component semantic descriptions to generate said semantic description.
5. A method as recited in claim 1 wherein said steps a) and b) are performed in response to a request for said semantic description.
6. A method as recited in claim 1 wherein said plurality of component semantic descriptions is distributively stored in a plurality of locations on a network.
7. A method as recited in claim 6 wherein said network is the Internet.
8. A method as recited in claim 7 further including the step of:
assigning a uniform resource identifier (URI) to each component semantic description stored on the Internet to facilitate access.
9. A method as recited in claim 1 wherein said plurality of component semantic descriptions is stored in a control dictionary.
10. A computer system comprising:
a bus;
a processor coupled to said bus; and
a memory device coupled to said bus and having computer-executable instructions for performing a method of forming a semantic description for content data, said method comprising the steps of:
a) retrieving one or more of a plurality of component semantic descriptions stored remotely from said content data according to reference information associated with said content data; and
b) generating said semantic description using said one or more component semantic descriptions and said reference information.
11. A computer system as recited in claim 10 wherein said step b) includes modifying said one or more component semantic descriptions to generate said semantic description.
12. A computer system as recited in claim 10 wherein said step b) includes extracting a partial semantic description from said one or more component semantic descriptions to generate said semantic description.
13. A computer system as recited in claim 10 wherein said step b) includes combining said one or more component semantic descriptions to generate said semantic description.
14. A computer system as recited in claim 10 wherein said steps a) and b) are performed in response to a request for said semantic description.
15. A computer system as recited in claim 10 wherein said plurality of component semantic descriptions is distributively stored in a plurality of locations on a network.
16. A computer system as recited in claim 15 wherein said network is the Internet.
17. A computer system as recited in claim 16 wherein each component semantic description stored on the Internet has a uniform resource identifier (URI) to facilitate access.
18. A computer system as recited in claim 10 wherein said plurality of component semantic descriptions is stored in a control dictionary.
19. A semantic description for content data, comprising:
one or more component semantic descriptions which are retrieved from a plurality of component semantic descriptions stored remotely from said content data according to reference information associated with said content data, wherein said one or more component semantic descriptions are processed based on said reference information to form said semantic description.
20. A semantic description as recited in claim 19 wherein said one or more component semantic descriptions are modified to form said semantic description.
21. A semantic description as recited in claim 19 wherein a partial semantic description is extracted from said one or more component semantic descriptions to form said semantic description.
22. A semantic description as recited in claim 19 wherein said one or more component semantic descriptions are combined to form said semantic description.
23. A semantic description as recited in claim 19 wherein said one or more component semantic descriptions are retrieved in response to a request for said semantic description.
24. A semantic description as recited in claim 19 wherein said plurality of component semantic descriptions is distributively stored in a plurality of locations on a network.
25. A semantic description as recited in claim 24 wherein said network is the Internet.
26. A semantic description as recited in claim 25 wherein each component semantic description stored on the Internet has a uniform resource identifier (URI) to facilitate access.
27. A semantic description as recited in claim 19 wherein said plurality of component semantic descriptions is stored in a control dictionary.
28. A method of forming a semantic description for content data, comprising the steps of:
a) retrieving one or more of a plurality of component semantic descriptions stored remotely from said content data; and
b) generating said semantic description using said one or more component semantic descriptions.
29. A method as recited in claim 28 wherein said step b) includes modifying said one or more component semantic descriptions to generate said semantic description.
30. A method as recited in claim 28 wherein said step b) includes extracting a partial semantic description from said one or more component semantic descriptions to generate said semantic description.
31. A method as recited in claim 28 wherein said step b) includes combining said one or more component semantic descriptions to generate said semantic description.
32. A method as recited in claim 28 wherein said plurality of component semantic descriptions is distributively stored in a plurality of locations on a network.
33. A method as recited in claim 28 wherein said network is the Internet.
34. A method as recited in claim 33 further including the step of:
assigning a uniform resource identifier (URI) to each component semantic description stored on the Internet to facilitate access.
35. A method as recited in claim 28 wherein said plurality of component semantic descriptions is stored in a control dictionary.
US09/809,578 2000-03-14 2001-03-14 Generating semantic descriptions for content data from component semantic descriptions stored remotely from the content data Abandoned US20020013783A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US09/809,578 US20020013783A1 (en) 2000-03-14 2001-03-14 Generating semantic descriptions for content data from component semantic descriptions stored remotely from the content data

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US18962600P 2000-03-14 2000-03-14
US18920200P 2000-03-14 2000-03-14
US19131200P 2000-03-21 2000-03-21
US09/809,578 US20020013783A1 (en) 2000-03-14 2001-03-14 Generating semantic descriptions for content data from component semantic descriptions stored remotely from the content data

Publications (1)

Publication Number Publication Date
US20020013783A1 true US20020013783A1 (en) 2002-01-31

Family

ID=27392537

Family Applications (2)

Application Number Title Priority Date Filing Date
US09/809,578 Abandoned US20020013783A1 (en) 2000-03-14 2001-03-14 Generating semantic descriptions for content data from component semantic descriptions stored remotely from the content data
US09/809,684 Expired - Fee Related US7319951B2 (en) 2000-03-14 2001-03-14 Application of category theory and cognitive science to design of semantic descriptions for content data

Family Applications After (1)

Application Number Title Priority Date Filing Date
US09/809,684 Expired - Fee Related US7319951B2 (en) 2000-03-14 2001-03-14 Application of category theory and cognitive science to design of semantic descriptions for content data

Country Status (3)

Country Link
US (2) US20020013783A1 (en)
AU (1) AU2001255806A1 (en)
WO (1) WO2001069445A2 (en)

Cited By (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030221171A1 (en) * 2001-11-21 2003-11-27 Godfrey Rust Data dictionary method
US20050131920A1 (en) * 2003-10-17 2005-06-16 Godfrey Rust Computer implemented methods and systems for representing multiple data schemas and transferring data between different data schemas within a contextual ontology
US20070001970A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002061A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001973A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001972A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001975A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001983A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001971A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002670A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001968A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Display device and electronic instrument
US20070001974A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070000971A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002063A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001984A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002509A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002188A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001969A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013074A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013684A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013687A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013685A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013634A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013707A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013706A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070016700A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070187762A1 (en) * 2006-02-10 2007-08-16 Seiko Epson Corporation Integrated circuit device and electronic instrument
KR100823289B1 (en) 2006-08-28 2008-04-17 한국과학기술정보연구원 Apparatus and method for managing outcomes based on URI
US20080112254A1 (en) * 2005-06-30 2008-05-15 Seiko Epson Corporation Integrated circuit device and electronic instrument
US8339352B2 (en) 2005-09-09 2012-12-25 Seiko Epson Corporation Integrated circuit device and electronic instrument
US9031889B1 (en) 2012-11-09 2015-05-12 DataInfoCom USA Inc. Analytics scripting systems and methods
US9230211B1 (en) * 2012-11-09 2016-01-05 DataInfoCom USA, Inc. Analytics scripting systems and methods

Families Citing this family (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7191250B1 (en) * 2001-03-19 2007-03-13 Palmsource, Inc. Communication protocol for wireless data exchange via a packet transport based system
US6741990B2 (en) * 2001-05-23 2004-05-25 Intel Corporation System and method for efficient and adaptive web accesses filtering
US7925139B2 (en) * 2001-12-03 2011-04-12 Sony Corporation Distributed semantic descriptions of audiovisual content
DK200301926A (en) * 2003-12-23 2005-06-24 Eskebaek Thomas Knowledge management system with ontology-based methods for the dissemination of knowledge and the search for knowledge
US7860873B2 (en) 2004-07-30 2010-12-28 International Business Machines Corporation System and method for automatic terminology discovery
US9177248B2 (en) 2005-03-30 2015-11-03 Primal Fusion Inc. Knowledge representation systems and methods incorporating customization
US9378203B2 (en) 2008-05-01 2016-06-28 Primal Fusion Inc. Methods and apparatus for providing information of interest to one or more users
US10002325B2 (en) 2005-03-30 2018-06-19 Primal Fusion Inc. Knowledge representation systems and methods incorporating inference rules
US8849860B2 (en) 2005-03-30 2014-09-30 Primal Fusion Inc. Systems and methods for applying statistical inference techniques to knowledge representations
US9104779B2 (en) 2005-03-30 2015-08-11 Primal Fusion Inc. Systems and methods for analyzing and synthesizing complex knowledge representations
US7596574B2 (en) * 2005-03-30 2009-09-29 Primal Fusion, Inc. Complex-adaptive system for providing a facted classification
US7606781B2 (en) * 2005-03-30 2009-10-20 Primal Fusion Inc. System, method and computer program for facet analysis
US7849090B2 (en) * 2005-03-30 2010-12-07 Primal Fusion Inc. System, method and computer program for faceted classification synthesis
US7844565B2 (en) 2005-03-30 2010-11-30 Primal Fusion Inc. System, method and computer program for using a multi-tiered knowledge representation model
US8955105B2 (en) * 2007-03-14 2015-02-10 Microsoft Corporation Endpoint enabled for enterprise security assessment sharing
US8413247B2 (en) * 2007-03-14 2013-04-02 Microsoft Corporation Adaptive data collection for root-cause analysis and intrusion detection
US8959568B2 (en) * 2007-03-14 2015-02-17 Microsoft Corporation Enterprise security assessment sharing
US20080229419A1 (en) * 2007-03-16 2008-09-18 Microsoft Corporation Automated identification of firewall malware scanner deficiencies
US8316036B2 (en) 2007-08-31 2012-11-20 Microsoft Corporation Checkpointing iterators during search
US8280721B2 (en) * 2007-08-31 2012-10-02 Microsoft Corporation Efficiently representing word sense probabilities
US8712758B2 (en) * 2007-08-31 2014-04-29 Microsoft Corporation Coreference resolution in an ambiguity-sensitive natural language processing system
US7984032B2 (en) * 2007-08-31 2011-07-19 Microsoft Corporation Iterators for applying term occurrence-level constraints in natural language searching
US8346756B2 (en) * 2007-08-31 2013-01-01 Microsoft Corporation Calculating valence of expressions within documents for searching a document index
US8041697B2 (en) * 2007-08-31 2011-10-18 Microsoft Corporation Semi-automatic example-based induction of semantic translation rules to support natural language search
US8229730B2 (en) * 2007-08-31 2012-07-24 Microsoft Corporation Indexing role hierarchies for words in a search index
US20090070322A1 (en) * 2007-08-31 2009-03-12 Powerset, Inc. Browsing knowledge on the basis of semantic relations
US8639708B2 (en) * 2007-08-31 2014-01-28 Microsoft Corporation Fact-based indexing for natural language search
US8463593B2 (en) * 2007-08-31 2013-06-11 Microsoft Corporation Natural language hypernym weighting for word sense disambiguation
US8868562B2 (en) * 2007-08-31 2014-10-21 Microsoft Corporation Identification of semantic relationships within reported speech
US8229970B2 (en) * 2007-08-31 2012-07-24 Microsoft Corporation Efficient storage and retrieval of posting lists
CN106845645B (en) 2008-05-01 2020-08-04 启创互联公司 Method and system for generating semantic network and for media composition
US8676732B2 (en) 2008-05-01 2014-03-18 Primal Fusion Inc. Methods and apparatus for providing information of interest to one or more users
US9361365B2 (en) 2008-05-01 2016-06-07 Primal Fusion Inc. Methods and apparatus for searching of content using semantic synthesis
US8374859B2 (en) * 2008-08-20 2013-02-12 Universal Entertainment Corporation Automatic answering device, automatic answering system, conversation scenario editing device, conversation server, and automatic answering method
CN106250371A (en) 2008-08-29 2016-12-21 启创互联公司 For utilizing the definition of existing territory to carry out the system and method that semantic concept definition and semantic concept relation is comprehensive
US8112365B2 (en) * 2008-12-19 2012-02-07 Foster Scott C System and method for online employment recruiting and evaluation
US20110060645A1 (en) * 2009-09-08 2011-03-10 Peter Sweeney Synthesizing messaging using context provided by consumers
US9292855B2 (en) 2009-09-08 2016-03-22 Primal Fusion Inc. Synthesizing messaging using context provided by consumers
US20110060644A1 (en) * 2009-09-08 2011-03-10 Peter Sweeney Synthesizing messaging using context provided by consumers
US8365059B2 (en) * 2009-11-03 2013-01-29 Oto Technologies, Llc E-reader semantic text manipulation
US9262520B2 (en) 2009-11-10 2016-02-16 Primal Fusion Inc. System, method and computer program for creating and manipulating data structures using an interactive graphical interface
US9244965B2 (en) * 2010-02-22 2016-01-26 Thoughtwire Holdings Corp. Method and system for sharing data between software systems
US9235806B2 (en) 2010-06-22 2016-01-12 Primal Fusion Inc. Methods and devices for customizing knowledge representation systems
US10474647B2 (en) 2010-06-22 2019-11-12 Primal Fusion Inc. Methods and devices for customizing knowledge representation systems
US11294977B2 (en) 2011-06-20 2022-04-05 Primal Fusion Inc. Techniques for presenting content to a user based on the user's preferences
US20120324367A1 (en) 2011-06-20 2012-12-20 Primal Fusion Inc. System and method for obtaining preferences with a user interface
US10282378B1 (en) * 2013-04-10 2019-05-07 Christopher A. Eusebi System and method for detecting and forecasting the emergence of technologies
US9355372B2 (en) * 2013-07-03 2016-05-31 Thomson Reuters Global Resources Method and system for simplifying implicit rhetorical relation prediction in large scale annotated corpus
EP3961508A1 (en) 2020-08-31 2022-03-02 Xephor Solutions GmbH System and method for data processing and computation
CN112767924A (en) * 2021-02-26 2021-05-07 北京百度网讯科技有限公司 Voice recognition method and device, electronic equipment and storage medium

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5197005A (en) * 1989-05-01 1993-03-23 Intelligent Business Systems Database retrieval system having a natural language interface
US6233545B1 (en) * 1997-05-01 2001-05-15 William E. Datig Universal machine translator of arbitrary languages utilizing epistemic moments
US6199068B1 (en) * 1997-09-11 2001-03-06 Abb Power T&D Company Inc. Mapping interface for a distributed server to translate between dissimilar file formats
US6778970B2 (en) * 1998-05-28 2004-08-17 Lawrence Au Topological methods to organize semantic network data flows for conversational applications
DE69932029D1 (en) * 1998-08-05 2006-08-03 Koninkl Philips Electronics Nv METHOD AND DEVICE FOR PRODUCING A STICK IMAGE
US6654813B1 (en) * 1998-08-17 2003-11-25 Alta Vista Company Dynamically categorizing entity information
US6240406B1 (en) * 1998-08-31 2001-05-29 The Trustees Of The University Of Pennsylvania System and method for optimizing queries with aggregates and collection conversions
US6236395B1 (en) * 1999-02-01 2001-05-22 Sharp Laboratories Of America, Inc. Audiovisual information management system
US6598039B1 (en) * 1999-06-08 2003-07-22 Albert-Inc. S.A. Natural language interface for searching database

Cited By (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030221171A1 (en) * 2001-11-21 2003-11-27 Godfrey Rust Data dictionary method
US7124364B2 (en) * 2001-11-21 2006-10-17 Contecs:Dd Llc Data dictionary method
US20050131920A1 (en) * 2003-10-17 2005-06-16 Godfrey Rust Computer implemented methods and systems for representing multiple data schemas and transferring data between different data schemas within a contextual ontology
US20070001969A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001972A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001973A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013684A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001975A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013687A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001971A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002670A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001968A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Display device and electronic instrument
US20070001974A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070000971A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002063A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001984A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002509A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002188A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001970A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013074A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070002061A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070001983A1 (en) * 2005-06-30 2007-01-04 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013685A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013634A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013707A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070013706A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070016700A1 (en) * 2005-06-30 2007-01-18 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20110128274A1 (en) * 2005-06-30 2011-06-02 Seiko Epson Corporation Integrated Circuit Device and Electronic Instrument
US7522441B2 (en) 2005-06-30 2009-04-21 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20080112254A1 (en) * 2005-06-30 2008-05-15 Seiko Epson Corporation Integrated circuit device and electronic instrument
US8339352B2 (en) 2005-09-09 2012-12-25 Seiko Epson Corporation Integrated circuit device and electronic instrument
US20070187762A1 (en) * 2006-02-10 2007-08-16 Seiko Epson Corporation Integrated circuit device and electronic instrument
US8188545B2 (en) 2006-02-10 2012-05-29 Seiko Epson Corporation Integrated circuit device and electronic instrument
KR100823289B1 (en) 2006-08-28 2008-04-17 한국과학기술정보연구원 Apparatus and method for managing outcomes based on URI
US9031889B1 (en) 2012-11-09 2015-05-12 DataInfoCom USA Inc. Analytics scripting systems and methods
US9230211B1 (en) * 2012-11-09 2016-01-05 DataInfoCom USA, Inc. Analytics scripting systems and methods
US9424518B1 (en) 2012-11-09 2016-08-23 DataInfoCom USA, Inc. Analytics scripting systems and methods
US10740679B1 (en) * 2012-11-09 2020-08-11 DataInfoCom USA, Inc. Analytics scripting systems and methods

Also Published As

Publication number Publication date
US20010047254A1 (en) 2001-11-29
US7319951B2 (en) 2008-01-15
AU2001255806A1 (en) 2001-09-24
WO2001069445A3 (en) 2004-01-08
WO2001069445A2 (en) 2001-09-20

Similar Documents

Publication Publication Date Title
US7319951B2 (en) Application of category theory and cognitive science to design of semantic descriptions for content data
US7089543B2 (en) Use of formal logic specification in construction of semantic descriptions
Schetinger et al. Doom or deliciousness: Challenges and opportunities for visualization in the age of generative models
Sebe et al. Personalized multimedia retrieval: the new trend?
Spala et al. Extending MPEG-7 for efficient annotation of complex web 3D scenes
Brunow Naming, shaming, framing: Ambivalence of queer visibility in audiovisual archives.
US20020059290A1 (en) Formal verification of semantic DS
Chmielewski Describing interactivity of 3d content
Seremeti et al. Multimedia ontologies
Benitez¹ et al. Description of a single multimedia document
US7925139B2 (en) Distributed semantic descriptions of audiovisual content
Ma et al. VideoMap: An interactive and scalable visualization for exploring video content
Bürger et al. Metadata standards and ontologies for multimedia content
Schellner et al. CULTOS: towards a world-wide digital collection of exchangeable units of multimedia content for intertextual studies
Asirelli et al. MultiMedia Metadata Management: a Proposal for an Infrastructure.
Radev et al. Film video modeling
Mohammed Application of Multimedia Technology in Database and IT Service Management
Chen et al. Development and application of the semantic annotation framework for digital images
Chmielewski Metadata model for interactions of 3D objects
JP2005512213A5 (en)
D’Armenio et al. Introduction: hyper-visuality: images in the era of social platforms, digital archives and computational economies
US20130156399A1 (en) Embedding content in rich media
Brasoveanu et al. The Role of Visual Rhetoric in Semantic Multimedia: Strategies for Decision Making in Times of Crisis
Bürger et al. CHAPTER III. 8 METADATA STANDARDS AND ONTOLOGIES FOR MULTIMEDIA CONTENT
Chris Authorship and Auteurism

Legal Events

Date Code Title Description
AS Assignment

Owner name: SONY ELECTRONICS, INC., NEW JERSEY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RISING, HAWLEY, III;REEL/FRAME:011624/0929

Effective date: 20010314

Owner name: SONY CORPORATION OF JAPAN, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RISING, HAWLEY, III;REEL/FRAME:011624/0929

Effective date: 20010314

AS Assignment

Owner name: SONY ELECTRONICS, INC., NEW JERSEY

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE NAME OF THE ASSIGNEE. FILED ON 03-14-2001, RECORDED ON REEL 011624 FRAME 0929;ASSIGNOR:RISING, HAWLEY III;REEL/FRAME:012727/0827

Effective date: 20010314

Owner name: SONY CORPORATION, JAPAN

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE NAME OF THE ASSIGNEE. FILED ON 03-14-2001, RECORDED ON REEL 011624 FRAME 0929;ASSIGNOR:RISING, HAWLEY III;REEL/FRAME:012727/0827

Effective date: 20010314

AS Assignment

Owner name: SONY ELECTRONICS INC., NEW JERSEY

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TABATABAI, ALI;REEL/FRAME:014405/0143

Effective date: 20030624

Owner name: SONY CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:TABATABAI, ALI;REEL/FRAME:014405/0143

Effective date: 20030624

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION

AS Assignment

Owner name: SONY CORPORATION, JAPAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:SONY CORPORATION OF JAPAN;REEL/FRAME:045208/0908

Effective date: 20180131