KR20110112211A - System and method for providing multimedia service in a communication system - Google Patents
System and method for providing multimedia service in a communication system Download PDFInfo
- Publication number
- KR20110112211A KR20110112211A KR1020110030397A KR20110030397A KR20110112211A KR 20110112211 A KR20110112211 A KR 20110112211A KR 1020110030397 A KR1020110030397 A KR 1020110030397A KR 20110030397 A KR20110030397 A KR 20110030397A KR 20110112211 A KR20110112211 A KR 20110112211A
- Authority
- KR
- South Korea
- Prior art keywords
- effect
- sensory
- multimedia
- information
- binary
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/235—Processing of additional data, e.g. scrambling of additional data or processing content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/25—Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
- H04N21/258—Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/85—Assembly of content; Generation of multimedia applications
- H04N21/854—Content authoring
- H04N21/8543—Content authoring using a description language, e.g. Multimedia and Hypermedia information coding Expert Group [MHEG], eXtensible Markup Language [XML]
Abstract
The present invention relates to a multimedia service providing system and method for providing a variety of high-capacity multimedia content and various sensory effects of the multimedia content to users in real time at high speed. According to the service request, the multimedia content of the multimedia service is generated, the sensory effect information indicating the sensory effects of the multimedia content is generated, and the sensory effect information is represented in binary notation using a binary representation encoding scheme. Encoding, converting the sensory effect information encoded in the binary notation into command information in binary notation, and converting the multimedia content and the sensory effects through a device command according to the control information of the binary notation. In real time It is provided to the user.
Description
BACKGROUND OF THE
In a communication system, active research is being conducted to provide users with services of various quality of service (QoS: QoS) having a high transmission speed. Such a communication system transmits various types of service data to users quickly and stably through limited resources according to service requests of users who want to receive various types of services, thereby providing services requested by each user. Measures have been proposed.
On the other hand, in the current communication system, a method for transmitting a large amount of service data at high speed has been proposed according to various service needs of users, and in particular, a large amount of multimedia data corresponding to the service needs of users who want to receive various multimedia services is proposed. There is an active research on the method for high speed transmission. In other words, through a communication system, users want to be provided with various multimedia services of higher quality, and in particular, multimedia contents corresponding to multimedia services are provided as well as various sensory effects of the multimedia contents, thereby providing higher quality multimedia services. To be provided.
However, in the current communication system, there is a limitation in providing multimedia services requested by the users by transmitting the multimedia contents according to the multimedia service requests of the users. In particular, in the current communication system, as described above, in order to meet various demands of higher quality multimedia services of users, a specific scheme for providing not only multimedia contents but also various sensory effects of the multimedia contents to users is still proposed. Not. That is, in the current communication system, a concrete method for providing various high quality multimedia services to each user in real time by transmitting the multimedia content and various sensory effects at high speed has not been proposed.
Accordingly, there is a need for a method for providing a high-quality, high-capacity multimedia service at high speed in accordance with the service needs of users in a communication system, and in particular, a high-quality, high-capacity multimedia service required by each user in real time.
Accordingly, an object of the present invention is to provide a system and method for providing a multimedia service in a communication system.
Another object of the present invention is to provide a multimedia service providing system and method for providing a variety of high quality multimedia services to users in real time in accordance with a service request of a user in a communication system.
Another object of the present invention is to transmit multimedia contents of a multimedia service that each user wants to be provided in a communication system, and various sensory effects of the multimedia contents at high speed, thereby realizing various high quality multimedia services in real time. The present invention provides a system and method for providing a multimedia service to each user.
A system of the present invention for achieving the above object, in a system for providing a multimedia service in a communication system, according to the service request of the multimedia service to be provided by users, the multimedia content of the multimedia service, and the multimedia content of A service provider for providing sensory effect information indicative of sensory effects; A user server configured to receive multimedia data including the multimedia content and the sensory effect information, and convert the sensory effect information into control information from the multimedia data and provide the command information; And user devices that provide the multimedia contents and the sensory effects to the users in real time through a device command according to the control information.
Another system of the present invention for achieving the above object, in a system for providing a multimedia service in a communication system, according to the service request of the multimedia service to be provided by the user, generates the multimedia content of the multimedia service, A generator for generating sensory effect information indicative of sensory effects of multimedia content; An encoder for encoding the sensory effect information in binary notation using a binary representation coding scheme; And a transmitter for transmitting the multimedia data including the multimedia content and sensory effect information encoded in the binary notation.
In accordance with another aspect of the present invention, there is provided a method of providing a multimedia service in a communication system, the method comprising: generating multimedia content of the multimedia service according to a service request of a multimedia service to be provided by users, and generating the multimedia service. Generating sensory effect information indicating sensory effects of the content; Encoding the sensory effect information in binary notation using a binary representation coding scheme; Converting sensory effect information encoded in the binary notation into command information in binary notation; And providing the multimedia contents and the sensory effects to the users in real time through a device command according to the control information of the binary notation.
The present invention reliably provides various high quality multimedia services that each user wants to receive in a communication system, and in particular, provides multimedia contents of the multimedia service and various sensory effects of the multimedia contents to each user. can do. In addition, the present invention encodes information indicating various sensory effects of the multimedia content using a binary representation, thereby transmitting the multimedia content and various sensory effects of the multimedia content at high speed, and accordingly, The multimedia contents and sensory effects may be provided to each user in real time, that is, the various high quality multimedia services may be provided to the users in real time.
1 is a view schematically showing the structure of a multimedia service providing system according to an embodiment of the present invention.
2 is a diagram schematically illustrating a structure of a service provider in a multimedia service providing system according to an exemplary embodiment of the present invention.
3 is a diagram schematically illustrating a structure of a user server in a multimedia service providing system according to an exemplary embodiment of the present invention.
4 is a diagram schematically illustrating a structure of a user device in a multimedia service providing system according to an exemplary embodiment of the present invention.
5 is a diagram illustrating a location model of sensory effect metadata in a multimedia service providing system according to an exemplary embodiment of the present invention.
6 is a diagram illustrating a movement pattern in the sensory effect of the multimedia service providing system according to an embodiment of the present invention.
7 is a diagram illustrating a motion trace sample pattern in a sensory effect of a multimedia service providing system according to an exemplary embodiment of the present invention.
8 is a diagram illustrating an incline pattern in a sensory effect of a multimedia service providing system according to an exemplary embodiment of the present invention.
9 is a diagram illustrating a shake pattern in a sensory effect of a multimedia service providing system according to an exemplary embodiment of the present invention.
FIG. 10 is a diagram illustrating a wave pattern in a sensory effect of a multimedia service providing system according to an exemplary embodiment of the present invention. FIG.
11 is a diagram illustrating a spin pattern in the sensory effect of the multimedia service providing system according to an embodiment of the present invention.
12 is a diagram illustrating a turn pattern in the sensory effect of the multimedia service providing system according to an exemplary embodiment of the present invention.
FIG. 13 is a diagram illustrating a collide pattern in a sensory effect of a multimedia service providing system according to an exemplary embodiment of the present invention. FIG.
14 is a diagram illustrating a horizontal movement pattern in the sensory effect of the multimedia service providing system according to an embodiment of the present invention.
15 is a view illustrating a vertical movement pattern in the sensory effect of the multimedia service providing system according to an embodiment of the present invention.
16 is a view illustrating a direction incline pattern in a sensory effect of a multimedia service providing system according to an embodiment of the present invention.
17 is a diagram illustrating a direction shake pattern in a sensory effect of a multimedia service providing system according to an exemplary embodiment of the present invention.
18 is a diagram illustrating a shake motion distance in the sensory effect of the multimedia service providing system according to an exemplary embodiment of the present invention.
19 and 20 illustrate wave motion directions in a sensory effect of a multimedia service providing system according to an exemplary embodiment of the present invention.
21 and 22 are views illustrating a wave motion starting direction in a sensory effect of a multimedia service providing system according to an exemplary embodiment of the present invention.
FIG. 23 is a diagram illustrating a wave motion distance in a sensory effect of a multimedia service providing system according to an embodiment of the present invention. FIG.
24 is a diagram illustrating a turn pattern direction in a sensory effect of a multimedia service providing system according to an embodiment of the present invention.
25 is a diagram illustrating a horizontal colloid pattern in a sensory effect of a multimedia service providing system according to an embodiment of the present invention.
FIG. 26 illustrates a vertical colloid pattern in the sensory effect of the multimedia service providing system according to an exemplary embodiment of the present invention. FIG.
27 is a diagram schematically illustrating a multimedia service providing process of a multimedia service providing system according to an embodiment of the present invention.
Hereinafter, exemplary embodiments of the present invention will be described in detail with reference to the accompanying drawings. It should be noted that in the following description, only parts necessary for understanding the operation according to the present invention will be described, and descriptions of other parts will be omitted so as not to distract from the gist of the present invention.
The present invention proposes a multimedia service providing system and method for providing various high quality multimedia services in a high speed and real time in a communication system. Here, according to an embodiment of the present invention, multimedia contents of a multimedia service to be provided to each user and various sensory effects of the multimedia contents according to a service request of users who want to receive various high quality services are provided. By transmitting at high speed, it provides various high quality multimedia services required by each user in real time.
In addition, in an embodiment of the present invention, the multimedia content of the aforementioned multimedia service and various sensory effects of the multimedia content are transmitted at high speed by maximizing the resources available for providing the multimedia service to users. At this time, the multimedia content of the multimedia service that users want to provide is a large amount of data, most of the available resources are used for the transmission of the multimedia content. Accordingly, the resources available for the transmission of various sensory effects of the multimedia content that must be transmitted and provided indispensable for providing various multimedia services of high quality required by users are further limited, and as a result, various high quality multimedia services can be quickly And in order to provide to users in real time, high-speed transmission of the various sensory effects as well as high-speed transmission of a large amount of multimedia content is essential.
That is, in an embodiment of the present invention, in order to provide a multimedia service required by each user at high speed and in real time through resources available for providing various high quality multimedia services, the multimedia content is coded, and in particular, Information indicative of various sensory effects (hereinafter referred to as 'sensory effects information') is encoded using a binary representation to minimize the data size of the sensory effect information and thus The multimedia content and various sensory effects of the multimedia content are transmitted at high speed, and as a result, the multimedia content and sensory effects are provided to the users in real time, that is, the various high quality multimedia services are provided to the users in real time. do .
In addition, according to an embodiment of the present invention, in the Moving Picture Experts Group (MPEG) -V, various sensory effect information of multimedia are transmitted at high speed using binary code notation, that is, sensory effect data through binary notation. Alternatively, sensory effect metadata may be transmitted at high speed to provide various sensory effects of the multimedia content in real time to each user who is provided with a multimedia service.
Here, the embodiment of the present invention relates to the high-speed transmission of sensory effect information, that is, sensory effect data or sensory effect metadata in Part 3 of MPEG-V, and generates various high-quality multimedia services according to service requests of each user. A service provider which provides or sells the multimedia content of the multimedia service and encodes the multimedia content at high speed. In particular, various sensory effects of the multimedia content are encoded in binary notation, that is, the sensory effect information is encoded in a binary code notation. do. The service provider transmits the multimedia content and the sensory effect information encoded in the binary notation to a user server such as a home server at high speed.
In this case, as the service provider encodes and transmits the sensory effect information in binary notation, as described above, an extremely limited available resource for transmitting the sensory effect information, that is, used for transmitting a large amount of multimedia content is used. The sensory effect information is transmitted at high speed by maximizing the available resources except resources. Accordingly, the service provider transmits the sensory effect information as well as the multimedia content to the user server at high speed, and as a result, provides each user with various sensory effects of the multimedia content in real time.
Here, the user server outputs the multimedia service and transmits the multimedia content and the sensory effect information to user devices that provide the actual multimedia service to each user. The effect information is encoded in binary notation and converted into command information for device control of each user device, and the control information converted into binary notation is transmitted to each user device. Each of the user devices is controlled according to the control information converted into the binary notation to output various sensory effects, that is, to provide users with various sensory effects of the multimedia content in real time.
For example, Part 3 of the above-described MPEG-V defines a schema for efficiently describing various sensory effects that may appear in a scene of a multimedia content or a real environment. For example, when a wind blows in a specific scene of a movie, a wind-like realistic effect is described using a predefined schema and inserted into multimedia data, and when a home server reproduces a movie through the multimedia data, the multimedia data After extracting the sensory effect information from, and synchronizes with the user device capable of outputting the wind effect, such as a fan, to provide the user with the windy effect. In another example, the learner (ie, the user) purchases user devices that can produce various sensational effects and is located in the house, while the instructor (ie, the service provider) remotely delivers the lecture (ie, transmits multimedia data). It delivers a more realistic education, that is, a higher quality multimedia service by transmitting various realistic effects to the students according to the lecture contents (ie, multimedia contents).
In order to provide high quality multimedia services, sensory effect information provided together with multimedia contents may be described as an extensible markup language (XML) document. For example, when the service provider describes the sensory effect information as an XML document, the sensory effect information is transmitted to the user server as an XML document, and the user server that has received the sensory effect information of the XML document may describe the XML document. After analyzing, the sensory effect information of the analyzed XML document is analyzed.
Here, although the user server analyzes an XML document and sensory effect information, there may be a limitation in providing various high quality multimedia services to users at high speed and in real time. However, in the embodiment of the present invention, as described above, As information is encoded and transmitted in binary notation, analysis of XML documents and sensory effect information is unnecessary, thereby providing various high quality multimedia services to users at high speed and in real time. In other words, in the embodiment of the present invention, in the Part 3 of MPEG-V, sensory effect information is compressed and transmitted through a binary code notation method, not an XML document, and the number of bits used for transmitting the sensory effect information is reduced, that is, It reduces the amount of resources used for the transmission of sensory effect information, and sends sensory effect information at high speed and efficiency by omitting the analysis process of the XML document and sensory effect information. Next, the multimedia service providing system according to an exemplary embodiment of the present invention will be described in more detail with reference to FIG. 1.
1 is a diagram schematically illustrating a structure of a multimedia service providing system according to an exemplary embodiment of the present invention.
Referring to FIG. 1, the multimedia service providing system includes a
As described above, the
In this case, the
Here, the
The
In this case, when receiving the sensory effect information encoded in the binary notation, the
Here, the
In addition, the
In this case, as described above, the
Here, the
As such, in a system for providing a multimedia service according to an embodiment of the present invention, the
In other words, in the system for providing a multimedia service according to an embodiment of the present invention, the
In addition, the
2 is a diagram schematically illustrating a structure of a service provider in a multimedia service providing system according to an exemplary embodiment of the present invention.
Referring to FIG. 2, the
The
The
Here, the encoder 2 240 defines syntax, binary representation, and semantics of sensory effects corresponding to sensory effect information in binary notation encoding of the sensory effect information. . In addition, as the sensory effect information is encoded in binary notation, the data size of the sensory effect information is minimized. As described above, the
The
3 is a diagram schematically illustrating a structure of a user server in a multimedia service providing system according to an exemplary embodiment of the present invention.
Referring to FIG. 3, the
As described above, the
The
The
Here, the
Therefore, the
The encoder 3 340 encodes the converted control information using a binary coding scheme, that is, the control information is encoded in binary notation, and the control information is encoded in a binary code in the form of a stream. In other words, the encoder 3 340 becomes a device control stream encoder and outputs a device control stream encoded in binary notation for control information for controlling the device.
Here, as the control information is encoded in binary notation, the control information of the binary notation becomes a control signal of each of the
More specifically, when the
When the
In addition, when the
For example, when the
The transmitter 2 350 transmits the multimedia content included in the multimedia data and the control information encoded in the binary notation to the
4 is a diagram schematically illustrating a structure of a user device in a multimedia service providing system according to an exemplary embodiment of the present invention.
Referring to FIG. 4, the user device decodes the control information encoded in the multimedia content or the binary notation, the receiver 2 410, which receives the control information encoded in the multimedia content or the binary notation from the
The receiver 2 410 receives multimedia content transmitted from the transmitter 2 350 of the
The decoder 2 420 decodes the control information of the binary notation received in the form of a stream. In this case, since the control information encoded in the binary notation is a control information stream encoded in the binary code in the form of a stream, the decoder 2 420 is a device control stream decoder, and decodes the control information stream encoded in the binary notation. The decoded control information is transmitted to the
The
In other words, when the receiver 2 410 receives the control information of the XML document, the decoder 2 420 analyzes and confirms the control information of the XML document, and the
The
First, the sensory effect information, that is, the base datatypes and elements of the sensory effect metadata will be described. Syntax can be expressed as shown in Table 1 below. Here, Table 1 is a table showing the syntax of the sensory effect metadata.
<!-SEM Base Attributes->
<!-############################################## ##->
<attributeGroup name = "SEMBaseAttributes">
<attribute name = "activate" type = "boolean" use = "optional"/>
<attribute name = "duration" type = "positiveInteger" use = "optional"/>
<attribute name = "fade" type = "positiveInteger" use = "optional"/>
<attribute name = "alt" type = "anyURI" use = "optional"/>
<attribute name = "priority" type = "positiveInteger" use = "optional"/>
<attribute name = "location" type = "mpeg7: termReferenceType"
use = "optional"/>
<attributeGroup ref = "sedl: SEMAdaptabilityAttributes"/>
</ attributeGroup>
<simpleType name = "intensityValueType">
<restriction base = "float"/>
</ simpleType>
<simpleType name = "intensityRangeType">
<restriction>
<simpleType>
<list itemType = "float"/>
</ simpleType>
<length value = "2" fixed = "true"/>
</ restriction>
</ simpleType>
<!-############################################## ##->
<!-SEM Adaptability Attributes->
<!-############################################## ##->
<attributeGroup name = "SEMAdaptabilityAttributes">
<attribute name = "adaptType" type = "sedl: adaptTypeType" use = "optional"/>
<attribute name = "adaptRange" type = "sedl: adaptRangeType" default = "10"
use = "optional"/>
</ attributeGroup>
<simpleType name = "adaptTypeType">
<restriction base = "NMTOKEN">
<enumeration value = "Strict"/>
<enumeration value = "Under"/>
<enumeration value = "Over"/>
<enumeration value = "Both"/>
</ restriction>
</ simpleType>
<simpleType name = "adaptRangeType">
<restriction base = "unsignedInt">
<minInclusive value = "0"/>
<maxInclusive value = "100"/>
</ restriction>
</ simpleType>
<!-############################################## ##->
<!-SEM Base type->
<!-############################################## ##->
<complexType name = "SEMBaseType" abstract = "true">
<complexContent>
<restriction base = "anyType">
<attribute name = "id" type = "ID" use = "optional"/>
</ restriction>
</ complexContent>
</ complexType>
The basic data type and the binary coding notation or binary representation of elements of the sensory effect metadata may be represented as in Table 2 below. Here, Table 2 is a table showing a binary representation of the basic data type and elements of the sensory effect metadata.
In addition, the semantics of the basic data type and elements of the sensory effect metadata may be represented as shown in Table 3 below. Here, Table 3 is a table showing the SEM basic properties semantics (Semantics of the SEMBaseAttributes).
NOTE 2 The alternative might be used in case the original effect cannot be processed.
EXAMPLE 1 The alternative effect is chosen because the original intended effect cannot be processed due to lack of devices supporting this effect.
NOTE 3 The priority might by used to process effects-Flag within a group of effects-according to the capabilities of the adaptation VR).
EXAMPLE 2 The adaptation VR processes the individual effects of a group of effects according to their priority in descending order due to its limited capabilities. That is, effects with low priority might get lost.
A classification scheme that may be used for this purpose is the LocationCS as Flag in Annex A.2.1. The terms from the LocationCS shall be concatenated with the ":" sign in order of the x-, y-, and z-axis to uniquely define a location within the three-dimensional space.
For referring to a group of locations, a wild card mechanism may be employed using the "*" sign.
EXAMPLE 4 urn: mpeg: mpeg-v: 01-SI-LocationCS-NS: center: middle: front defines the location as follows: center on the x-axis, middle on the y-axis, and front on the z-axis . That is, it describes all effects at the center, middle, front side of the user.
EXAMPLE 5 urn: mpeg: mpeg-v: 01-SI-LocationCS-NS: left: *: midway defines the location as follows: left on the x-axis, any location on the y-axis, and midway on the z- axis. That is, it describes all effects at the left, midway side of the user.
EXAMPLE 6 urn: mpeg: mpeg-v: 01-SI-LocationCS-NS: *: *: back defines the location as follows: any location on the x-axis, any location on the y-axis, and back on the z -axis. That is, it describes all effects at the back of the user.
In the binary description, the following mapping table is used location.
In the SEM basic characteristic semantics shown in Table 3 above, location uses a location model for sensory effect metadata as shown in FIG. 5. 5 is a diagram illustrating a location model of sensory effect metadata in a multimedia service providing system according to an exemplary embodiment of the present invention.
That is, the positional model of the sensory effect metadata includes a back 502, a midway 504, a front 506, and a bottom on the spatial coordinates of xyz as shown in FIG. 5. (bottom) 508, middle 510, left 512,
As illustrated in FIG. 5, each position of the sensory effect metadata on the spatial coordinates of xyz may be represented in binary notation as shown in Table 4 below. That is, in the SEM basic characteristic semantics shown in Table 3, the position is encoded in binary notation. Here, Table 4 is a table showing the binary representation of the position on the spatial coordinates of xyz.
And, the semantics of the basic data type and elements of the sensory effect metadata can be expressed as shown in Table 5 below. Here, Table 5 is a table showing the SEM adaptability characteristics semantics (Semantics of the SEM Adaptability Attributes).
Strict: An adaptation by approximation may not be performed.
Under: An adaptation by approximation may be performed with a smaller effect value than the specified effect value.
Over: An adaptation by approximation may be performed with a greater effect value than the specified effect value.
Both: An adaptation by approximation may be performed between the upper and lower bound specified by adaptRange.
In the SEM adaptability characteristic semantics shown in Table 5, the adapt type can be represented as shown in Table 6 below and is encoded in binary notation. Here, Table 6 is a table showing the binary notation of the adaptive type.
In addition, the semantics of the basic data type and elements of the sensory effect metadata may be represented as in Table 7 below. Here, Table 7 is a table showing the SEM base type semantics (Semantics of the SEMBaseType).
Next, when describing the sensory effect information, that is, the root element of the sensory effect metadata, the syntax may be represented as shown in Table 8. Here, Table 8 is a table showing syntax of the root element.
<!-Definition of the SEM root element->
<!-############################################## ##->
<element name = "SEM">
<complexType>
<sequence>
<element name = "DescriptionMetadata" type = "sedl: DescriptionMetadataType"
minOccurs = "0" maxOccurs = "1"/>
<choice maxOccurs = "unbounded">
<element name = "Declarations" type = "sedl: DeclarationsType"/>
<element name = "GroupOfEffects" type = "sedl: GroupOfEffectsType"/>
<element name = "Effect" type = "sedl: EffectBaseType"/>
<element name = "ReferenceEffect" type = "sedl: ReferenceEffectType"/>
</ choice>
</ sequence>
<attribute name = "autoExtraction" type = "sedl: autoExtractionType"/>
<anyAttribute namespace = "## other" processContents = "lax"/>
</ complexType>
</ element>
<simpleType name = "autoExtractionType">
<restriction base = "string">
<enumeration value = "audio"/>
<enumeration value = "visual"/>
<enumeration value = "both"/>
</ restriction>
</ simpleType>
The binary encoding notation or binary notation of the root elements of the sensory effect metadata may be represented as in Table 9 below. Here, Table 9 is a table showing the binary representation of the root elements of the sensory effect metadata.
In addition, semantics of the root elements of the sensory effect metadata may be expressed as shown in Table 10 below. Here, Table 10 is a table showing the SEM root element semantics.
EXAMPLE-Creation information or Classification Scheme Alias.
In the binary description, the following mapping table is used.
audio: the automatic extraction of sensory effects from the audio part of the media resource, which is described by this sensory effect metadata, is desirable.
visual: the automatic extraction of sensory effects from the visual part of the media resource, which is described by this sensory effect metadata, is desirable.
both: the automatic extraction of sensory effects from both the audio and visual part of the media resource, which is described by this sensory effect metadata, is desirable.
In the binary description, the following mapping table is used.
In the SEM root element semantics shown in Table 10, the element ID (ElementID) may be represented in binary notation as shown in Table 11 below. Here, Table 11 is a table showing the binary representation of the element ID.
In addition, in the SEM root element semantics shown in Table 10 above, autoextractionID may be represented in binary notation as shown in Table 12 below. Here, Table 12 is a table showing the binary notation of the automatic extraction ID.
Here, in addition to the si attribute list (si attribute list), first, the XML representation syntax of the si attribute list (XML representation syntax) can be represented as shown in Table 13. Table 13 is a table showing XML notation syntax of the sensory effect metadata.
<!-Digital Item Adaptation ISO / IEC 21000-7 Second Edition->
<!-Schema for XML Streaming Instructions->
<schema
version = "ISO / IEC 21000-7 2nd"
id = "XSI-2nd.xsd"
xmlns = "http://www.w3.org/2001/XMLSchema"
xmlns: si = "urn: mpeg: mpeg21: 2003: 01-DIA-XSI-NS"
targetNamespace = "urn: mpeg: mpeg21: 2003: 01-DIA-XSI-NS"
elementFormDefault = "qualified">
<annotation>
<documentation>
Declaration of attributes used for XML streaming instructions
</ documentation>
</ annotation>
<!-The following attribute defines the process units->
<attribute name = "anchorElement" type = "boolean"/>
<!-The following attribute indicates that the PU shall be encoded as Random Access Point->
<attribute name = "encodeAsRAP" type = "boolean"/>
<attribute name = "puMode" type = "si: puModeType"/>
<simpleType name = "puModeType">
<restriction base = "string">
<enumeration value = "self"/>
<enumeration value = "ancestors"/>
<enumeration value = "descendants"/>
<enumeration value = "ancestorsDescendants"/>
<enumeration value = "preceding"/>
<enumeration value = "precedingSiblings"/>
<enumeration value = "sequential"/>
</ restriction>
</ simpleType>
<!-The following attributes define the time properties->
<attribute name = "timeScale" type = "unsignedInt"/>
<attribute name = "ptsDelta" type = "unsignedInt"/>
<attribute name = "absTimeScheme" type = "string"/>
<attribute name = "absTime" type = "string"/>
<attribute name = "pts" type = "nonNegativeInteger"/>
</ schema>
The binary encoding notation or binary notation of the syntax shown in Table 13 may be expressed as shown in Table 14 below. Here, Table 14 is a table showing a binary representation syntax.
The semantics of the si attribute list are shown in Table 15 below. Here, Table 15 is a table showing the semantics of the siAttributeList.
The anchorElement allows one to indicate whether an XML element is an anchor element, ie, the starting point for composing the process unit.
PuMode = descendants means that the process unit contains the anchor element and its descendant elements. Note that the anchor elements are pictured in white.
In the binary description, the following mapping table is used.
In the semantics of the si attribute list shown in Table 15, the put mode may be represented in binary notation as shown in Table 16 below. That is, in the semantics of the si attribute list shown in Table 15, the put mode is encoded in binary notation. Here, Table 16 shows a binary representation of the put mode.
Next, when describing the sensory effect information, that is, description metadata of the sensory effect metadata, the syntax may be expressed as shown in Table 17 below. Here, Table 17 is a table showing the technical metadata syntax.
<!-Definition of Description Metadata Type->
<!-############################################## ##->
<complexType name = "DescriptionMetadataType">
<complexContent>
<extension base = "mpeg7: DescriptionMetadataType">
<sequence>
<element name = "ClassificationSchemeAlias" minOccurs = "0"
maxOccurs = "unbounded">
<complexType>
<complexContent>
<extension base = "sedl: SEMBaseType">
<attribute name = "alias" type = "NMTOKEN" use = "required"/>
<attribute name = "href" type = "anyURI" use = "required"/>
</ extension>
</ complexContent>
</ complexType>
</ element>
</ sequence>
</ extension>
</ complexContent>
</ complexType>
The binary encoding notation or binary notation of the description metadata of the sensory effect metadata may be represented as shown in Table 18 below. Here, Table 18 is a table showing the binary notation of the description metadata of the sensory effect metadata.
In addition, the semantics of the description metadata of the sensory effect metadata may be expressed as shown in Table 19 below. Here, Table 19 is a table showing the description metadata semantics (Semantics of the DescriptionMetadata).
Next, when the sensory effect information, that is, the declarations of the sensory effect metadata, will be described, the syntax may be represented as shown in Table 20 below. Here, Table 20 is a table showing the declaration syntax.
<!-Declarations type->
<!-############################################## ##->
<complexType name = "DeclarationsType">
<complexContent>
<extension base = "sedl: SEMBaseType">
<choice maxOccurs = "unbounded">
<element ref = "sedl: GroupOfEffects"/>
<element ref = "sedl: Effect"/>
<element ref = "sedl: Parameter"/>
</ choice>
</ extension>
</ complexContent>
</ complexType>
The binary encoding notation or binary notation of the declaration of the sensory effect metadata may be represented as in Table 21 below. Here, Table 21 is a table showing the binary notation of the declaration of the sensory effect metadata.
In addition, the semantics of the declaration of the sensory effect metadata may be expressed as shown in Table 22 below. Here, Table 22 shows semantics of the DeclarationsType.
In the binary description, make referece to Table 3. Element ID
Next, when describing the sensory effect information, that is, a group of effects of the sensory effect metadata, the syntax may be expressed as shown in Table 23 below. Here, Table 23 is a table showing syntax of the effect group.
<!-Group of Effects type->
<!-############################################## ##->
<complexType name = "GroupOfEffectsType">
<complexContent>
<extension base = "sedl: SEMBaseType">
<choice minOccurs = "2" maxOccurs = "unbounded">
<element ref = "sedl: Effect"/>
<element ref = "sedl: ReferenceEffect"/>
</ choice>
<attributeGroup ref = "sedl: SEMBaseAttributes"/>
<anyAttribute namespace = "## other" processContents = "lax"/>
</ extension>
</ complexContent>
</ complexType>
The binary coding notation or binary notation of the effect group of the sensory effect metadata may be represented as in Table 24 below. Here, Table 24 shows a binary representation of the effect group of the sensory effect metadata.
In addition, the semantics of the effect group of the sensory effect metadata may be expressed as shown in Table 25 below. Here, Table 25 is a table showing the Semantics of the Group Of Effects Type.
In the binary description, make referece to Table 3. Element ID
NOTE ElementID restricted 3, 4
Next, the sensory effect information, that is, the effect (effect) of the sensory effect metadata will be described. Syntax may be expressed as shown in Table 26 below. Here, Table 26 is a table showing the effect syntax.
<!-Effect base type->
<!-############################################## ##->
<complexType name = "EffectBaseType" abstract = "true">
<complexContent>
<extension base = "sedl: SEMBaseType">
<sequence minOccurs = "0">
<element name = "SupplementalInformation" type = "sedl: SupplementalInformationType" minOccurs = "0"/>
</ sequence>
<attribute name = "autoExtraction" type = "sedl: autoExtractionType"/>
<attributeGroup ref = "sedl: SEMBaseAttributes"/>
<anyAttribute namespace = "## other" processContents = "lax"/>
</ extension>
</ complexContent>
</ complexType>
<complexType name = "SupplementalInformationType">
<sequence>
<element name = "ReferenceRegion" type = "mpeg7: SpatioTemporalLocatorType"/>
<element name = "Operator" type = "sedl: OperatorType" minOccurs = "0"/>
</ sequence>
</ complexType>
<simpleType name = "OperatorType">
<restriction base = "NMTOKEN">
<enumeration value = "Average"/>
<enumeration value = "Dominant"/>
</ restriction>
</ simpleType>
In addition, the binary coding notation or binary notation of the effect of the sensory effect metadata may be represented as shown in Table 27 below. Here, Table 27 is a table showing the binary notation of the effect of the sensory effect metadata.
In the binary notation of the effects shown in Table 27, the effect type ID may be represented as shown in Table 28 below. Here, Table 28 is a table showing the effect type ID in the binary notation of the effect.
MotionType
Forcetype
In addition, the semantics of the effect of the sensory effect metadata may be expressed as shown in Table 29 below. Here, Table 29 is a table showing the Semantics of the EffectBaseType (Semantics of the EffectBaseType).
EXAMPLE-si: pts describes the point in time when the associated information shall become available to the application for processing.
In addition, the semantics of the effect of the sensory effect metadata may be expressed as shown in Table 30 below. Here, Table 30 is a table showing the Semantics of the Supplemental Information Type in the binary notation of the effect of the sensory effect metadata shown in Table 27.
Average: extracts sensory effects from the reference region by calculating average value.
Dominant: extracts sensory effects from the reference region by calculating dominant value.
In Table 30, an operator may be represented by a binary notation as shown in Table 31 below. That is, in the supplemental information type semantics shown in Table 30, the operator is encoded in binary notation. Here, Table 31 is a table which shows the binary notation of an operator.
Next, referring to the sensory effect information, that is, the reference effect of the sensory effect metadata, the syntax may be expressed as shown in Table 32 below. Here, Table 32 is a table showing the reference effect syntax.
<!-Reference Effect type->
<!-############################################## ##->
<complexType name = "ReferenceEffectType">
<complexContent>
<extension base = "sedl: SEMBaseType">
<attribute name = "uri" type = "anyURI" use = "required"/>
<attributeGroup ref = "sedl: SEMBaseAttributes"/>
<anyAttribute namespace = "## other" processContents = "lax"/>
</ extension>
</ complexContent>
</ complexType>
The binary encoding notation or binary notation of the reference effect of the sensory effect metadata may be represented as shown in Table 33 below. Here, Table 33 is a table showing the binary notation of the reference effect of the sensory effect metadata.
In addition, the semantics of the reference effect of the sensory effect metadata may be expressed as shown in Table 34 below. Here, Table 34 is a table showing the Semantics of the ReferenceEffectType.
Attributes included here override the attribute values possibly defined within the sensory effect, group of effects or parameter referenced by the uri.
EXAMPLE-si: pts describes the point in time when the associated information shall become available to the application for processing.
Next, the sensory effect information, that is, the parameters (parameters) of the sensory effect metadata will be described. The syntax may be represented as in Table 35 below. Here, Table 35 is a table showing parameter syntax.
<!-Parameter Base type->
<!-############################################## ##->
<complexType name = "ParameterBaseType" abstract = "true">
<complexContent>
<extension base = "sedl: SEMBaseType"/>
</ complexContent>
</ complexType>
The binary coding notation or binary notation of the parameter of the sensory effect metadata may be represented as in Table 36 below. Here, Table 36 is a table which shows the binary notation of the parameter of the sensory effect metadata.
In addition, the semantics of the parameter of the sensory effect metadata may be expressed as shown in Table 37 below. Here, Table 37 is a table showing the parameter base type semantics.
Next, the sensory effect information, that is, the color correction parameter type of the sensory effect metadata will be described. First, the XML representation syntax of the color correction parameter type may be represented as shown in Table 38 below. Table 38 shows an XML notation syntax of the color correction parameter type.
<!-Definition of Color Correction Parameter type->
<!-############################################## ##->
<complexType name = "ColorCorrectionParameterType">
<complexContent>
<extension base = "sedl: ParameterBaseType">
<sequence>
<element name = "ToneReproductionCurves"
type = "sedl: ToneReproductionCurvesType" minOccurs = "0"/>
<element name = "ConversionLUT" type = "sedl: ConversionLUTType"/>
<element name = "ColorTemperature" type = "sedl: IlluminantType" minOccurs = "0"/>
<element name = "InputDeviceColorGamut"
type = "sedl: InputDeviceColorGamutType" minOccurs = "0"/>
<element name = "IlluminanceOfSurround" type = "mpeg7: unsigned12"
minOccurs = "0"/>
</ sequence>
</ extension>
</ complexContent>
</ complexType>
<complexType name = "ToneReproductionCurvesType">
<sequence maxOccurs = "256">
<element name = "DAC_Value" type = "mpeg7: unsigned8"/>
<element name = "RGB_Value" type = "mpeg7: doubleVector"/>
</ sequence>
</ complexType>
<complexType name = "ConversionLUTType">
<sequence>
<element name = "RGB2XYZ_LUT" type = "mpeg7: DoubleMatrixType"/>
<element name = "RGBScalar_Max" type = "mpeg7: doubleVector"/>
<element name = "Offset_Value" type = "mpeg7: doubleVector"/>
<element name = "Gain_Offset_Gamma" type = "mpeg7: DoubleMatrixType"/>
<element name = "InverseLUT" type = "mpeg7: DoubleMatrixType"/>
</ sequence>
</ complexType>
<complexType name = "IlluminantType">
<choice>
<sequence>
<element name = "XY_Value" type = "dia: ChromaticityType"/>
<element name = "Y_Value" type = "mpeg7: unsigned7"/>
</ sequence>
<element name = "Correlated_CT" type = "mpeg7: unsigned8"/>
</ choice>
</ complexType>
<complexType name = "InputDeviceColorGamutType">
<sequence>
<element name = "IDCG_Type" type = "string"/>
<element name = "IDCG_Value" type = "mpeg7: DoubleMatrixType"/>
</ sequence>
</ complexType>
The binary encoding notation or binary notation of the syntax shown in Table 38 may be expressed as shown in Table 39 below. Here, Table 39 is a table showing binary representation syntax.
The semantics of the color correction parameter type are shown in Table 40 below. Here, Table 40 is a table showing the color correction parameter type semantics (Semantics of the ColorCorrectionParameterType).
The semantics of the tone reproduction curves in the color correction parameter type semantics shown in Table 40 are as shown in Table 41 below. Here, Table 41 is a table showing the tone reproduction curve type semantics (Semantics of the ToneReproductionCurvesType).
In addition, the semantics of the conversion LUT (conversion LUT) in the color correction parameter type semantics shown in Table 40 are as shown in Table 42 below. Here, Table 42 shows a table of conversion LUT type semantics.
The size of the conversion matrix is 3x3 such as . The way of describing the values in the binary representation is in the order of [ , , ; , , ; , , ].
The semantics of the color correction parameter type are as shown in Table 43 below. Here, Table 43 is a table showing the semantics of the semantics (Semantics of the IlluminantType).
In the binary description, the following mapping table is used.
In the illuminant type semantics shown in Table 43, the illuminant of the element type may be represented in binary notation as shown in Table 44 below. That is, in the illuminant type semantics shown in Table 43, the element type is encoded in binary notation. Here, Table 44 shows a binary representation of element types.
The semantics of the input device color gamut in the color correction parameter type semantics shown in Table 40 are as shown in Table 45 below. Here, Table 45 is a table showing the input device color gamut type semantics (Semantics of the InputDeviceColorGamutType).
Hereinafter, the sensory effect term (Sensory effect vocabulary), that is, the binary representation of the sensory effect information and the sensory effect information through an example of various sensory effects will be described in more detail with respect to the coding using a binary notation scheme. . Here, the various sensory effects of the multimedia content, as described above, the lighting effect, color tone effect, glare effect, temperature effect, wind effect, vibration effect, spray effect as a water spray effect, aroma effect, smoke effect, color correction effect , Motion and sensory effects (e.g., rigid body motion effects, passive kinesthetic motion effects, passive kinesthetic force effects, active kinetic effects) (active kinesthetic effect, tactile effect), and the like.
First, the lighting effect will be described in detail. The syntax of the lighting effect can be expressed as shown in Table 46 below. Here, Table 46 is a table | surface which showed the syntax of the said lighting effect.
<!-SEV Light type->
<!-############################################## ##->
<complexType name = "LightType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<attribute name = "color" type = "sev: colorType" use = "optional"/>
<attribute name = "intensity-value" type = "sedl: intensityValueType"
use = "optional"/>
<attribute name = "intensity-range" type = "sedl: intensityRangeType"
use = "optional"/>
</ extension>
</ complexContent>
</ complexType>
<simpleType name = "colorType">
<union memberTypes = "mpeg7: termReferenceType sev: colorRGBType"/>
</ simpleType>
<simpleType name = "colorRGBType">
<restriction base = "NMTOKEN">
<whiteSpace value = "collapse"/>
<pattern value = "# [0-9A-Fa-f] {6}"/>
</ restriction>
</ simpleType>
In addition, the binary coding notation or binary notation of the lighting effect can be represented as shown in Table 47. Here, Table 47 is a table showing the binary notation of the lighting effect.
In addition, the semantics of the lighting effect can be expressed as shown in Table 48 below. Here, Table 48 is a table showing the illumination type semantics (Semantics of the LightType).
(Describes the color of the light effect as a reference to a classification scheme term or as RGB value.A CS that may be used for this purpose is the ColorCS defined in Annex A.2.1).
In the illumination type semantics shown in Table 48, the color may be represented in binary notation as shown in Table 49 below. That is, in the illumination type semantics shown in Table 48, the colors are encoded in binary notation. Here, Table 49 is a table showing binary notation of colors, namely named color types.
And, the semantics of the lighting effect can be expressed as shown in Table 50 below. Here, Table 50 is a table showing color RGB type semantics.
Next, the flash effect will be described in detail. The syntax of the flash effect can be expressed as shown in Table 51 below. Here, Table 51 is a table | surface which showed the syntax of the said flash effect.
<!-SEV Flash type->
<!-############################################## ##->
<complexType name = "FlashType">
<complexContent>
<extension base = "sev: LightType">
<attribute name = "frequency" type = "positiveInteger" use = "optional"/>
</ extension>
</ complexContent>
</ complexType>
In addition, the binary encoding notation or binary notation of the flash effect may be represented as shown in Table 52 below. Here, Table 52 is a table which shows the binary notation of the said flash effect.
In addition, the semantics of the scintillation effect can be shown as Table 53 below. Here, Table 53 is a table showing the flash type semantics (Semantics of the FlashType).
EXAMPLE-The value 10 means it will flicker 10 times for each second.
Next, with reference to the temperature effect in detail, the syntax of the temperature effect can be expressed as shown in Table 54 below. Here, Table 54 is a table | surface which showed the syntax of the said temperature effect.
<!-SEV Temperature type->
<!-############################################## ##->
<complexType name = "TemperatureType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<attribute name = "intensity-value" type = "sedl: intensityValueType"
use = "optional"/>
<attribute name = "intensity-range" type = "sedl: intensityRangeType"
use = "optional"/>
</ extension>
</ complexContent>
</ complexType>
In addition, the binary encoding notation or binary notation of the temperature effect may be represented as shown in Table 55 below. Here, Table 55 is a table which shows the binary notation of the said temperature effect.
In addition, the semantics of the temperature effect can be expressed as shown in Table 56 below. Here, Table 56 is a table showing the temperature type semantics (Semantics of the TemperatureType).
intensity-range [0]: minmum intensity
intensity-range [1]: maximum intensity
EXAMPLE-[0.0, 100.0] on the Celsius scale or [32.0, 212.0] on the Fahrenheit scale.
Next, the wind effect will be described in detail. The syntax of the wind effect can be expressed as shown in Table 57 below. Here, Table 57 is a table which shows the syntax of the said wind effect.
<!-SEV Wind type->
<!-############################################## ##->
<complexType name = "WindType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<attribute name = "intensity-value" type = "sedl: intensityValueType"
use = "optional"/>
<attribute name = "intensity-range" type = "sedl: intensityRangeType"
use = "optional"/>
</ extension>
</ complexContent>
</ complexType>
The binary encoding notation or binary notation of the wind effect may be represented as shown in Table 58 below. Here, Table 58 is a table showing the binary notation of the wind effect.
In addition, the semantics of the wind effect can be expressed as shown in Table 59 below. Here, Table 59 is a table showing the Semantics of the WindType.
intensity-range [0]: minmum intensity
intensity-range [1]: maximum intensity
EXAMPLE-[0.0, 12.0] on the Beaufort scale.
Next, when the vibration effect is described in detail, the syntax of the vibration effect can be expressed as shown in Table 60 below. Here, Table 60 is a table which shows the syntax of the said vibration effect.
<!-SEV Vibration type->
<!-############################################## ##->
<complexType name = "VibrationType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<attribute name = "intensity-value" type = "sedl: intensityValueType"
use = "optional"/>
<attribute name = "intensity-range" type = "sedl: intensityRangeType"
use = "optional"/>
</ extension>
</ complexContent>
</ complexType>
The binary coding notation or binary notation of the vibration effect may be represented as in Table 61 below. Here, Table 61 is a table showing the binary notation of the vibration effect.
In addition, the semantics of the vibration effect can be expressed as shown in Table 62 below. Here, Table 62 is a table showing vibration type semantics (Semantics of the VibrationType).
intensity-range [0]: minmum intensity
intensity-range [1]: maximum intensity
EXAMPLE-[0.0, 10.0] on the Richter magnitude scale
Next, the spraying effect (spraying effect) will be described in detail, the syntax of the spraying effect can be expressed as shown in Table 64 below. Here, Table 64 is a table | surface which showed the syntax of the said spray effect.
<!-Definition of Spraying type->
<!-############################################## ##->
<complexType name = "SprayingType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<attribute name = "intensity-value" type = "sedl: intensityValueType"
use = "optional"/>
<attribute name = "intensity-range" type = "sedl: intensityRangeType"
use = "optional"/>
<attribute name = "sprayingType" type = "mpeg7: termReferenceType"/>
</ extension>
</ complexContent>
</ complexType>
The binary encoding notation or binary notation of the spray effect may be represented as in Table 65 below. Here, Table 65 is a table which shows the binary notation of the said spray effect.
In addition, the semantics of the spray effect can be shown as Table 66 below. Here, Table 66 is a table showing the spray type semantics (Semantics of the SprayingType).
intensity-range [0]: minmum intensity
intensity-range [1]: maximum intensity
EXAMPLE-[0.0, 10.0] ml / h.
In the semantics of the spray effect shown in Table 66, the spraying type can be represented in binary notation as shown in Table 67 below. That is, in the spray type semantics shown in Table 66, the spray type is encoded in binary notation. Here, Table 67 is a table which shows the binary notation of a spray type.
~ 11111111
Next, specifically describing the fragrance effect, the syntax of the fragrance effect may be represented as shown in Table 68 below. Here, Table 68 is a table | surface which showed the syntax of the said fragrance effect.
<!-Definition of Scent type->
<!-############################################## ##->
<complexType name = "ScentType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<attribute name = "scent" type = "mpeg7: termReferenceType"
use = "optional"/>
<attribute name = "intensity-value" type = "sedl: intensityValueType"
use = "optional"/>
<attribute name = "intensity-range" type = "sedl: intensityRangeType"
use = "optional"/>
</ extension>
</ complexContent>
</ complexType>
The binary encoding notation or binary notation of the fragrance effect may be represented as shown in Table 69 below. Here, Table 69 is a table which shows the binary notation of the said fragrance effect.
In addition, the semantics of the fragrance effect can be shown as Table 70 below. Here, Table 70 is a table showing the fragrance type semantics (Semantics of the ScentType).
intensity-range [0]: minmum intensity
intensity-range [1]: maximum intensity
EXAMPLE-[0.0, 10.0] ml / h.
In the semantics of the fragrance effect shown in Table 70, the scent can be represented in binary notation as shown in Table 71 below. That is, in the fragrance type semantics shown in Table 70, the fragrance is encoded in binary notation. Here, Table 71 is a table which shows the binary notation of fragrance.
~ 1111111111111111
Next, with reference to the smoke effect in detail, the syntax of the smoke effect can be expressed as shown in Table 72 below. Here, Table 72 is a table | surface which showed the syntax of the said smoke effect.
<!-Definition of Fog type->
<!-############################################## ##->
<complexType name = "FogType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<attribute name = "intensity-value" type = "sedl: intensityValueType"
use = "optional"/>
<attribute name = "intensity-range" type = "sedl: intensityRangeType"
use = "optional"/>
</ extension>
</ complexContent>
</ complexType>
In addition, the binary encoding notation or binary notation of the deferred effect may be represented as shown in Table 73 below. Here, Table 73 is a table which shows the binary notation of the said smoke effect.
In addition, the semantics of the smoke effect can be expressed as shown in Table 74 below. Here, Table 74 is a table showing the smoke type semantics (Semantics of the FogType).
intensity-range [0]: minmum intensity
intensity-range [1]: maximum intensity
EXAMPLE-[0.0, 10.0] ml / h.
Next, the color correction effect will be described in detail. The syntax of the color correction effect can be expressed as shown in Table 75 below. Here, Table 75 is a table showing the syntax of the color correction effect.
<!-Definition of Color Correction type->
<!-############################################## ##->
<complexType name = "ColorCorrectionType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<choice minOccurs = "0">
<element name = "SpatioTemporalLocator" type = "mpeg7: SpatioTemporalLocatorType"/>
<element name = "SpatioTemporalMask" type = "mpeg7: SpatioTemporalMaskType"/>
</ choice>
<attribute name = "intensity-value" type = "sedl: intensityValueType"
use = "optional"/>
<attribute name = "intensity-range" type = "sedl: intensityRangeType"
use = "optional" fixed = "0 1"/>
</ extension>
</ complexContent>
</ complexType>
The binary encoding notation or binary notation of the color correction effect may be represented as shown in Table 76 below. Here, Table 76 is a table showing the binary notation of the color correction effect.
In addition, the semantics of the color correction effect may be expressed as shown in Table 77 below. Here, Table 77 is a table showing color correction type semantics (Semantics of the ColorCorrectionType).
intensity-range [0]: minmum intensity
intensity-range [1]: maximum intensity
Next, the rigid body motion effect as a motion effect will be described in detail. The syntax of the ridge body motion effect can be expressed as shown in Table 78 below. Here, Table 78 is a table which shows the syntax of the said ridge body motion effect.
<!-Definition of Rigid Body Motion type->
<!-############################################## ##->
<complexType name = "RigidBodyMotionType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<sequence>
<element name = "MoveToward" type = "sev: MoveTowardType"
minOccurs = "0"/>
<element name = "TrajectorySamples" type = "mpeg7: FloatMatrixType"
minOccurs = "0" maxOccurs = "unbounded"/>
<element name = "Incline" type = "sev: InclineType" minOccurs = "0"/>
<element name = "Shake" type = "sev: ShakeType" minOccurs = "0"/>
<element name = "Wave" type = "sev: WaveType" minOccurs = "0"/>
<element name = "Spin" type = "sev: SpinType" minOccurs = "0"/>
<element name = "Turn" type = "sev: TurnType" minOccurs = "0"/>
<element name = "Collide" type = "sev: CollideType" minOccurs = "0"/>
</ sequence>
</ extension>
</ complexContent>
</ complexType>
<!-############################################## ##->
<!-Definition of Move Toward type->
<!-############################################## ##->
<complexType name = "MoveTowardType">
<choice minOccurs = "0">
<element name = "Speed" type = "float"/>
<element name = "Acceleration" type = "float"/>
</ choice>
<attribute name = "directionV" type = "MoveTowardAngleType" use = "optional" default = "0"/>
<attribute name = "directionH" type = "MoveTowardAngleType" use = "optional" default = "0"/>
<attribute name = "distance" type = "float" use = "optional"/>
</ complexType>
<!-############################################## ##->
<!-Definition of Incline type->
<!-############################################## ##->
<complexType name = "InclineType">
<sequence>
<choice minOccurs = "0">
<element name = "PitchSpeed" type = "float"/>
<element name = "PitchAcceleration" type = "float"/>
</ choice>
<choice minOccurs = "0">
<element name = "rollSpeed" type = "float"/>
<element name = "rollAcceleration" type = "float"/>
</ choice>
<choice minOccurs = "0">
<element name = "yawSpeed" type = "float"/>
<element name = "yawAcceleration" type = "float"/>
</ choice>
</ sequence>
<attribute name = "pitch" type = "sev: InclineAngleType" use = "optional" default = "0"/>
<attribute name = "roll" type = "sev: InclineAngleType" use = "optional" default = "0"/>
<attribute name = "yaw" type = "sev: InclineAngleType" use = "optional" default = "0"/>
</ complexType>
<!-############################################## ##->
<!-Definition of Shake type->
<!-############################################## ##->
<complexType name = "ShakeType">
<attribute name = "direction" type = "mpeg7: termReferenceType"
use = "optional"/>
<attribute name = "count" type = "float" use = "optional"/>
<attribute name = "distance" type = "float" use = "optional"/>
</ complexType>
<!-############################################## ##->
<!-Definition of Wave type->
<!-############################################## ##->
<complexType name = "WaveType">
<attribute name = "direction" type = "mpeg7: termReferenceType"
use = "optional"/>
<attribute name = "startDirection" type = "mpeg7: termReferenceType"
use = "optional"/>
<attribute name = "count" type = "float" use = "optional"/>
<attribute name = "distance" type = "float" use = "optional"/>
</ complexType>
<!-############################################## ##->
<!-Definition of Spin type->
<!-############################################## ##->
<complexType name = "SpinType">
<attribute name = "direction" type = "mpeg7: termReferenceType"
use = "optional"/>
<attribute name = "count" type = "float" use = "optional"/>
</ complexType>
<!-############################################## ##->
<!-Definition of Turn type->
<!-############################################## ##->
<complexType name = "TurnType">
<attribute name = "direction" type = "sev: TurnAngleType" use = "optional"/>
<attribute name = "speed" type = "float" use = "optional"/>
</ complexType>
<!-############################################## ##->
<!-Definition of Collide type->
<!-############################################## ##->
<complexType name = "CollideType">
<attribute name = "directionH" type = "sev: MoveTowardAngleType"
use = "optional" default = "0"/>
<attribute name = "directionV" type = "sev: MoveTowardAngleType"
use = "optional" default = "0"/>
<attribute name = "speed" type = "float" use = "optional"/>
</ complexType>
<!-############################################## ##->
<!-Definition of Rigid Body Motion base type->
<!-############################################## ##->
<simpleType name = "TurnAngleType">
<restriction base = "integer">
<minInclusive value = "-180"/>
<maxInclusive value = "180"/>
</ restriction>
</ simpleType>
<simpleType name = "InclineAngleType">
<restriction base = "integer">
<minInclusive value = "-359"/>
<maxInclusive value = "359"/>
</ restriction>
</ simpleType>
<simpleType name = "MoveTowardAngleType">
<restriction base = "integer">
<minInclusive value = "0"/>
<maxInclusive value = "359"/>
</ restriction>
</ simpleType>
The binary coding notation or binary notation of the ridge body motion effect may be represented as in Table 79 below. Here, Table 79 is a table showing the binary notation of the ridge body motion effect.
SizeOfIntensityColumn); k ++) {
In addition, the semantics of the ridge body motion effect can be shown in Table 80 below. Here, Table 80 is a table showing the ridge body motion type semantics (Semantics of the RigidBodyMotionType).
In the ridge body motion type semantics shown in Table 80, Move Torward represents the
In addition, in the ridge body motion type semantics shown in Table 80, the transactional sample (TrajectorySamples) represents a set of negative coordinate
In the ridge body motion type semantics shown in Table 80, the incline is pitched 810,
In addition, in the ridge body motion type semantics shown in Table 80, the shake shows the
Further, in the ridge body motion type semantics shown in Table 80, the wave represents a
In the ridge body motion type semantics shown in Table 80, spin represents a
In addition, in the ridge body motion type semantics shown in Table 80, a turn rotates the turn-
In addition, in the ridge body motion type semantics shown in Table 80, the collide, as shown in FIG. 13, the impact (1350) due to the collision according to the
In addition, the semantics of the ridge body motion effect can be expressed as shown in Table 81 below. Here, Table 81 is a table | surface which shows Move forward type semantics.
In the move forward type semantics shown in Table 81, the direction H (direction H), as shown in Figure 14, represents the magnitude of the horizontal direction movement through the angle unit, wherein
In the move forward type semantics shown in Table 81, the direction V indicates the magnitude of the vertical direction movement through an angle unit, as shown in FIG. 15, wherein
In addition, the semantics of the ridge body motion effect can be shown in Table 82 below. Here, Table 82 is a table | surface which shows Incline type semantics.
In the incline type semantics shown in Table 82, the pitch, roll0, and yaw is the
In addition, the semantics of the ridge body motion effect can be expressed as shown in Table 83 below. Here, Table 83 is a table | surface which shows the Semantics of the ShakeType.
In the shake type semantics shown in Table 83, the direction is the direction of the
In the shake type semantics shown in Table 83, the distance indicates the moving
In addition, the semantics of the ridge body motion effect can be expressed as shown in Table 85 below. Here, Table 85 is a table showing wave type semantics.
In the wave type semantics shown in Table 85, the direction represents a continuous wave pattern such as wave ripple at a predetermined position (1900, 2000) as shown in Figs. 19 and 20, in particular the wave pattern. Front-rear 1910 and left-right (2010). 19 and 20 illustrate wave motion directions in a realistic effect of a multimedia service providing system according to an exemplary embodiment of the present invention. In addition, the direction of the wave pattern may be represented in binary notation as shown in Table 86 below. That is, in the wave type semantics shown in Table 85, the direction is encoded in binary notation. Here, Table 86 is a table showing the binary notation of directions.
In the wave type semantics shown in Table 85, the start direction indicates the start direction of the
In addition, in the wave type semantics shown in Table 85, the distance represents the
In addition, the semantics of the ridge body motion effect can be expressed as shown in Table 88 below. Here, Table 88 is a table | surface which shows turn type semantics.
In the turn type semantics shown in Table 88, the direction indicates the turn direction, as shown in FIG. 24, in particular the direction -90 (2410) and the direction 90 (2420) of the turn pattern. 24 is a diagram illustrating a turn pattern direction in the sensory effect of the multimedia service providing system according to an embodiment of the present invention.
In addition, the semantics of the ridge body motion effect can be expressed as shown in Table 89 below. Here, Table 89 is a table showing spin type semantics.
In the spin type semantics shown in Table 89, the direction may be represented in binary notation as shown in Table 90 below. That is, in the spin type semantics shown in Table 89, the direction is encoded in binary notation. Here, Table 90 is a table showing the binary notation of the direction.
In addition, the semantics of the ridge body motion effect can be shown in Table 91 below. Here, Table 91 is a table | surface which shows the Collide type semantics.
In the collide type semantics shown in Table 91, the direction H indicates the magnitude of the horizontal direction movement through the angle unit, as shown in FIG.
In the colloid type semantics shown in Table 91, the direction V indicates the magnitude of the vertical direction movement through an angle unit, as shown in FIG.
Next, the passive kinesthetic motion effect as a motion effect will be described in detail. The syntax of the passive kinetic motion effect can be expressed as shown in Table 92 below. Here, Table 92 is a table showing the syntax of the passive kinematic motion effect.
<!-SEV Passive Kinesthetic Motion type->
<!-############################################## ##->
<complexType name = "PassiveKinestheticMotionType">
<complexContent>
<extension base = "sev: RigidBodyMotionType">
<attribute name = "updaterate" type = "positiveInteger" use = "required"/>
</ extension>
</ complexContent>
</ complexType>
The binary coding notation or binary notation of the passive kinematic motion effect may be represented as shown in Table 93 below. Here, Table 93 is a table showing the binary notation of the passive kinematic motion effect.
In addition, the semantics of the passive kinematic motion effect can be shown as Table 94 below. Here, Table 94 is a table showing the passive kinematic motion type semantics (Semantics of the Passive Kinesthetic Motion Type).
ex. The value 20 means the kinesthetic device will move to 20 different positions and orientations for each second.
Next, the passive kinesthetic force effect (passive kinesthetic force effect) will be described in detail, the syntax of the passive kinetic force effect can be expressed as shown in Table 95 below. Here, Table 95 is a table showing the syntax of the passive kinetic force effect.
<!-SEV Passive Kinesthetic Force type->
<!-############################################## ##->
<complexType name = "PassiveKinestheticForceType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<sequence>
<element name = "passivekinestheticforce"
type = "mpeg7: FloatMatrixType"/>
</ sequence>
<attribute name = "updaterate" type = "positiveInteger" use = "required"/>
</ extension>
</ complexContent>
</ complexType>
The binary coding notation or binary notation of the passive kinematic force effect may be represented as shown in Table 96 below. Here, Table 96 is a table showing the binary notation of the passive kinetic force effect.
SizeOfforceColumn); k ++) {
In addition, the semantics of the passive kinetic force effect can be expressed as shown in Table 97 below. Here, Table 97 is a table | surface which shows the passive kinematic force type semantics.
Next, the active kinesthetic effect (active kinesthetic effect) will be described in detail, the syntax of the active kinetic effect can be expressed as shown in Table 98 below. Here, Table 98 is a table showing the syntax of the active kinetic effect.
<!-SEV Active Kinesthetic type->
<!-############################################## ##->
<complexType name = "ActiveKinestheticType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<sequence>
<element name = "activekinesthetic"
type = "sev: ActiveKinestheticForceType"/>
</ sequence>
</ extension>
</ complexContent>
</ complexType>
<complexType name = "ActiveKinestheticForceType">
<attribute name = "Fx" type = "float"/>
<attribute name = "Fy" type = "float"/>
<attribute name = "Fz" type = "float"/>
<attribute name = "Tx" type = "float" use = "optional"/>
<attribute name = "Ty" type = "float" use = "optional"/>
<attribute name = "Tz" type = "float" use = "optional"/>
</ complexType>
The binary coding notation or binary notation of the active kinematics effect may be represented as in Table 99 below. Here, Table 99 is a table showing the binary notation of the active kinematics effect.
In addition, the semantics of the active kinesmatic effect can be expressed as shown in Table 100 below. Here, Table 100 is a table showing the Active Kinesetic Type semantics (Semantics of the Active Kinesthetic Type).
Next, the tactile effect will be described in detail. The syntax of the tactile effect can be expressed as shown in Table 101 below. Here, Table 101 is a table | surface which showed the syntax of the said tactile effect.
<!-SEV Tactile type->
<!-############################################## ##->
<complexType name = "TactileType">
<complexContent>
<extension base = "sedl: EffectBaseType">
<sequence>
<choice>
<element name = "ArrayIntensity" type = "mpeg7: FloatMatrixType"/>
<element name = "TactileVideo" type = "anyURI"/>
</ choice>
</ sequence>
<attribute name = "tactileEffect" type = "mpeg7: termReferenceType" use = "optional"/>
<attribute name = "updaterate" type = "positiveInteger" use = "optional"/>
</ extension>
</ complexContent>
</ complexType>
The binary coding notation or binary notation of the tactile effect may be represented as shown in Table 102 below. Here, Table 102 is a table showing the binary notation of the haptic effect.
SizeOfIntensityColumn); k ++) {
In addition, the semantics of the haptic effect may be represented as shown in Table 103 below. Here, Table 103 is a table which shows tactile type semantics.
In the binary description, the following mapping table is used.
In the tactile type semantics shown in Table 103, the tactile effect can be represented in binary notation as shown in Table 104 below. That is, in the tactile type semantics shown in Table 103, the haptic effect is encoded in binary notation. Here, Table 104 is a table which shows the binary notation of a tactile effect.
Next, an operation of the multimedia service providing system according to an exemplary embodiment of the present invention will be described in more detail with reference to FIG. 27.
27 is a diagram schematically illustrating a process of providing a multimedia service of a multimedia service providing system according to an embodiment of the present invention.
Referring to FIG. 27, in
In
In
In
In
Then, in
In
While the invention has been shown and described with reference to certain preferred embodiments thereof, it will be understood by those skilled in the art that various changes and modifications may be made without departing from the spirit and scope of the invention. Therefore, the scope of the present invention should not be limited to the described embodiments, but should be determined not only by the scope of the following claims, but also by the equivalents of the claims.
Claims (16)
A service provider for providing multimedia content of the multimedia service and sensory effect information indicating sensory effects of the multimedia content according to a service request of a multimedia service that users want to receive;
A user server configured to receive multimedia data including the multimedia content and the sensory effect information, and convert the sensory effect information into control information from the multimedia data and provide the command information; And
And user devices that provide the multimedia content and the sensory effects to the users in real time through a device command according to the control information.
The service provider, the multimedia service providing system, characterized in that for encoding the sensory effect information in a binary representation (binary representation).
The multimedia data, the multimedia content and the multimedia service providing system, characterized in that the sensory effect information encoded in the binary notation.
A generator for generating the multimedia content and the sensory effect information;
An encoder for encoding the sensory effect information by using a binary notation coding scheme; And
And a transmitter for transmitting the multimedia data.
And the encoder encodes the sensory effect information into a sensory effect stream in binary notation.
The sensory effects may include a light effect, a colored light effect, a flash light effect, a temperature effect, a wind effect, a vibration effect, Spraying effect, scent effect, smoke effect, color correction effect, rigid body motion effect, passive kinetic motion effect (passive kinesthetic) A multimedia service providing system comprising a motion effect, a passive kinesthetic force effect, an active kinesthetic effect, and a tactile effect.
And the service provider defines syntax, binary representation, and semantics of the sensory effects.
A generator for generating multimedia content of the multimedia service and generating sensory effect information indicative of sensory effects of the multimedia content according to a service request of a multimedia service to be provided by users;
An encoder for encoding the sensory effect information in binary notation using a binary representation coding scheme; And
And a transmitter for transmitting the multimedia data including the multimedia content and sensory effect information encoded in the binary notation.
And the encoder encodes the sensory effect information into a sensory effect stream in binary notation.
The sensory effects may include a light effect, a colored light effect, a flash light effect, a temperature effect, a wind effect, a vibration effect, Spraying effect, scent effect, smoke effect, color correction effect, rigid body motion effect, passive kinetic motion effect (passive kinesthetic) A multimedia service providing system comprising a motion effect, a passive kinesthetic force effect, an active kinesthetic effect, and a tactile effect.
The encoder defines a syntax, binary representation, and semantics of the sensory effects.
Generating multimedia content of the multimedia service and generating sensory effect information indicative of sensory effects of the multimedia content according to a service request of a multimedia service to be provided by users;
Encoding the sensory effect information in binary notation using a binary representation coding scheme;
Converting sensory effect information encoded in the binary notation into command information in binary notation; And
And providing the multimedia contents and the sensory effects to the users in real time through a device command according to the control information of the binary notation.
And transmitting multimedia data including the multimedia content and sensory effect information encoded in the binary notation.
The encoding in the binary notation may include encoding the sensory effect information into a sensory effect stream in binary notation.
The sensory effects may include a light effect, a colored light effect, a flash light effect, a temperature effect, a wind effect, a vibration effect, Spraying effect, scent effect, smoke effect, color correction effect, rigid body motion effect, passive kinetic motion effect (passive kinesthetic) A multimedia service providing method comprising a motion effect, a passive kinesthetic force effect, an active kinesthetic effect, and a tactile effect.
The encoding in binary notation may include defining syntax, binary representation, and semantics of the sensory effects.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/080,095 US20110282967A1 (en) | 2010-04-05 | 2011-04-05 | System and method for providing multimedia service in a communication system |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR20100031129 | 2010-04-05 | ||
KR1020100031129 | 2010-04-05 |
Publications (1)
Publication Number | Publication Date |
---|---|
KR20110112211A true KR20110112211A (en) | 2011-10-12 |
Family
ID=45028081
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
KR1020110030397A KR20110112211A (en) | 2010-04-05 | 2011-04-01 | System and method for providing multimedia service in a communication system |
Country Status (1)
Country | Link |
---|---|
KR (1) | KR20110112211A (en) |
-
2011
- 2011-04-01 KR KR1020110030397A patent/KR20110112211A/en not_active Application Discontinuation
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20110241908A1 (en) | System and method for processing sensory effect | |
US20130103703A1 (en) | System and method for processing sensory effects | |
CN107771395B (en) | Method and apparatus for generating and transmitting metadata for virtual reality | |
JP4334470B2 (en) | Ambient light control | |
US20080291218A1 (en) | System And Method For Generating Interactive Video Images | |
US10897646B2 (en) | Video stream transmission method and related device and system | |
JP5781435B2 (en) | Multimedia application system and method using metadata related to sensory playback device | |
KR101571283B1 (en) | Media content transmission method and apparatus, and reception method and apparatus for providing augmenting media content using graphic object | |
US20100274817A1 (en) | Method and apparatus for representing sensory effects using user's sensory effect preference metadata | |
US20100268745A1 (en) | Method and apparatus for representing sensory effects using sensory device capability metadata | |
RU2721410C2 (en) | Method and device for providing tactile feedback and interactivity based on user's tactile space (hapspace) | |
US10861221B2 (en) | Sensory effect adaptation method, and adaptation engine and sensory device to perform the same | |
KR20110030656A (en) | Data transmission device, method for transmitting data, audio-visual environment control device, audio-visual environment control system, and method for controlling audio-visual environment | |
JP2007248861A (en) | Image display device and receiving device | |
CN108833937A (en) | Method for processing video frequency and device | |
Kim | Authoring multisensorial content | |
Yoon et al. | MPEG-V: Bridging the virtual and real world | |
Saleme et al. | Coping with the challenges of delivering multiple sensorial media | |
Timmerer et al. | Assessing the quality of sensory experience for multimedia presentations | |
CN102282849A (en) | Data transmission device, data transmission mthod, audio-visual environment control devcice, audio-visual environment control method, and audio-visual environment control system | |
KR20110111251A (en) | Method and apparatus for providing metadata for sensory effect, computer readable record medium on which metadata for sensory effect is recorded, method and apparatus for representating sensory effect | |
US20110282967A1 (en) | System and method for providing multimedia service in a communication system | |
US20110276659A1 (en) | System and method for providing multimedia service in a communication system | |
US20080240669A1 (en) | Mpeg-based user interface device and method of controlling function using the same | |
KR20110112211A (en) | System and method for providing multimedia service in a communication system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WITN | Withdrawal due to no request for examination |