WO2015105327A1 - Procede d'affichage de profondeur de bits pour la lecture d'une video utilisant un service de diffusion en flux adaptatif sur http (dash) - Google Patents

Procede d'affichage de profondeur de bits pour la lecture d'une video utilisant un service de diffusion en flux adaptatif sur http (dash) Download PDF

Info

Publication number
WO2015105327A1
WO2015105327A1 PCT/KR2015/000137 KR2015000137W WO2015105327A1 WO 2015105327 A1 WO2015105327 A1 WO 2015105327A1 KR 2015000137 W KR2015000137 W KR 2015000137W WO 2015105327 A1 WO2015105327 A1 WO 2015105327A1
Authority
WO
WIPO (PCT)
Prior art keywords
media
representation
representations
attribute
metadata
Prior art date
Application number
PCT/KR2015/000137
Other languages
English (en)
Korean (ko)
Inventor
이진영
윤국진
정원식
이광순
허남호
Original Assignee
한국전자통신연구원
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from KR1020150000758A external-priority patent/KR20150083429A/ko
Application filed by 한국전자통신연구원 filed Critical 한국전자통신연구원
Priority to EP15735102.4A priority Critical patent/EP3094097A4/fr
Priority to CN201580004175.2A priority patent/CN106063279A/zh
Priority to US15/110,500 priority patent/US20160337679A1/en
Publication of WO2015105327A1 publication Critical patent/WO2015105327A1/fr

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs
    • H04N21/2343Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements
    • H04N21/23439Processing of video elementary streams, e.g. splicing of video streams or manipulating encoded video stream scene graphs involving reformatting operations of video signals for distribution or compliance with end-user requests or end-user device requirements for generating different versions
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/258Client or end-user data management, e.g. managing client capabilities, user preferences or demographics, processing of multiple end-users preferences to derive collaborative data
    • H04N21/25808Management of client data
    • H04N21/25825Management of client data involving client display capabilities, e.g. screen resolution of a mobile phone
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/25Management operations performed by the server for facilitating the content distribution or administrating data related to end-users or client devices, e.g. end-user or client device authentication, learning user preferences for recommending movies
    • H04N21/266Channel or content management, e.g. generation and management of keys and entitlement messages in a conditional access system, merging a VOD unicast channel into a multicast channel
    • H04N21/2662Controlling the complexity of the video stream, e.g. by scaling the resolution or bitrate of the video stream based on the client capabilities
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/60Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client 
    • H04N21/61Network physical structure; Signal processing
    • H04N21/6106Network physical structure; Signal processing specially adapted to the downstream path of the transmission network
    • H04N21/6125Network physical structure; Signal processing specially adapted to the downstream path of the transmission network involving transmission via Internet
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/84Generation or processing of descriptive data, e.g. content descriptors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/83Generation or processing of protective or descriptive data associated with content; Content structuring
    • H04N21/845Structuring of content, e.g. decomposing content into time segments
    • H04N21/8456Structuring of content, e.g. decomposing content into time segments by decomposing the content in the time domain, e.g. in time segments

Definitions

  • the technical field of the present invention relates to streaming content providing technology, and more particularly, to a bit depth representation of video being played.
  • Streaming is one of the methods for transmitting and playing multimedia content such as sound and video.
  • the client can play the content while receiving the content through streaming.
  • the adaptive streaming service means providing a streaming service through a communication scheme consisting of a client request and a server response to the request.
  • the client may request a media sequence suitable for its environment (for example, its transport channel) through an adaptive streaming service, and the server may select a media sequence that matches the client's request among various quality media sequences that it has. Can provide.
  • the adaptive streaming service can be provided based on various protocols.
  • the HTTP adaptive streaming service refers to an adaptive streaming service provided based on the HTTP protocol.
  • Clients of the HTTP adaptive streaming service can be provided with content from a server using the HTTP protocol, and can send a request related to the streaming service to the server.
  • One embodiment of the present invention can provide an apparatus and method for playing media content using representations including accessibility information.
  • One embodiment of the present invention can provide an apparatus and method for playing media content using information describing a relationship between representations of the media content.
  • One embodiment of the present invention may provide an apparatus and method for playing media content using information of different representations for terminals having different display bit depths.
  • receiving metadata of media content comprising one or more periods, accessing segments of the media based on the information provided by the metadata and the media content included in the segments Decoding and rendering data of each cycle, wherein each cycle includes one or more representations of the media, each representation containing information for assisting persons with disabilities who have difficulty in perceiving the media content.
  • Including the media providing method is provided.
  • the media component of the representation is increased to address the problem of the handicapped, the increase being one or more of an increase in contrast and an increase in color correction.
  • the information indicates the intended use of the representation and the intended use is one or more of sign language, subtitles, captions, and description.
  • the information may include an attribute that provides a reference to an external MPEG-21 Digital Item Adaptation (DIA).
  • DIA Digital Item Adaptation
  • a first representation of the one or more representations may include an attribute indicating a list of representations of some of the one or more representations, wherein the first representation is the portion of the representation Can be used with one of the presentations.
  • the first representation may be a representation for audio description.
  • the some representations may each be representations of a video component.
  • the first representation of the one or more representations may include an attribute indicating a list of representations of some of the one or more representations.
  • One representation of the some representations may be used in place of the first representation.
  • receiving metadata of media content including one or more periods, accessing segments of the media based on information provided by the metadata, and including within the segments Decoding and rendering data of media content, each period comprising one or more representations of the media, each representation selecting a different representation for terminals with different display bit depths.
  • a media presentation method can be provided that includes a bit depth attribute used for the purpose.
  • the bit depth attribute may indicate the number of bits for representing luma / chroma samples of visual content.
  • receiving metadata of media content including one or more periods, receiving segments of the media based on information provided by the metadata, and receiving the data of the media included in the segments.
  • a media engine that receives data of the media from the access engine and outputs the media, wherein each cycle includes one or more representations of the media, each representation comprising the media;
  • a terminal may be provided that includes information to assist disabled people who have difficulty in perceiving content.
  • receiving metadata of media content including one or more periods, receiving segments of the media based on information provided by the metadata, and receiving the data of the media included in the segments.
  • a media engine that receives data of the media from the access engine and outputs the media, wherein each cycle includes one or more representations of the media, and wherein the metadata includes the one or more representations of the media;
  • a terminal may be provided that includes an attribute that describes a relationship between representations.
  • receiving metadata of media content including one or more periods, receiving segments of the media based on information provided by the metadata, and receiving the data of the media included in the segments.
  • a media engine that receives data of the media from the access engine and outputs the media, wherein each cycle includes one or more representations of the media, each representation having a different display;
  • a terminal may be provided that includes a bit depth attribute used to select a different representation for terminals with bit depths.
  • the method and apparatus according to an embodiment of the present invention may play media content using representations that include accessibility information.
  • the method and apparatus according to an embodiment of the present invention may play the media content using information describing the relationship between the representations of the media content.
  • the method and apparatus according to an embodiment of the present invention can play media content using information of different representations for terminals with different display bit depths.
  • 1 shows a high-level data model of DASH.
  • FIG. 2 is a signal flowchart of a content processing method according to an embodiment of the present invention.
  • FIG 3 illustrates categories of signaling information according to an embodiment of the present invention.
  • FIG 4 illustrates layers of content divisions and levels of signaling information according to an embodiment of the present invention.
  • FIG. 5 is a structural diagram of a terminal according to an embodiment of the present invention.
  • the index of a particular element is 1 to N or 0 to N
  • the above element may be one or more.
  • N may represent an integer of 0 or more.
  • Dynamic Adaptive Streaming over HTTP allows 1) the delivery of media content from a standard HTTP server to an HTTP client, and 2) by standard HTTP caches. You can specify formats that enable caching of content.
  • the format may be an XML format or a binary format.
  • the media content may be one media content period or a contiguous sequence of media content periods.
  • media content, media and content may be used interchangeably with each other.
  • Media content may be a collection of media content components having a common timeline, such as video, audio, and subtitles, for example.
  • media components may have relationships as a program or movie (eg, individually, jointly, or mutually exclusive).
  • the media content may be content on demand or live content.
  • the content may be divided into one or more intervals. That is, content can include one or more intervals.
  • an interval and a period may be used in the same meaning.
  • the term may be a term used in Generation Partnership Project (3GPP) adaptive HTTP streaming.
  • the period may be an interval of media presentation. Consecutive sequences of all periods can constitute a media presentation.
  • One or more intervals may be a basic unit.
  • One or more intervals may be described by signaling metadata. That is, metadata may describe each of one or more intervals.
  • the metadata may be MPD to be described later.
  • the media content component may be a subsequent component of media content with an assigned media component type.
  • Media content components may be encoded into separate media streams.
  • a media component is, for example, audio, video or timed text with specific attributes, such as bandwidth, language or resolution. It may be an encoded version of individual media types.
  • the media stream may be an encoded version of the media content component.
  • the media component type may be a single type of media content such as audio, video or text.
  • the media presentation may be a collection of data that establishes a bounded or unbounded presentation of the media content.
  • the media presentation may be a collection of data accessible to clients of DASH to provide streaming services to the user.
  • Media Presentation Description may be a stylized description of a media presentation for providing a streaming service.
  • the MPD may be a document containing metadata required by a client of DASH to construct a suitable HTTP-URL to access the segment and to provide a streaming service to the user.
  • the media presentation may include possible updates of the MPD and may be described by the MPD.
  • the MPD may be a document that contains metadata required by the DASH client to 1) access segments, and 2) configure HTTP-URLs suitable for providing a streaming service to the user.
  • HTTP-URLs can be absolute or relative.
  • the MPD may be an XML-document.
  • the MPD may define a format for informing resource identifiers for the segment.
  • the MPD may provide a context for the identified resources within the media presentation.
  • Resource identifiers may be HTTP-URL. URLs may be limited by the byte range attribute.
  • the period may be an interval of media presentation.
  • a continuous sequence of all periods may constitute a media presentation. That is, the media presentation can include one or more periods. Or, the media presentation may consist of a sequence of one or more periods.
  • Representation may be a collection and encapsulation of one or more media streams in a delivery format and may be associated with descriptive metadata.
  • the representation may be a structured collection of one or more media content components within one period. That is, the representation may be one of alternative choices of a complete set or subset of media content components that make up the media content for a defined period.
  • Representation may include one or more media streams.
  • Representation may begin at the beginning of a period (ie, a period including itself) and continue to the end of the period.
  • Representation is an alternative to media content or a subset of media content, with different encoding choices, such as bitrate, resolution, language, codec, and the like. alternative) may be one of the choices.
  • the MPD (or MPD element) may provide descriptive information that enables the client to select one or more representations.
  • representation and media representation may be used in the same sense.
  • a segment may be the smallest accessible unit in an MPD with a defined format.
  • segments and media segments may be used in the same sense.
  • Each interval may be divided into segments.
  • segments and fragments may be used in the same sense.
  • Segment may be a term of Generation Partnership Project (3GPP) adaptive HTTP streaming.
  • a segment may be an object (eg, an object in response to an HTTP / 1.1 GET request for an HTTP-URL (or a GET request for a portion indicated by a byte range), as defined in RFC 2616).
  • entity may refer to a body.
  • the terminal can play the media content using the received bytes (ie, the segment).
  • a sub-segment can mean the smallest unit in segments that can be indexed by a segment index at the segment level.
  • the sub-representation may be part of the representation described in the MPD staged over the entire period.
  • Random Access Point is a specific location within a media segment.
  • the RAP is identified as being a location that can continue playback from the location of the RAP using only the information contained within the media segment.
  • Dynamic Adaptive Streaming over HTTP may provide a media-streaming model for the delivery of media content.
  • the client can exclusively control the session of the delivery.
  • the client can use the HTTP protocol to request data from a standard web server without DASH-specific functions.
  • the DASH standard can focus on the data format used to provide the DASH media presentation.
  • a collection of encoded and deliverable versions of media content and appropriate descriptions of these versions can form a media presentation.
  • Media content may consist of one or more media content cycles that continue over time.
  • Each media content period may consist of one or more media content components.
  • audio components and video components in various languages may be examples of one or more media content components.
  • Each media content component may be assigned a media content component type.
  • An example of a media content component type may be audio or video.
  • Each media content component may have one or more encoded versions.
  • the encoded version of the media content component may be named a media stream.
  • Each media stream may inherit the properties of media content, media content cycles, or media content used for encoding the media stream.
  • the media stream may be assigned properties of an encoding process, such as, for example, sub-sampling, codec parameters, encoding bitrate, and the like.
  • the metadata may change depending on the static or dynamic choice for media content components and media streams.
  • 1 shows a high-level data model of DASH.
  • DASH may be based on a hierarchical data model such as shown in FIG. 1.
  • DASH media presentation can be described by an MPD document.
  • the MPD may describe a sequence of periods over time.
  • the sequence of periods can make a media presentation.
  • the cycle may generally represent the media content cycle while a consistent set of encoded versions of the media content is available. That is, bitrates, languages, captions and subtitles may not be changed during one period.
  • the period may include one or more adaptation sets.
  • the adaptation set may represent a set of replaceable encoded versions of one or more media content components. For example, there may be one adaptation set for the primary video component, and there may be one separate adaptation set for the primary audio component. For example, if there are other available materials, such as captions and audio descriptions, the other available materials may have separate adaptation sets.
  • the adaptation set may include one or more representations.
  • Representation may describe a deliverable encoded version of one or more media content components.
  • Representation may include one or more media streams. Any single representation in the adaptation set may be sufficient to render the included media content components.
  • a client can switch from one representation to another to adapt to network conditions or other factors within one period. The client may ignore representations that the client itself does not support or rely on for some reason not suitable codec or other rendering techniques.
  • the content can be divided into segments over time. That is, the representation may include one or more segments.
  • a segment may be the basic unit of data, as represented in the MPD.
  • the URL of each segment may be provided. Providing the URL of a segment may mean that the segment is the largest unit of data that can be retrieved by a single HTTP request.
  • the MPD may include a byte range of the URL along with the URL. Thus, a segment may be included within the byte range of another larger resource.
  • MPD may comprise a chain of one or more periods.
  • Each period may comprise one or more adaptation sets. If the adaptation set includes one or more media content components, each media content component can be defined individually.
  • Each adaptation set may include one or more representations.
  • Each representation may include one or more sub-representations.
  • Each representation may include one or more segments.
  • a segment may comprise metadata for accessing, decoding and displaying media data and / or media content comprising the segment.
  • Adaptation sets, representations, and sub-representations may share common attributes and elements.
  • Each segment may comprise one or more subsegments.
  • An MPD document representing an MPD may contain an MPD element.
  • Table 1 below may represent the XML schema of the MPD according to an example.
  • Table 1 may represent an initial part of the XML schema of the MPD. Initial portions of the above may include namespaces and other definitions.
  • Mechanisms for referencing remote elements from the local MPD may be defined.
  • a subset of W3C XLINK simple links with 1) limited syntax and semantics, and 2) a processing model can be defined.
  • the XLINK attributes used as part of ISO / IEC 23009 may be as follows.
  • the xlink: type attribute can define the type of W3C XLINK that is used.
  • the xlink: href attribute can use a URI to identify a remote element, as defined in IETF RFC 3986.
  • the xlink: show attribute can define the desired behavior when a remote element is dereferenced once in the MPD, as defined by the W3C XLINK.
  • the xlink: actuate attribute may define the desired timing at which the remote object is referenced in the MPD, as defined in W3C XLINK.
  • Table 2 below may represent an XML schema of XLINK according to an example.
  • href attribute may be as follows:
  • URI references to remote elements that cannot be resolved may be treated as invalid references and invalidate the MPD.
  • URI references to remote elements that are inappropriate targets for a given reference may be treated as invalid references and may invalidate the MPD.
  • URI references referencing themselves directly or indirectly may be treated as invalid circular references and may invalidate the MPD.
  • Any reference to the remote element can be an HTTP-URL.
  • URI reference is relative
  • reference resolution may apply.
  • the MPD element may have the following attributes.
  • the id attribute may specify an identifier of the media presentation.
  • the id attribute may be an identifier that is unique within the scope of the published MIDI presentation.
  • the type attribute may specify whether the MPD can be updated.
  • the mediaPresentationDuration attribute may specify the duration of the entire media presentation. If the mediaPresentationDuration attribute is absent, the duration of the media presentation may be unknown.
  • the maxSegmentDuration attribute may specify the maximum duration of any segment in any representation in the media presentation.
  • the MPD element may have the following elements.
  • the ProgramInformation element may have an index of 0 to N and may specify descriptive information about the program. Having an index of N may mean that the number of instances of the element may be up to unlimited. Having an index of zero may mean that an element is optional. Having an index of at least 1 may indicate that the element is mandatory. That is, the range of the index may indicate an occurrence number or cardinality.
  • the BaseURL element may have an index of 0 to N and may specify a base URL that may be used for reference resolution and alternative URL selection.
  • the Location element may have an index of 0 to N and may specify a location available to the MPD.
  • the Period element may have an index of 1 to N and may specify information of a period.
  • the Metrics element may have an index of 0 to N and may specify DASH metrics.
  • Table 3 below may indicate XML syntax of an MPD element according to an example.
  • the media presentation may include one or more periods.
  • the period may be defined by the Period element within the MPD element.
  • the period may be divided into a regular period and an early available period.
  • PeriodStart time of the regular period may be defined as follows.
  • the period may be a regular period and the PeriodStart time may be equal to the value of the start attribute.
  • the period may be a regular period.
  • the start time of the period PeriodStart may be the sum of the start time of the previous period and the value of the duration attribute of the previous period.
  • the PeriodStart time may be zero.
  • the start attribute is absent, 2) the previous Period element does not contain a duration attribute, or the Period element is the first in the MPD, and 3) the MPD's type attribute is 'dynamic' May be an early available cycle.
  • Period element can have the following attributes.
  • the xlink: href attribute may specify a reference to an external Period element.
  • the xlink: actuate attribute may specify processing instructions.
  • the value of the link: actuate attribute may be "onload” or "onRequest”.
  • the id attribute may specify an identifier of a cycle.
  • the id attribute may be an identifier that is unique within the scope of the MIDI presentation.
  • the start attribute may specify the PeriodStart time of the period.
  • the PeriodStart time can be used as an anchor to determine the MPD start time of each media segment and can be used to determine the presentation time of each access unit in the media presentation timeline.
  • the duration attribute may specify a duration of a period for determining a PeriodStart time of a next period.
  • Period element may have the following elements.
  • the BaseURL element may have an index of 0 to N and may be used for reference solution and alternative URL selection.
  • AdaptationSet element may have an index of 1 to N, and may specify information of the adaptation set.
  • the Subset element may have an index of 0 to N and may specify a subset.
  • Table 4 below may indicate XML syntax of a Period element according to an example.
  • Each period may include one or more adaptation sets.
  • the adaptation set may be described by the AdaptationSet element included in the Period element.
  • the adaptation set may include alternative encodings of one or more media components. Each alternative encoding may be included in the representation.
  • One or more representations included in one adaptation set may represent the same media content component and may include one or more media streams that are considered perceptually the same.
  • One or more representations may be arranged in an adaptation set according to media content component properties presented within the one or more representations above.
  • the media content component properties are assigned to 1) the language described by the lang attribute, 2) the media content type described by the contentType attribute, 3) the picture aspect ratio described by the par attribute, and 4) the Role element.
  • the AdaptationSet element may include basic values of elements and attributes associated with one or more representations that the adaptation set includes.
  • a list of possible present elements and attributes common to the AdaptationSet element and the Representation element is described below.
  • the Adaptation element may support the description of a range for the bandwidth attribute, width attribute, height attribute, and framerate attribute associated with one or more representations included in the adaptation set.
  • the above description may provide a summary of all values for all of the one or more representations that the adaptation set includes.
  • One or more representations that an adaptation set includes may not have values outside the range defined within the adaptation set.
  • the adaptation set can be classified into groups by using the group attribute.
  • AdaptationSet element may have the following attributes.
  • the xlink: href attribute may specify a reference to an external AdaptationSet element.
  • the xlink: actuate attribute may specify processing instructions.
  • the value of the link: actuate attribute may be "onload” or "onRequest”.
  • the id attribute may specify an identifier of the adaptation set within a range of periods.
  • the id attribute may be an identifier that is unique within the range containing the period.
  • the id attribute may be absent within the remote element.
  • the group attribute may be an identifier for a group that is unique within a range of periods.
  • the lang attribute may declare the language code (s) of the adaptation set. Syntax and semantics according to IETF RFC 5646 may be used.
  • the contentType attribute may specify the media content component type of the adaptation set.
  • the value of "type", the highest-level content-type, may be defined in RFC1521.
  • the par attribute may specify the photo length ratio.
  • the par attribute may contain two integers. If the par attribute is present, the width and height attributes of the representations may also be present.
  • the minBandWidth attribute may specify the minimum bandwidth attribute value of all representations in the adaptation set.
  • the maxBandWidth attribute may specify a maximum bandwidth attribute value of all representations in the adaptation set.
  • the minWidth attribute may specify the minimum width attribute value of all representations in the adaptation set.
  • the maxWidth attribute may specify the maximum width attribute value of all representations in the adaptation set.
  • the minHeight attribute may specify the minimum height attribute value of all representations in the adaptation set.
  • the maxHeight attribute may specify the maximum height attribute value of all representations in the adaptation set.
  • the minFrameRate attribute may specify the minimum frameRate attribute value of all representations in the adaptation set.
  • the maxFrameRate attribute may specify the maximum frameRate attribute value of all representations in the adaptation set.
  • AdaptationSet element may have the following elements.
  • the Accessibility element may have an index of 0 to N and may specify information about an accessibility scheme.
  • the role element may have an index of 0 to N and may specify information on a role annotation scheme.
  • the Rating element may have an index of 0 to N and may specify information about a rating scheme.
  • the Viewpoint element may have an index of 0 to N and may specify information on the viewpoint annotation scheme.
  • the ContentComponent element may have an index of 0 to N and may specify properties of a media content component included in the adaptation set.
  • the BaseURL element may have an index of 0 to N and may be used for reference solution and alternative URL selection.
  • the Representation element may have an index of 0 to N and may specify a representation. Within each adaptation set there may be at least one representation element. The Representation element may be part of a remote element.
  • Table 5 below may indicate XML syntax of an AdaptationSet element according to an example.
  • the adaptation set may include one or more media content components. Properties of each media content component may be described by a ContentComponet element, and may be directly described by an AdaptationSet element when there is one media content component in the adaptation set.
  • the ContentComponent element may have the following attributes.
  • the id attribute may specify an identifier of a media component.
  • the id attribute can be unique within the scope of the adaptation set.
  • the lang attribute may declare the language code (s) of the media content component. Syntax and semantics according to IETF RFC 5646 may be used.
  • the contentType attribute may specify the type of the media content component.
  • the value of "type", the highest-level content-type, may be defined in RFC1521.
  • the par attribute may specify the photo length ratio.
  • the par attribute may contain two integers. If the par attribute is present, the width and height attributes of the representations may also be present.
  • the ContentComponent element may have the following element.
  • the Accessibility element may have an index of 0 to N and may specify information about an accessibility scheme.
  • the role element may have an index of 0 to N and may specify information on a role annotation scheme.
  • the Rating element may have an index of 0 to N and may specify information about a rating scheme.
  • the Viewpoint element may have an index of 0 to N and may specify information on the viewpoint annotation scheme.
  • Table 6 below may indicate XML syntax of a ContentComponent element according to an example.
  • the AdaptationSet element may include one or more Representation elements.
  • the representation may be one of alternative choices of a complete set or subset of media content components that make up the media content for a defined period.
  • Representation may begin at the beginning of PeriodStart of the period and continue until the end of the period.
  • the end of the period may be the beginning of the next period or the end of the media presentation.
  • Each representation may include one or more media streams.
  • Each media stream may be an encoded version of a media content component.
  • the representation may include one or more segments.
  • the Representation element may have the following attributes. '
  • the id attribute may specify the identifier of the representation.
  • the id attribute may be unique within a range of periods unless the representation is functionally identical to other representations within the same period.
  • the id attribute may not include whitespace characters.
  • the bandwidth attribute may specify the boundary of the data rate and data rate variation of the representation.
  • the qualityRanking attribute may specify a quality ranking of the representation relative to other representations in the same adaptation set.
  • the dependencyID attribute may specify all complementary representations that depend on the decoding and / or presentation process.
  • the dependencyID attribute may be a space separated list of values of the id attributes.
  • the Representation element may have the following elements.
  • the BaseURL element may have an index of 0 to N and may be used for reference solution and alternative URL selection.
  • the SubRepresentation element may have an index of 0 to N and may specify information about a sub-representation embedded in the representation. It may be used for reference solution and alternative URL selection.
  • Table 7 below may indicate XML syntax of a Representation element according to an example.
  • Sub-representations may be embedded within regular representations and may be described by SubRepresentation elements.
  • the Representation element may include a SubRepresentation element.
  • the SubRepresentation element may describe the properties of one or more media content components embedded within the representation. Representations and sub-representations may share common attributes and elements.
  • the SubRepresentation element may have the following attributes.
  • the level attribute may specify the level of the sub-representation.
  • the dependency attribute may specify a set of sub-representations in the representation that the sub-representation depends on in the decoding and presentation process as a space separated list of values of the level attribute.
  • the bandwidth attribute is the same as the bandwidth attribute of the representation, but may be applied to the sub-representation.
  • the contentComponet attribute may specify the set of all media content components included in the sub-representation as a space separated list of values of the id attribute of the ContentComponent element.
  • Table 8 below may indicate XML syntax of a SubRepresentation element according to an example.
  • the AdaptationSet element, Representation element, and SubRepresentation element may have the following attributes in common.
  • the profile attribute may specify a profile between associated presentation (s) according to the media presentation profiles.
  • the value of the profile attribute may be a subset of any higher level value of the document hierarchy (representation, adaptation set, and MPD). If the profile attribute is absent, the value of the profile attribute may be considered equal to the value of the next higher level of the document hierarchy. For example, if the profile attribute for the representation is absent, the profile attribute for the adaptation set may be valid for the representation.
  • the width attribute may specify the horizontal visual presentation size of the video media type on the grid determined by the sar attribute. In the absence of a sar attribute, the width and height may be specified as if the value of the sar attribute is "1: 1".
  • the height attribute may specify the vertical visual presentation size of the video media type on the grid determined by the sar attribute.
  • the sar attribute may specify an example length ratio of the video media component type.
  • the sar attribute can take the form of a string containing two integers separated by ":".
  • the first integer may specify a horizontal size based on any unit of encoded video pixels.
  • the second integer may specify a vertical size based on any unit of encoded video pixels.
  • the frameRate attribute may specify the output frame rate of the video media type in the representation. If the frame or frame rate varies, the value of the frameRate attribute may be half the average frame or average frame rate over the entire duration of the representation.
  • the audioSampleRate attribute may be a decimal integer value that specifies the sampling rate of the audio media component type, and may be a pair of decimal integer values separated by spaces indicating the minimum sampling rate and the maximum sampling rate of the audio media component type, respectively. .
  • the value of the audioSamplingRate attribute may be samples per second.
  • the mimeType attribute may specify the MIME type of the concatenation of the initial content segment.
  • the codecs attribute may specify codecs to be played within the representation.
  • the AdaptationSet element, Representation element, and SubRepresentation element may have the following elements in common.
  • the FramePacking element may have an index of 0 to N, and may specify frame-packing arrangement information of a video media component type.
  • the AudioChannelConfiguration element may have an index of 0 to N and may specify an audio channel configuration of an audio media component type.
  • the ContentProtection element which may have an index of 0 to N, specifies information about the content protection schemes used for the associated representation.
  • Table 9 below may indicate common XML syntax of an AdaptationSet element, a Representation element, and a SubRepresentation element according to an example.
  • a segment may be a minimum addressable unit that has a defined format and is described by the MPD.
  • HTTP-URL may be defined as ⁇ absolute-URI> according to RFC 3986.
  • HTTP-URL can have a fixed scheme "http: //" or "https: //" and can be limited by the byte range if the range attribute is provided in the URL.
  • the byte range may be expressed as a byte-range-spec defined in RFC 2616.
  • the byte range may be limited to a single expression that identifies a continuous range of bytes.
  • the representation may be assigned segment information through the presence of a BaseURL element, a SegmentBase element, a SegmentTemplate element, and / or a SegmentList element.
  • Segment information may provide information about the location, availability, and properties of all of the one or more segments that the presentation includes. In particular, the initial content setting, the presence and location of media, index and bitstream switching segments can be provided.
  • SegmentBase element, SegmentTemplate element, and SegmentList element may exist in Representation element.
  • the SegmentBase element, the SegmentTemplate element, and the SegmentList element may exist in a Period element or an AdaptationSet element to indicate default values. If a SegmentTemplate element or SegmentList element exists at one level of the hierarchy, it may not be present at a lower hierarchy level.
  • the SegmentBase element, SegmentTemplate element, and SegmentList element may inherit the attributes and elements of the same element at a higher level. If the same attribute or element exists at both levels, the lower level attribute or element may have precedence over the higher level attribute or element.
  • the SegmentBase element may contain sufficient information if a single media segment is provided per representation and the media segment URL is included in the BaseURL element. If multiple media segments are present, a SegmentList element or a Segment Template element sharing the multi-segment base information may be used.
  • a duration attribute or a SegmentTimeLine element may be present.
  • the segment list may be defined by one or more SegmentList elements.
  • Each SegmentList element may include a list of SegmentURL elements for a concatenated list of segment URLs.
  • Each segment URL may include a media segment URL and may include a byte range.
  • the segment URL element may include an index segment.
  • Segment templates can be defined by the SegmentTemplate element. Certain identifiers that are replaced by dynamic values to generate a list of segments may be handed to the segments.
  • Segment information may provide the following information.
  • the segment element or the child element of the segment element may include the following attributes.
  • the duration element may indicate a fixed approximate segment duration. All segments in the Representation element may have the same duration. However, the last segment in the period may have a different duration.
  • the sourceURL attribute may specify part of the source URL.
  • sourceURL may have a format of ⁇ absolute-URI> according to RFC 3986.
  • sourceURL may have a fixed scheme of ⁇ relative-ref> according to "http: //", "https: //" or RFC 3986. If the sourceURL attribute does not exist, any BaseURL element may be mapped to the sourceURL attribute. In addition, if the sourceURL attribute does not exist, a range attribute may be provided.
  • the range attribute may specify a restriction on the byte range of the HTTP-URL.
  • the byte range can be represented and formatted according to the byte-range-spec defined in RFC 2616. If the range attribute is absent, the element can refer to the entire resource referenced in the sourceURL attribute.
  • Table 10 below may show XML syntax for information related to a segment in an example.
  • representations can be described by various metadata.
  • the characteristics of the representation may be mostly for selecting representations based on the characteristics of the terminal.
  • characteristics of the terminal may include a resolution and a connection bandwidth.
  • the accessibility characteristics of representations can be used to support disabled users who suffer from difficulties in perceiving multimedia content.
  • the following embodiment for supporting accessibility is intended to enable 1) MPEG-21 Digital Item Adaptation (DIA) accessibility description and 2) enhancement / replacement of content components. (facilitate) May contain some hints.
  • the MPEG-21 DIA accessibility description which includes the VisualImpairmentType and / or AuditoryImpairmentType elements of ISO / IEC 21000-7, provides a variety of visual or auditory impairments for the user. It makes it possible to explain the symptoms. Visual or auditory disturbances of the user may include color deficiency or low vision.
  • URIs for representations including video and / or audio content may be defined as shown in Table 11 below.
  • the URI may indicate accessibility schemes based on the accessibility definition of ISO / IEC 21000-7.
  • ⁇ value> is "s0"
  • a suitable representation is prepared in advance, and the associated accessibility description may be described by the Accessibility element. Different ways may be used to convey the above description.
  • the accessibility description may be carried by the MP21Access element and may be referenced by the xlink: href attribute defined below.
  • the representation may be used for the handicapped person in which specific symptoms of the handicapped are described by the accessibility description.
  • the media of the representation can be augmented to cope with the problem of the handicapped.
  • the increase in representation may be contrast or adjusting of color.
  • the aforementioned Accessibility element can be extended as follows.
  • the new child element "MP21Access” can be used as a container for describing MPEG-21 DIA accessibility.
  • an additional xlink: href attribute may be added to the Accessibility element to refer to an external description that includes MPEG-21 DIA accessibility elements.
  • a new attribute "use” may be employed as a hint for the client to inform the intended use of the representation for the purpose of accessibility.
  • the Group element, Adaptation Set element, Representation element, or SubRepresentation element may further have the following attributes to support accessibility.
  • use attribute may indicate the intended use of the representation (s).
  • the value of the use attribute may be signLang, subtitle, caption, description or plusDes.
  • signLang can be used as a sign language.
  • subtitle can indicate use as a subtitle.
  • caption can indicate use as a caption.
  • the caption may be a subtitle with additional description.
  • description may indicate use as a description, such as, for example, an audio description of a movie.
  • plusDesc may indicate that the content of the corresponding representation has been augmented with a description. For example, it may point to an augmented sound track including an audio description.
  • the xlink: href attribute may provide a reference to an external MPEG-21 DIA that includes the AuditoryImpairmentType and / or VisualImpairmentType elements of the ISO / IEC IS 21000-7 element.
  • the representation or representation group associated with the above description may be suitable for people with disabilities having the problems described by the above description.
  • the xlink: actuate attribute can provide processing instructions.
  • the processing instructions may be "onLoad” or "onRequest”.
  • the xlink: actuate attribute cannot exist if the xlink: href attribute is absent.
  • the Group element, Adaptation Set element, Representation element, or SubRepresentation element may further have the following elements to support accessibility.
  • the Accessibility element may have an index of 0 to N and may provide information about an accessibility information scheme.
  • the MP21Access element may have an index of 0 to 1 and may include an MPEG-21 DIA description together with AuditoryImpairmentType and / or VisualImpairmentType elements of ISO / IEC 21000-7. Representations or representation groups associated with the above description may be suitable for people with disabilities having the problems described by the above description.
  • Table 12 below may indicate XML syntax of each element for supporting accessibility of a Group element, an Adaptation Set element, a Representation element, or a SubRepresentation element according to an example.
  • Metadata describing the relationship between representation groups is relatively rare. For example, it is difficult to tell a client which video stream a given audio description is associated with.
  • two representation groups may be conceptually equivalent, and only one representation group may have to be selected depending on the context.
  • two simple attributes are proposed for the above-mentioned purposes.
  • the metadata described below may be useful not only for accessibility but also for general adaptability.
  • the Group element, Adaptation Set element, Representation element, or SubRepresentation element may further have the following attributes to describe the relationship between representations.
  • the usedWith attribute may point to a list of one or more representations or representation groups. If the current representation or representation group is used, the usedWith attribute can be used with any item in the list above. For example, a representation group for audio description may be used with a particular representation group of video components.
  • the value of the usedWith attribute may be a whitespace-separated list. Each item of the space separated list may be an identification of one representation group or one representation.
  • the identification of the representation group may be a value of the group attribute or the adaptationSet attribute.
  • the identification of the representation may have the format of the repid attribute.
  • the repid attribute may be a value of the id attribute of the representation.
  • the equivalentTo attribute may point to a list of one or more representations or representation groups.
  • the current representation can be used in place of any of the items in the list above.
  • a representation group for subtitles can be used in place of a particular representation group of audio components.
  • the format of the equivalentTo attribute may be the same as the format of the usedWith attribute.
  • Table 13 below may indicate an XML syntax for describing a relationship between representations of a Group element, an Adaptation Set element, a Representation element, or a SubRepresentation element according to an example.
  • Bit depth may refer to the number of bits used to represent a luma / chroma sample of visual content.
  • the Group element, Adaptation Set element, Representation element, or SubRepresentation element may further have the following attributes to provide bit depth.
  • the bitDepth attribute may indicate the number of bits for representing a luma / chroma sample of visual content. For example, in different video pictures, if the bit depth varies, the value of the bitDepth attribute may be the maximum value of the bit depths.
  • Table 14 below may indicate XML syntax for supporting bit depths of a Group element, an Adaptation Set element, a Representation element, or a SubRepresentation element according to an example.
  • two attributes minBitDepth and maxBitDepth may be added to the Representation Group element or the Adaptation Set element.
  • the Representation Group element, Group element, or Adaptation Set element may further have the following attributes to describe the maximum bit depth and the minimum bit depth.
  • the xlink: href attribute may indicate a reference to an external Group element or an Adaptation Set element.
  • the xlink: actuate attribute can provide processing instructions.
  • the processing instructions may be "onLoad” or "onRequest”.
  • the minBitDepth attribute may indicate the minimum bit depth value for all representations in the group or adaptation set.
  • the maxBitDepth attribute may indicate a maximum bit depth value, for all representations in a group or adaptation set.
  • Table 15 below may indicate XML syntax for supporting a minimum bit depth value and a maximum bit depth value of a Representation Group element, a Group element, or an Adaptation Set element according to an example.
  • FIG. 2 is a signal flowchart of a content processing method according to an embodiment of the present invention.
  • the terminal 200 may be a client of the above-described DASH.
  • the client of DASH may be compatible with the client specified in RFC 2616.
  • the client of the DASH may generally use the HTTP GET method or the HTTP partial GET method, as specified in RFC 2616, to access the segments or portions of the segments.
  • Server 210 may host segments of DASH. Server 210 may be compatible with the server specified in RFC 2616.
  • the terminal 200 may receive metadata of media content from the server 210. That is, the server 210 transmits metadata about the media content to the terminal 200.
  • Media content may include one or more periods.
  • the metadata may be MPD.
  • the MPD may provide enough information for the terminal 200 to provide a streaming service to the user by accessing segments through a protocol specified within a scheme of defined resources.
  • the specified protocol may be HTTP / 1.1.
  • the server 210 may provide an MPD delivery function. MPDs may be generated in accordance with DASH media presentation preparation.
  • the terminal 200 may process the received metadata. In operation 230, the terminal 200 may extract information provided from the metadata or information included in the metadata.
  • the terminal 200 may access a segment of media content based on the information provided by the metadata.
  • Server 210 may be an HTTP server that provides a DASH segment delivery function.
  • each period may include one or more groups, and each group may include one or more representations of media content. That is, each period may include one or more representations of media content. Thus, each representation may begin at the beginning of a period that includes it and continue until the end of the period. Each representation may include one or more segments.
  • the terminal 200 may request the server 210 for a segment of media content using the URL of the segment.
  • the URL may be interpreted correlated with the base URL element described above.
  • the URL of the segment may be generated based on the base URL element.
  • the MPD provides the terminal 200 with sufficient information to provide a streaming service to the user by 1) requesting the segments from the server 210 and 2) demultiplexing, decoding and rendering the media streams contained in the segment. can do.
  • the terminal 200 may request the server 210 for a suitable segment for a specific period based on the metadata processing. That is, the requested segment may be selected based on metadata.
  • the terminal 200 may request the segment 210 from the server 210 by executing the HTTP GET method.
  • the metadata may include a "range” attribute.
  • the request may be a request for bytes specified by the "range” attribute of the resource pointed to by the URL.
  • the URL of the segment can be an absolute URL or a relative URL.
  • the terminal 200 may select a period, an adaptation set, and a representation using the elements or attributes included in the metadata, and may request segments based on the selected period, the adaptation set, or the elements and attributes of the representation. Can be selected.
  • each representation may include information to assist people with disabilities who have difficulty in perceiving media content.
  • the media component of the representation can be augmented to address the problem of the handicapped.
  • the increase may be one or more of an increase in contrast and an increase in color correction.
  • the above information may indicate the intended use of the representation.
  • intended use may be one or more of sign language, subtitles, captions, and description.
  • the information may also include an attribute that provides a reference to an external MPEG-21 Digital Item Adaptation (DIA).
  • DIA MPEG-21 Digital Item Adaptation
  • the metadata may include an attribute that describes a relationship between one or more representations.
  • the first representation of the one or more representations may include an attribute that points to a list of representations of some of the one or more representations, the first representation of the some of the representations Can be used with one.
  • the first representation may be a representation for audio description, and some representations may be representations of video components, respectively.
  • the second representation of the one or more representations may include an attribute that points to a list of representations of some of the one or more representations above, and one of some of the representations above The representation of may be used in place of the second representation.
  • each representation may include a bit depth attribute used to select a different representation for terminals with different display bit depths.
  • the bit depth attribute may indicate the number of bits to represent the luma / chroma sample of the visual content.
  • server 210 needs to parse media content (eg, MP4 file for SVC) and may extract data parts suitable for the requested segment. .
  • media content eg, MP4 file for SVC
  • the server 210 may transmit appropriate segments to the terminal 200 for each request from the terminal 200.
  • the terminal 200 may receive segments from the server 210.
  • the terminal 200 may play the media content by decoding and rendering data of the media content included in the segment.
  • Terminal 200 may play the media content using the received segment by repeating steps 220-260.
  • FIG 3 illustrates categories of signaling information according to an embodiment of the present invention.
  • the signaling information (ie, metadata) may be divided into the following categories 1) to 4).
  • General information 310 may include a general description of the content and a general description of each interval, such as, for example, duration and start time.
  • Quality of Service (QoS information) 320 Describe the characteristics of each alternative, such as bitrate, resolution, quality. That is, the quality information of the service may describe the characteristics of each of the alternatives of the content.
  • Alternatives may be physical (ie, pre-generated) and virtual (ie, generated on-the-fly). Based on the information of the alternatives, the client can select the appropriate alternative fragment. Thus, it is possible to support the adaptability to the contexts of terminals and networks.
  • Mapping Information 330 Describes locations for retrieving content. Depending on the specific cases, different alternatives may have the same or different positions.
  • Client request 340 This kind of signaling information may conform to the format of an HTTP 1.1 request message.
  • the parameters requested by the client may derive from the information of the categories 1) to 3) above.
  • FIG 4 illustrates layers of content divisions and levels of signaling information according to an embodiment of the present invention.
  • the signaling of metadata may be physically separated into content-level information 410, interval level information 420, QoS information 430, and mapping 440 information.
  • Linking of the content-level information 310, the interval level information 420, the QoS information 430, and the relevant portions of the mapping 440 may be accomplished by referencing.
  • the processing model when only content-level information 310 and interval level information 420 are sent to the client may be “server-based”.
  • Metadata parts may enable efficiency in storage and delivery. For example, during a session, metadata of content-level information 410 may be sent once, and only interval level information 420 may be updated periodically. Similarly, one file containing QoSInfo part 430 may be used for different intervals and different content.
  • FIG. 5 is a structural diagram of a terminal according to an embodiment of the present invention.
  • the structure of the terminal 100 to be described below may represent logical components of a conceptual DASH client model.
  • the terminal 100 may include an access engine 510 and a media engine 520.
  • the access engine 510 may be a DASH access engine.
  • the access engine 510 may receive metadata (eg, MPD) from the server 110.
  • metadata eg, MPD
  • the access engine 510 may configure the requests and issue the configured requests to the server 110.
  • the access engine 510 may receive the content (eg, calendar, segments, or portions of segments) from the server 110.
  • content eg, calendar, segments, or portions of segments
  • the access engine 510 can provide the content to the media engine 520.
  • the output of the access engine 510 is media (or media) of an MPEG container (e.g., ISO / IEC 14492-12 ISO Base Media File Format or ISO / IEC 13818-2 MPEG-2 Transport Stream). May be included).
  • the output of the access engine 510 may include timing information that maps the internal timing of the media to the timeline of the media presentation.
  • the access engine 510 may perform the functions described as being performed by the terminal 200 in the steps 220 to 260 described above with reference to FIG. 2.
  • the media engine 520 may play the provided content. That is, the media engine 520 may output media using the media and timing information output from the access engine 510.
  • bit depth attribute is described.
  • bit depth is one of the characteristics of the content and may be included in the Media Presentation description (MPD).
  • MPD Media Presentation description
  • Bit depth represents the number of bits used to represent a sample of media content.
  • content can be encoded using different bit depths.
  • Some profiles in AVC vary in bit depth from 8 to 16 bits.
  • each terminal can accept different bit depths. The following attributes are added to the adaptation set of the representation so that these terminals can select a representation representing a bit depth suitable for each terminal.
  • Option 1 expresses several bit depths using one attribute
  • Option 2 uses a corresponding attribute to express each bit depth
  • Table 16 below shows common adaptation sets, representations and sub-representation attributes and elements.
  • Table 17 below shows syntax of an attribute for describing bit depth.
  • Table 18 shows semantics of attributes describing maximum bit depths and minimum bit depths.
  • Table 19 shows syntax of attributes describing maximum bit depth and minimum bit depth.
  • Table 20 below shows common adaptation sets, representations and sub-representation attributes and elements.
  • Table 21 shows syntax of an attribute for describing bit depth.
  • Table 22 below shows the semantics of attributes describing maximum bit depths and minimum bit depths.
  • Table 23 shows the syntax of attributes describing maximum bit depth and minimum bit depth.
  • Method according to an embodiment of the present invention is implemented in the form of program instructions that can be executed by various computer means may be recorded on a computer readable medium.
  • the computer readable medium may include program instructions, data files, data structures, etc. alone or in combination.
  • Program instructions recorded on the media may be those specially designed and constructed for the purposes of the present invention, or they may be of the kind well-known and available to those having skill in the computer software arts.
  • Examples of computer-readable recording media include magnetic media such as hard disks, floppy disks, and magnetic tape, optical media such as CD-ROMs, DVDs, and magnetic disks, such as floppy disks.
  • Examples of program instructions include not only machine code generated by a compiler, but also high-level language code that can be executed by a computer using an interpreter or the like.
  • the hardware device described above may be configured to operate as one or more software modules to perform the operations of the present invention, and vice versa.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Databases & Information Systems (AREA)
  • Computer Security & Cryptography (AREA)
  • Computer Graphics (AREA)
  • Information Transfer Between Computers (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

L'invention concerne un procédé et un appareil destinés à un service de diffusion en flux adaptatif sur HTTP qui utilise des métadonnées de contenu multimédia. Le contenu multimédia comprend une série d'un ou de plusieurs cycles. Chaque cycle comprend une ou plusieurs représentations. Les métadonnées comprennent des informations décrivant la relation entre les représentations, et comprennent des informations relatives à des terminaux comportant différentes profondeurs de bits d'affichage.
PCT/KR2015/000137 2014-01-08 2015-01-07 Procede d'affichage de profondeur de bits pour la lecture d'une video utilisant un service de diffusion en flux adaptatif sur http (dash) WO2015105327A1 (fr)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP15735102.4A EP3094097A4 (fr) 2014-01-08 2015-01-07 Procede d'affichage de profondeur de bits pour la lecture d'une video utilisant un service de diffusion en flux adaptatif sur http (dash)
CN201580004175.2A CN106063279A (zh) 2014-01-08 2015-01-07 用于播放使用dash视频的位深表现方法
US15/110,500 US20160337679A1 (en) 2014-01-08 2015-01-07 Method for displaying bit depth for playing video using dash

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
KR10-2014-0002487 2014-01-08
KR20140002487 2014-01-08
KR10-2015-0000758 2015-01-05
KR1020150000758A KR20150083429A (ko) 2014-01-08 2015-01-05 Dash를 사용하는 비디오 재생을 위한 비트 깊이 표현 방법

Publications (1)

Publication Number Publication Date
WO2015105327A1 true WO2015105327A1 (fr) 2015-07-16

Family

ID=53524106

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/KR2015/000137 WO2015105327A1 (fr) 2014-01-08 2015-01-07 Procede d'affichage de profondeur de bits pour la lecture d'une video utilisant un service de diffusion en flux adaptatif sur http (dash)

Country Status (1)

Country Link
WO (1) WO2015105327A1 (fr)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20120106631A (ko) * 2011-03-16 2012-09-26 한국전자통신연구원 레프리젠테이션을 사용하는 스트리밍 콘텐츠 제공 장치 및 방법
KR20120120916A (ko) * 2010-10-06 2012-11-02 한국항공대학교산학협력단 스트리밍 콘텐츠 제공 장치 및 방법
KR20130009740A (ko) * 2010-02-19 2013-01-23 텔레폰악티에볼라겟엘엠에릭슨(펍) 에이치티티피 스트리밍에서 적응화를 위한 방법 및 장치
US20130103849A1 (en) * 2011-09-21 2013-04-25 Qualcomm Incorporated Signaling characteristics of segments for network streaming of media data
KR20130063413A (ko) * 2011-12-06 2013-06-14 삼성전자주식회사 멀티미디어 컨텐트 전송 시스템에서 적응적 스트리밍을 이용한 트래픽 제어 방법 및 장치

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20130009740A (ko) * 2010-02-19 2013-01-23 텔레폰악티에볼라겟엘엠에릭슨(펍) 에이치티티피 스트리밍에서 적응화를 위한 방법 및 장치
KR20120120916A (ko) * 2010-10-06 2012-11-02 한국항공대학교산학협력단 스트리밍 콘텐츠 제공 장치 및 방법
KR20120106631A (ko) * 2011-03-16 2012-09-26 한국전자통신연구원 레프리젠테이션을 사용하는 스트리밍 콘텐츠 제공 장치 및 방법
US20130103849A1 (en) * 2011-09-21 2013-04-25 Qualcomm Incorporated Signaling characteristics of segments for network streaming of media data
KR20130063413A (ko) * 2011-12-06 2013-06-14 삼성전자주식회사 멀티미디어 컨텐트 전송 시스템에서 적응적 스트리밍을 이용한 트래픽 제어 방법 및 장치

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
"Dynamic adaptive streaming over Hypertext Transfer Protocol (HTTP) (DASH) - Part 1: Media presentation description and segment formats/ Amd 2", ISO/IEC DIS 23009-1, 24 October 2014 (2014-10-24)
See also references of EP3094097A4 *

Similar Documents

Publication Publication Date Title
WO2012125006A2 (fr) Appareil et procédé de fourniture de contenu diffusé en continu à l'aide de représentations
WO2020080873A1 (fr) Procédé et appareil pour diffuser en continu des données
WO2014209057A1 (fr) Procédé et dispositif pour émettre et recevoir un service de diffusion dans un système de diffusion hybride sur la base d'une connexion d'un réseau de diffusion terrestre et d'un réseau à protocole internet
WO2014171718A1 (fr) Dispositif de transmission par diffusion, dispositif de réception par diffusion, procédé fonctionnel pour dispositif de transmission par diffusion et procédé fonctionnel pour dispositif de réception par diffusion
WO2009131391A1 (fr) Procédé de génération et de lecture de contenus audio basés sur un objet et support d'enregistrement lisible par ordinateur pour l'enregistrement de données présentant une structure de format fichier pour un service audio basé sur un objet
WO2016018042A1 (fr) Appareil et procédé pour des processus d'émission/réception d'un signal de diffusion
WO2013048148A2 (fr) Procédé et appareil d'émission et de réception de contenu
WO2014007571A1 (fr) Procédé et appareil de traitement de signaux de service numérique
WO2014014252A1 (fr) Procédé et appareil pour le traitement de signaux de service numériques
WO2012011743A2 (fr) Appareil et procédé de fourniture de contenus diffusés en continu
WO2012033319A2 (fr) Appareil et procédé pour fournir un contenu en flux continu
WO2014025207A1 (fr) Procédé et appareil pour traiter un signal de diffusion contenant un service de radiodiffusion interactive
WO2015080414A1 (fr) Procédé et dispositif d'émission et de réception d'un signal de diffusion pour assurer un service de lecture spéciale
WO2011034283A1 (fr) Procédé de traitement de métadonnées epg dans un dispositif de réseau et dispositif de réseau pour commander ce traitement
WO2010008235A2 (fr) Procédé et appareil d'expression d'effets sensoriels, et support d'enregistrement lisible par ordinateur sur lequel sont enregistrées des métadonnées concernant la commande d'un dispositif sensoriel
EP3811618A1 (fr) Procédé et appareil pour diffuser en continu des données
WO2015034306A1 (fr) Procédé et dispositif pour transmettre et recevoir un contenu de diffusion uhd perfectionné dans un système de diffusion numérique
WO2012050405A2 (fr) Récepteur numérique et procédé permettant de traiter le contenu 3d d'un récepteur numérique
WO2016171518A2 (fr) Émetteur de signal de radiodiffusion, récepteur de signal de radiodiffusion, procédé d'émission d'un signal de radiodiffusion et procédé de réception d'un signal de radiodiffusion
WO2017061796A1 (fr) Dispositif d'émission de signal de radiodiffusion, dispositif de réception de signal de radiodiffusion, procédé d'émission de signal de radiodiffusion, et procédé de réception de signal de radiodiffusion
WO2013077670A1 (fr) Procédé et appareil pour un service de streaming servant à fournir des informations sur l'extensibilité et la visualisation
WO2015137669A1 (fr) Dispositif de réception de diffusion et procédé de fonctionnement, et dispositif compagnon interagissant avec le dispositif de réception de diffusion et procédé de fonctionnement
WO2017135673A1 (fr) Dispositif d'émission de signal de diffusion, dispositif de réception de signal de diffusion, procédé d'émission de signal de diffusion et procédé de réception de signal de diffusion
WO2018062641A1 (fr) Fourniture d'un service de réalité virtuelle en tenant compte de la zone d'intérêt
WO2021242066A1 (fr) Appareil et procédé de réalisation d'un codage par intelligence artificielle et d'un décodage par intelligence artificielle sur une image

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 15735102

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 15110500

Country of ref document: US

REEP Request for entry into the european phase

Ref document number: 2015735102

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2015735102

Country of ref document: EP