WO2012009245A1 - Method of picture-in-picture for multimedia applications - Google Patents

Method of picture-in-picture for multimedia applications Download PDF

Info

Publication number
WO2012009245A1
WO2012009245A1 PCT/US2011/043494 US2011043494W WO2012009245A1 WO 2012009245 A1 WO2012009245 A1 WO 2012009245A1 US 2011043494 W US2011043494 W US 2011043494W WO 2012009245 A1 WO2012009245 A1 WO 2012009245A1
Authority
WO
WIPO (PCT)
Prior art keywords
picture
information
stream
component
media stream
Prior art date
Application number
PCT/US2011/043494
Other languages
French (fr)
Inventor
Li Hua Zhu
Jun Li
Zhenyu Wu
Original Assignee
Thomson Licensing
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Thomson Licensing filed Critical Thomson Licensing
Priority to US13/809,812 priority Critical patent/US8767124B2/en
Priority to KR1020137003063A priority patent/KR20130127423A/en
Priority to JP2013519735A priority patent/JP2013538482A/en
Priority to EP11733971.3A priority patent/EP2594070A1/en
Priority to CN2011800347149A priority patent/CN103004221A/en
Publication of WO2012009245A1 publication Critical patent/WO2012009245A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N5/00Details of television systems
    • H04N5/44Receiver circuitry for the reception of television signals according to analogue transmission standards
    • H04N5/445Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information
    • H04N5/45Picture in picture, e.g. displaying simultaneously another television channel in a region of the screen
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/234Processing of video elementary streams, e.g. splicing of video streams, manipulating MPEG-4 scene graphs
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/20Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
    • H04N21/23Processing of content or additional data; Elementary server operations; Server middleware
    • H04N21/236Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
    • H04N21/2365Multiplexing of several video streams
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/431Generation of visual interfaces for content selection or interaction; Content or additional data rendering
    • H04N21/4312Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
    • H04N21/4316Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations for displaying supplemental content in a region of the screen, e.g. an advertisement in a separate window
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/43Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
    • H04N21/434Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/80Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
    • H04N21/85Assembly of content; Generation of multimedia applications
    • H04N21/854Content authoring
    • H04N21/8543Content authoring using a description language, e.g. Multimedia and Hypermedia information coding Expert Group [MHEG], eXtensible Markup Language [XML]

Definitions

  • the present invention relates generally to picture-in-picture for multimedia applications such as transmission and rendering. More specifically, the invention relates to support of picture-in-picture functionality and improvement of the flexibility of picture-in- picture for multimedia applications.
  • Picture-in-picture allows viewers to view multiple separate video sources simultaneously.
  • some Blu-ray Disc titles include a picture-in-picture track that allows the viewer to see the director's comment on a film they are watching.
  • the picture-in-picture is implemented by generating a hard coded PIP video, i.e. by replacing the regions in the background video with a foreground video.
  • the hard coded PIP video is compressed and transmitted to the receiver.
  • viewers are not able to dynamically adjust the PIP, such as to enable/disable the PIP feature (unless a copy of the background video is sent separately), to change the position of the foreground video, etc.
  • Another traditional PIP application is to overlay two independent video streams at the player side, where video transport cannot provide any correlation information of the PIP video streams.
  • PIP media multiple video components can be correlated and form a set of media, i.e. PIP media.
  • the rendering of the PIP media can be dynamic, which means the position, scaling and alpha blending of the foreground videos can vary during playback, determined by either content creation or user interactions.
  • dynamic PIP enables the effect that the director points to different positions of the background video by e.g. moving the foreground video.
  • One deficiency of the current media file formats is that they cannot live or dynamically update the information of position, layer and scaling for the PIP stream in the system layer (i.e. the transport layer). Without the dynamic position and scaling information, it is not possible to reliably fit or overlay a video source on a display region that does not share the same resolution.
  • One possibility is to retrieve the information from the video decoder. Depending on the codec and the output of a particular encoder, such information may not exist or may not be reliable. It may also be difficult for a system to extract this information from the codec as well.
  • a system-level approach is a better approach to offering a consistent experience regardless of the underlying video codec used.
  • the present invention provides solutions to support picture-in-picture functionality and improve the flexibility of picture-in-picture for multimedia applications such as transmission and rendering.
  • This invention directs to methods and apparatuses for preparing a media stream containing a plurality of component streams for picture-in-picture application and for rendering such a media stream in a picture-in-picture mode.
  • a method for preparing a media stream containing a plurality of component streams for picture-in-picture applications Picture-in-picture information is generated based on the plurality of component streams contained in the media stream. Such picture-in-picture information is embedded into the media stream before sending to the receiver.
  • An apparatus for performing such a method is also provided.
  • the apparatus comprises a correlator, which generates picture-in-picture information from the plurality of component streams contained in the media stream, and an embedder for inserting the picture-in-picture information into the media stream.
  • a method for rendering a media stream containing a plurality of component streams in a picture-in-picture mode there is provided.
  • the picture-in-picture information related to the media stream is obtained.
  • a plurality of component streams is extracted from the media stream and is rendered in a picture-in-picture mode.
  • An apparatus for implementing the method is also provided.
  • Such an apparatus comprises an extractor and a processor.
  • the extractor extracts, from the media stream, picture-in-picture information, and a plurality of component streams according to the picture-in-picture information.
  • the extracted component streams are processed in the processor for rendering in a picture-in-picture mode.
  • Figure 1 illustrates a picture-in-picture display mode with one slave layer picture overlaying on a master layer picture.
  • Figure 2 shows the structure of an apparatus used to prepare, according to the present invention, a media stream which contains multiple component streams for picture-in-picture purpose.
  • Figure 3 shows one embodiment of the current invention to prepare a media stream for use in a picture-in-picture application.
  • Figure 4 shows the structure of a picture-in-picture rendering device used to read a media stream containing picture-in-picture information and render the stream in a picture-in-picture mode.
  • Figure 5 shows the process of reading a media stream containing picture-in- picture information and rendering the stream in a picture-in-picture mode according to the present invention.
  • a system level PIP solution is proposed by inserting the PIP information, such as the resolution/size, the location and the aspect ratio of each component stream in a PIP mode, to the bit stream at system level, so that the receiver can dynamically operate the PIP without accessing the information from the decoder.
  • Figure 1 illustrates the picture-in-picture display mode with two picture sources.
  • the background picture is labeled as a master layer 110 and the foreground picture is labeled as a slave layer 120.
  • the location of the slave layer 120 is defined with respect to the master layer 110 as a pair of offsets (a horizontal offset and a vertical offset).
  • the aspect ratio and the resolution of the slave layer 120 are also determined according to the desired display region in the master layer 110.
  • Other parameters, such as alpha blending of the master and slave layer pictures, can also be defined according to the application.
  • the present invention can support multiple slave layers. Each layer can be defined in a similar way.
  • the master layer and slave layers can be defined with respect to a fixed picture/video called anchor picture/video so that the location, scaling and other PIP information are aligned for all layers.
  • anchor picture/video a fixed picture/video
  • the content stream that consists of a master layer or a slave layer can be a video, an image, subtitles, or a base layer video, an enhancement layer video in SVC or one view of the video in MVC.
  • video component streams are used to describe the invention. The same principles apply to other component types.
  • Figure 2 shows an apparatus for preparing a media stream which contains multiple component streams for picture-in-picture purposes.
  • the media stream can be, for example, a program stream or a transport stream.
  • a stream identifier 210 identifies the component streams 215 contained in the media stream 205.
  • a correlator 220 correlates those identified component streams according to picture-in-picture relationships. Such PIP relationships are determined by the application.
  • a video of satellite images as a slave video may be, for example, associated/correlated with a video of a meteorologist commenting on the weather forecast as a master video.
  • the video of satellite images as a master video is associated with a video recording a local town hit by a tornado as a slave video for education purposes.
  • the correlator 220 also generates the PIP information 225 based on the information of the master and slave component videos.
  • the PIP information 225 includes, but is not limited to, Layer information that indicates the overlapping relationship among different layers, such as which component video is the master layer video, and which component video is the slave layer video. For applications that support multiple slave layers, the layer information also includes the overlapping relationship among the slave layers.
  • Position information that indicates the position of each layer when displayed in a picture-in-picture mode.
  • the position information can be calculated with respect to the master layer component video.
  • the position information is with respect to an anchor component video.
  • Size information that indicates the size or resolution of each layer when displayed a picture-in-picture mode.
  • the size or resolution of a slave layer video is determined by the size of the region in the master layer video that is intended for displaying the slave layer video.
  • Aspect ratio information that indicates the aspect ratio of each layer when displayed in a picture-in-picture mode.
  • the aspect ratio of a slave layer video is determined by the aspect ratio of the region in the master layer video that is intended for displaying the slave layer video.
  • the identified PIP information 225 may be inserted into the media stream by an embedder 240.
  • embedding is realized by feeding the PIP information 225 into a PIP information builder 230 to generate PIP information data units 235.
  • the information data units 235 carry the same information as the PIP information 225, but conform to the format of the component streams 215.
  • the embedder 240 inserts the format compliant PIP information data units 235 into the media stream along with the component stream 215 so that the PIP information can be associated with component streams to enable PIP at the receiver side.
  • the output media stream 245 is different from the input media stream 205 in that the output media stream 245 contains the PIP information inserted by the embedder 240.
  • Fig. 3 shows the process of preparing a media stream which contains multiple component streams for picture-in-picture applications.
  • each component stream contained in the media stream retrieved at step 310 is examined.
  • the PIP relationship of the current component stream with other component streams is identified in step 330.
  • the PIP information as mentioned above is generated.
  • PIP information is encapsulated into PIP information data units at step 340 which are compliant with the format of the media stream.
  • the PIP information data units are inserted into the media stream at step 350 to be sent to the receiver.
  • the PIP information for the component streams is stable for a certain period of time, such as within one scene of the video. That is, the value of position, size and other PIP information remain the same for all the frames within that scene.
  • the PIP information may need to be updated.
  • One example of such a change is that the topic of the director's commenting changes which requires moving the video of the director to a different position and with a different size and aspect ratio.
  • Another example could be that the source of the slave layer video switches to a different component stream.
  • Such changes of PIP information need to be updated and inserted into the media stream. If PIP information data units are used for carrying the PIP information, the PIP information data units should also be rebuilt based on the updated PIP information in compliant with the format of the media stream and embedded into the media stream.
  • the present invention targets on supporting the multiple component relationship. It can be used in many multimedia applications such as the HTTP streaming and enriching the functionalities of HTML5.
  • two embodiments of PIP information data unit design are disclosed for two streaming standards that shall support HTTP streaming: MPEG 2 Transport stream and ISO media format streaming. Since the current versions of both standards do not support system level PIP implementation, the proposed embodiments would be extensions to the current standards. It would be appreciated that those skilled in the art will be able to devise various versions of the PIP information data units that, although not explicitly described or shown herein, embody the present principles and are included within its spirit and scope.
  • pip_descriptor Listed in Table 1 is the proposed PIP information data unit for the MPEG-2 TS specification, referred to as a pip_descriptor.
  • the descriptor contains details of a video source's position, scaling and aspect ratio information.
  • the syntax of the pip_descritor is explained as follows:
  • descriptor_tag is an 8 bit field which identifies each descriptor in MPEG-2 standard. For more details about descriptors in MPEG-2, see Section 2.6 of Information technology - Generic coding of moving pictures and associated audio information: systems, ISO/IEC 13818-1:2007.
  • descriptor_length is an 8 bit field specifying the number of bytes of the descriptor immediately following the descriptor_length field.
  • layer_num This field indicates how a master layer component video and slave layer component videos overlap. A value of 0 is the lowest layer during overlay. In general, for one display, if there is any overlap between layer i content with layer j content and i ⁇ j, layer j content will be rendered for the overlapped region. Using the parameter of layer_num, the support for multiple-display devices can be easily implemented.
  • master_layer This field indicates which component video is the master layer of current component video layer. If the value of master layer is same as the value of layer number, it means the current component video is a master layer.
  • horizontal_offset - This field indicates the horizontal position of the top left pixel of a current component video display region with respect to the top left pixel of the master layer component video or display region as shown in Fig. 1. This value can also be measured with respect to an anchor component video.
  • vertical_offset - This field indicates the vertical position of the top left pixel of a current component video display region with respect to the top left pixel of the master layer component video display region as shown in Fig. 1. This value can also be measured with respect to an anchor component video,
  • This field is the horizontal size of a display region for a current component video. Video frames of the current component video are scaled to this size.
  • This field is the horizontal size of a display region for a current component video. Video frames of the current component video are scaled to this size.
  • aspect_ratio_info_present_flag If the value of this field equals to 1, it means that aspect_ratio_idc is present. If aspect_ratio_info_present_flag equals to 0, then aspect_ratio_idc is not present.
  • aspect_ratio_idc This field specifies the value of the sample aspect ratio of the luma samples in the component video. Table-2 shows the meaning of the code. When aspect_ratio_idc indicates Extended_SAR, the sample aspect ratio is represented by sar_width and sar_height. When the aspect_ratio_idc syntax element is not present, aspect_ratio_idc value shall be inferred to be equal to 0.
  • sar_width indicates the horizontal size of the sample aspect ratio (in arbitrary units).
  • sar_height indicates the vertical size of the sample aspect ratio (in the same arbitrary units as sar_width).
  • sar_width and sar_height shall be relatively prime or equal to 0.
  • aspect_ratio_idc is equal to 0 or sar_width is equal to 0 or sar_height is equal to 0 the sample aspect ratio shall be considered unspecified.
  • flexibility indicates the PIP behavior for a player.
  • it is a type of hard coded PIP and means player cannot alternated it or remove it which can be used in applications such as none removable Ads.
  • the flexibility field is set to 1
  • player can alternate the PIP according to an algorithm such as screen size and point-of-interest mask layer etc.
  • the flexibility field is set to 2
  • player allows user to change the PIP and even turn off PIP, which can enhance the interactivity experience and also reduce the control overhead.
  • the target_background_descriptor and pip_descriptors are used to place the video at a specific location in the display as shown in Figure 1.
  • the target_background_descriptor is explained in more detail in section 2.6.12 of Information technology - Generic coding of moving pictures and associated audio information: systems, ISO/IEC 13818-1:2007.
  • the pip_descriptor shall be included in the descriptor loop of the program map table/section (PMT) defined in Section 2.4.4.8 of Information technology - Generic coding of moving pictures and associated audio information: systems, ISO/IEC 13818-1:2007.
  • PMT program map table/section
  • the pip_descriptor can be dynamically updated along with the PMT information.
  • the live updating PIP information box is used during live streaming, wherein the PIP information box is included in the metadata portion of the component stream and can be updated constantly.
  • the offline updating PIP information box is used with offline applications such as DVD, whereby the PIP information is only sent once in the box and the information is inserted in the data sample portion of the stream. When the PIP information changes, the PIP information box is updated and inserted into the current data sample portion of the stream.
  • entry_count - is an integer that gives the number of entries in the following table. Each entry corresponds to a component stream in the media stream. pip_sample_number - gives the number of the samples for which that are PIP points in the stream. The details about the definition of a sample can be found at Section 3.1.10 of
  • PIP information is extracted from the media stream.
  • a plurality of component streams including a master layer component stream and one or more slave layer component streams, are determined and extracted from the media stream.
  • the component streams, including the master layer and slave layer component streams, are rendered according to the PIP information.
  • the component stream needs to be pre-processed before rendering.
  • Such a pre-processing relates to a scaling operation of the component streams.
  • Other pre-processing would be necessary depending on the component streams and the PIP information.
  • the rendering process will constantly monitor the PIP information, and make changes whenever PIP information has updates.
  • FIG. 4 shows the structure of a PIP rendering device used to read a media stream containing PIP information.
  • Extractor 410 extracts PIP information from the input media stream, and further extracts a master component stream and all slave component streams from the media stream according to the PIP information.
  • a processor 430 processes the extracted component streams according to the picture-in-picture information to generate output to a display 440.
  • extractor 410 extracts PIP information data units, such as pip_descriptors or PIP information boxes from the media stream and an analyzer 420 is employed to analyze the extracted PIP information data units.
  • One output from the analyzer 420 is the references to the master and slave layer component video streams. With those references, the extractor 410 is able to extract the necessary component streams for the processor 430.
  • Analyzer 420 further parses the PIP information data units to generate PIP information that is understandable to the processor 430.
  • the processor 430 can also accept outside control signals, e.g. from viewers, for controlling the PIP effect, such as enabling or disabling the PIP feature.
  • the control signal can contain the desired position or size information for the component streams. Such information overwrites the corresponding information provided in PIP information from the media stream.
  • the processor 430 may comprise a video decoder which decodes the extracted master and slave component streams. The decoding of master and slave layer component streams can be performed in parallel.
  • the processor 430 further performs such pre-processing by calling its pre-processor unit before generating the PIP video for displaying at display 440.
  • Display 440 may also include more than one display screen for multi-display applications as discussed before.
  • FIG. 5 shows a process for reading a media stream containing PIP information and rendering the stream in a PIP mode according to the present invention.
  • PIP information data units such as pip_descriptors or PIP information boxes are extracted in step 510.
  • PIP information is further identified in step 520 by analyzing the extracted PIP information data units, such as which streams are the master and slave layer component streams. With the PIP information, the process is able to extract the required component streams from the media stream in step 530.
  • a decision step 550 determines whether the component stream requires pre-processing. If pre-processing is needed, the component stream content is processed according to the PIP information in step 560, such as resizing the content to a desired size and an aspect ratio. If pre-processing is determined to be unnecessary, the process moves directly to step 570 for rending.
  • the loop for each of the component streams may be realized in parallel using multi-core processors or through multi- thread implementations.

Abstract

A method and a device for preparing a media stream containing more than one component stream for picture-in-picture applications are described along with the corresponding method and device for rendering the media stream in a picture-in-picture mode. The invention allows for live and dynamic picture-in-picture rendering of the component streams contained in the media stream. Extensions to the media formats that shall support various multimedia applications, such as the MPEG-2 Transport Stream and ISO media format, are proposed to enable such a dynamic picture-in-picture functionality.

Description

METHOD OF PICTURE-IN-PICTURE FOR MULTIMEDIA APPLICATIONS
CROSS-REFERENCE TO RELATED APPLICATIONS
This application claims the benefit of priority from U.S. Provisional Patent Application Serial No. 61/363,697, entitled "Method of Picture-in-Picture in the MPEG-2 TS for the HTTP Streaming," filed on July 13, 2010, and U.S. Provisional Patent Application Serial No. 61/363,699, entitled "Method of Picture-in-Picture in the ISO Base Media File Format for th HTTP Streaming," filed on July 13, 2010. The teachings of the above- identified provisional patent application are expressly incorporated herein by reference.
TECHNICAL FIELD
The present invention relates generally to picture-in-picture for multimedia applications such as transmission and rendering. More specifically, the invention relates to support of picture-in-picture functionality and improvement of the flexibility of picture-in- picture for multimedia applications.
BACKGROUND OF THE INVENTION
Picture-in-picture (PIP) allows viewers to view multiple separate video sources simultaneously. For example, some Blu-ray Disc titles include a picture-in-picture track that allows the viewer to see the director's comment on a film they are watching. Traditionally, such as in Blu-ray Disc applications, the picture-in-picture is implemented by generating a hard coded PIP video, i.e. by replacing the regions in the background video with a foreground video. The hard coded PIP video is compressed and transmitted to the receiver. As a result, viewers are not able to dynamically adjust the PIP, such as to enable/disable the PIP feature (unless a copy of the background video is sent separately), to change the position of the foreground video, etc. Another traditional PIP application is to overlay two independent video streams at the player side, where video transport cannot provide any correlation information of the PIP video streams.
With the development of interactive media technology, multiple video components can be correlated and form a set of media, i.e. PIP media. The rendering of the PIP media can be dynamic, which means the position, scaling and alpha blending of the foreground videos can vary during playback, determined by either content creation or user interactions. In the previous example wherein a foreground video shows a director commenting on the background video, dynamic PIP enables the effect that the director points to different positions of the background video by e.g. moving the foreground video.
One deficiency of the current media file formats, such as MPEG2 transport stream and ISO media file format, is that they cannot live or dynamically update the information of position, layer and scaling for the PIP stream in the system layer (i.e. the transport layer). Without the dynamic position and scaling information, it is not possible to reliably fit or overlay a video source on a display region that does not share the same resolution. One possibility is to retrieve the information from the video decoder. Depending on the codec and the output of a particular encoder, such information may not exist or may not be reliable. It may also be difficult for a system to extract this information from the codec as well. A system-level approach is a better approach to offering a consistent experience regardless of the underlying video codec used.
The present invention provides solutions to support picture-in-picture functionality and improve the flexibility of picture-in-picture for multimedia applications such as transmission and rendering.
SUMMARY OF THE INVENTION
This invention directs to methods and apparatuses for preparing a media stream containing a plurality of component streams for picture-in-picture application and for rendering such a media stream in a picture-in-picture mode.
According to an aspect of the present invention, there is provided a method for preparing a media stream containing a plurality of component streams for picture-in-picture applications. Picture-in-picture information is generated based on the plurality of component streams contained in the media stream. Such picture-in-picture information is embedded into the media stream before sending to the receiver. An apparatus for performing such a method is also provided. The apparatus comprises a correlator, which generates picture-in-picture information from the plurality of component streams contained in the media stream, and an embedder for inserting the picture-in-picture information into the media stream. According to another aspect of the present invention, there is provided a method for rendering a media stream containing a plurality of component streams in a picture-in-picture mode. The picture-in-picture information related to the media stream is obtained. According to such picture-in-picture information, a plurality of component streams is extracted from the media stream and is rendered in a picture-in-picture mode. An apparatus for implementing the method is also provided. Such an apparatus comprises an extractor and a processor. The extractor extracts, from the media stream, picture-in-picture information, and a plurality of component streams according to the picture-in-picture information. The extracted component streams are processed in the processor for rendering in a picture-in-picture mode.
BRIEF DESCRIPTION OF THE DRAWINGS
The above features of the present invention will become more apparent by describing in detail exemplary embodiments thereof with reference to the attached drawings in which:
Figure 1 illustrates a picture-in-picture display mode with one slave layer picture overlaying on a master layer picture.
Figure 2 shows the structure of an apparatus used to prepare, according to the present invention, a media stream which contains multiple component streams for picture-in-picture purpose.
Figure 3 shows one embodiment of the current invention to prepare a media stream for use in a picture-in-picture application.
Figure 4 shows the structure of a picture-in-picture rendering device used to read a media stream containing picture-in-picture information and render the stream in a picture-in-picture mode.
Figure 5 shows the process of reading a media stream containing picture-in- picture information and rendering the stream in a picture-in-picture mode according to the present invention.
DETAILED DESCRIPTION
In the present invention, a system level PIP solution is proposed by inserting the PIP information, such as the resolution/size, the location and the aspect ratio of each component stream in a PIP mode, to the bit stream at system level, so that the receiver can dynamically operate the PIP without accessing the information from the decoder.
Figure 1 illustrates the picture-in-picture display mode with two picture sources. The background picture is labeled as a master layer 110 and the foreground picture is labeled as a slave layer 120. In one embodiment, the location of the slave layer 120 is defined with respect to the master layer 110 as a pair of offsets (a horizontal offset and a vertical offset). The aspect ratio and the resolution of the slave layer 120 are also determined according to the desired display region in the master layer 110. Other parameters, such as alpha blending of the master and slave layer pictures, can also be defined according to the application. The present invention can support multiple slave layers. Each layer can be defined in a similar way. In a different embodiment, the master layer and slave layers can be defined with respect to a fixed picture/video called anchor picture/video so that the location, scaling and other PIP information are aligned for all layers. In the present invention, we refer to the content stream that consists of a master layer or a slave layer as a component stream. Note that a component stream can be a video, an image, subtitles, or a base layer video, an enhancement layer video in SVC or one view of the video in MVC. In the following, video component streams are used to describe the invention. The same principles apply to other component types.
Figure 2 shows an apparatus for preparing a media stream which contains multiple component streams for picture-in-picture purposes. The media stream can be, for example, a program stream or a transport stream. A stream identifier 210 identifies the component streams 215 contained in the media stream 205. A correlator 220 correlates those identified component streams according to picture-in-picture relationships. Such PIP relationships are determined by the application. In one application, a video of satellite images as a slave video may be, for example, associated/correlated with a video of a meteorologist commenting on the weather forecast as a master video. In a different application, the video of satellite images as a master video is associated with a video recording a local town hit by a tornado as a slave video for education purposes. The correlator 220 also generates the PIP information 225 based on the information of the master and slave component videos. The PIP information 225 includes, but is not limited to, Layer information that indicates the overlapping relationship among different layers, such as which component video is the master layer video, and which component video is the slave layer video. For applications that support multiple slave layers, the layer information also includes the overlapping relationship among the slave layers.
Position information that indicates the position of each layer when displayed in a picture-in-picture mode. In one embodiment, the position information can be calculated with respect to the master layer component video. In a different embodiment, the position information is with respect to an anchor component video.
Size information that indicates the size or resolution of each layer when displayed a picture-in-picture mode. In one instance, the size or resolution of a slave layer video is determined by the size of the region in the master layer video that is intended for displaying the slave layer video.
Aspect ratio information that indicates the aspect ratio of each layer when displayed in a picture-in-picture mode. Similarly, in one instance, the aspect ratio of a slave layer video is determined by the aspect ratio of the region in the master layer video that is intended for displaying the slave layer video.
Other information, such as alpha blending of the layers, can also be included.
The identified PIP information 225 may be inserted into the media stream by an embedder 240. In one embodiment, such embedding is realized by feeding the PIP information 225 into a PIP information builder 230 to generate PIP information data units 235. The information data units 235 carry the same information as the PIP information 225, but conform to the format of the component streams 215. The embedder 240 inserts the format compliant PIP information data units 235 into the media stream along with the component stream 215 so that the PIP information can be associated with component streams to enable PIP at the receiver side. The output media stream 245 is different from the input media stream 205 in that the output media stream 245 contains the PIP information inserted by the embedder 240.
Fig. 3 shows the process of preparing a media stream which contains multiple component streams for picture-in-picture applications. At step 320, each component stream contained in the media stream retrieved at step 310 is examined. First, the PIP relationship of the current component stream with other component streams is identified in step 330. Meanwhile, the PIP information as mentioned above is generated. Such PIP information is encapsulated into PIP information data units at step 340 which are compliant with the format of the media stream. The PIP information data units are inserted into the media stream at step 350 to be sent to the receiver.
In general, the PIP information for the component streams is stable for a certain period of time, such as within one scene of the video. That is, the value of position, size and other PIP information remain the same for all the frames within that scene. However, as the content of the component stream for either the master layer or the slave layer changes, the PIP information may need to be updated. One example of such a change is that the topic of the director's commenting changes which requires moving the video of the director to a different position and with a different size and aspect ratio. Another example could be that the source of the slave layer video switches to a different component stream. Such changes of PIP information need to be updated and inserted into the media stream. If PIP information data units are used for carrying the PIP information, the PIP information data units should also be rebuilt based on the updated PIP information in compliant with the format of the media stream and embedded into the media stream.
The present invention targets on supporting the multiple component relationship. It can be used in many multimedia applications such as the HTTP streaming and enriching the functionalities of HTML5. In the following, two embodiments of PIP information data unit design are disclosed for two streaming standards that shall support HTTP streaming: MPEG 2 Transport stream and ISO media format streaming. Since the current versions of both standards do not support system level PIP implementation, the proposed embodiments would be extensions to the current standards. It would be appreciated that those skilled in the art will be able to devise various versions of the PIP information data units that, although not explicitly described or shown herein, embody the present principles and are included within its spirit and scope.
Listed in Table 1 is the proposed PIP information data unit for the MPEG-2 TS specification, referred to as a pip_descriptor. The descriptor contains details of a video source's position, scaling and aspect ratio information. The syntax of the pip_descritor is explained as follows:
Table-1 - PiP descriptor
Figure imgf000008_0001
descriptor_tag - The descriptor_tag is an 8 bit field which identifies each descriptor in MPEG-2 standard. For more details about descriptors in MPEG-2, see Section 2.6 of Information technology - Generic coding of moving pictures and associated audio information: systems, ISO/IEC 13818-1:2007.
descriptor_length - The descriptor_length is an 8 bit field specifying the number of bytes of the descriptor immediately following the descriptor_length field. layer_num - This field indicates how a master layer component video and slave layer component videos overlap. A value of 0 is the lowest layer during overlay. In general, for one display, if there is any overlap between layer i content with layer j content and i < j, layer j content will be rendered for the overlapped region. Using the parameter of layer_num, the support for multiple-display devices can be easily implemented. In a system with multiple displays/screens, such as using a TV as a major display screen and an iPad as a secondary display screen, the application can assign a higher probability for the content in the component stream with higher layer_num value to be rendered on a secondary display screen. master_layer— This field indicates which component video is the master layer of current component video layer. If the value of master layer is same as the value of layer number, it means the current component video is a master layer.
horizontal_offset - This field indicates the horizontal position of the top left pixel of a current component video display region with respect to the top left pixel of the master layer component video or display region as shown in Fig. 1. This value can also be measured with respect to an anchor component video.
vertical_offset - This field indicates the vertical position of the top left pixel of a current component video display region with respect to the top left pixel of the master layer component video display region as shown in Fig. 1. This value can also be measured with respect to an anchor component video,
width - This field is the horizontal size of a display region for a current component video. Video frames of the current component video are scaled to this size.
height - This field is the horizontal size of a display region for a current component video. Video frames of the current component video are scaled to this size.
aspect_ratio_info_present_flag - If the value of this field equals to 1, it means that aspect_ratio_idc is present. If aspect_ratio_info_present_flag equals to 0, then aspect_ratio_idc is not present.
aspect_ratio_idc - This field specifies the value of the sample aspect ratio of the luma samples in the component video. Table-2 shows the meaning of the code. When aspect_ratio_idc indicates Extended_SAR, the sample aspect ratio is represented by sar_width and sar_height. When the aspect_ratio_idc syntax element is not present, aspect_ratio_idc value shall be inferred to be equal to 0.
Table-2— Meaning of sample aspect ratio indicator
Figure imgf000010_0001
sar_width indicates the horizontal size of the sample aspect ratio (in arbitrary units).
sar_height indicates the vertical size of the sample aspect ratio (in the same arbitrary units as sar_width).
sar_width and sar_height shall be relatively prime or equal to 0. When aspect_ratio_idc is equal to 0 or sar_width is equal to 0 or sar_height is equal to 0, the sample aspect ratio shall be considered unspecified.
flexibility indicates the PIP behavior for a player. When it is set to 0, it is a type of hard coded PIP and means player cannot alternated it or remove it which can be used in applications such as none removable Ads. When the flexibility field is set to 1, player can alternate the PIP according to an algorithm such as screen size and point-of-interest mask layer etc. When the flexibility field is set to 2, player allows user to change the PIP and even turn off PIP, which can enhance the interactivity experience and also reduce the control overhead. In a picture-in-picture mode, the target_background_descriptor and pip_descriptors are used to place the video at a specific location in the display as shown in Figure 1. The target_background_descriptor is explained in more detail in section 2.6.12 of Information technology - Generic coding of moving pictures and associated audio information: systems, ISO/IEC 13818-1:2007. The pip_descriptor shall be included in the descriptor loop of the program map table/section (PMT) defined in Section 2.4.4.8 of Information technology - Generic coding of moving pictures and associated audio information: systems, ISO/IEC 13818-1:2007. When the updating of PIP information is necessary, the pip_descriptor can be dynamically updated along with the PMT information.
For ISO media file format, two PIP information boxes are proposed, one for live updating and one for offline updating. The live updating PIP information box is used during live streaming, wherein the PIP information box is included in the metadata portion of the component stream and can be updated constantly. The offline updating PIP information box is used with offline applications such as DVD, whereby the PIP information is only sent once in the box and the information is inserted in the data sample portion of the stream. When the PIP information changes, the PIP information box is updated and inserted into the current data sample portion of the stream.
1. PIP information box for the live updating
Definition:
Box Type: 'pipl'
Container: Track Fragment Box ('traf') or Track Box ('trak')
Mandatory: No
Quantity: Zero or one
Syntax:
aligned(8) class PiPInfoBox extends Box ('pipl') {
template int( 16) layer_num = 0;
template int(16) master_layer = 0;
unsigned int(32) horizontal_offset; unsigned int(32) vertical_offset;
unsigned int(32) width;
unsigned int(32) height;
}
Semantics
Each entry is the same as explained before. . PIP information box for the offline updating
Definition:
Box Type: 'pips'
Container: Sample Table Box ('stbl')
Mandatory: No
Quantity: Zero or one Syntax:
aligned(8) class PiPInfoBox extends Box ('pips') {
unsigned int(32) entry_count;
int i;
for (i=0; i < entry_count; i++) {
unsigned int(32) pip_sample_number;
template int(16) layer_num = 0;
template int(16) master_layer = 0;
unsigned int(32) horizontal_offset;
unsigned int(32) vertical_offset;
unsigned int(32) width;
unsigned int(32) height;
}
}
Semantics
entry_count - is an integer that gives the number of entries in the following table. Each entry corresponds to a component stream in the media stream. pip_sample_number - gives the number of the samples for which that are PIP points in the stream. The details about the definition of a sample can be found at Section 3.1.10 of
Information Technology - Coding of Audio-Visual Objects - Part 12: ISO base media file format.
At the receiver side, in order to render the picture-in-picture for a media stream that contains multiple component streams, PIP information is extracted from the media stream. According to the PIP information, a plurality of component streams, including a master layer component stream and one or more slave layer component streams, are determined and extracted from the media stream. The component streams, including the master layer and slave layer component streams, are rendered according to the PIP information. In some cases, for example when the video content in the component stream has different size/resolution and aspect ratio from the required size and aspect ratio for PIP, the component stream needs to be pre-processed before rendering. Such a pre-processing relates to a scaling operation of the component streams. Other pre-processing would be necessary depending on the component streams and the PIP information. The rendering process will constantly monitor the PIP information, and make changes whenever PIP information has updates.
Figure 4 shows the structure of a PIP rendering device used to read a media stream containing PIP information. Extractor 410 extracts PIP information from the input media stream, and further extracts a master component stream and all slave component streams from the media stream according to the PIP information. A processor 430 processes the extracted component streams according to the picture-in-picture information to generate output to a display 440. In one embodiment, extractor 410 extracts PIP information data units, such as pip_descriptors or PIP information boxes from the media stream and an analyzer 420 is employed to analyze the extracted PIP information data units. One output from the analyzer 420 is the references to the master and slave layer component video streams. With those references, the extractor 410 is able to extract the necessary component streams for the processor 430. Analyzer 420 further parses the PIP information data units to generate PIP information that is understandable to the processor 430. The processor 430 can also accept outside control signals, e.g. from viewers, for controlling the PIP effect, such as enabling or disabling the PIP feature. The control signal can contain the desired position or size information for the component streams. Such information overwrites the corresponding information provided in PIP information from the media stream. The processor 430 may comprise a video decoder which decodes the extracted master and slave component streams. The decoding of master and slave layer component streams can be performed in parallel. If the pre-processing as mentioned before is necessary for the component streams, the processor 430 further performs such pre-processing by calling its pre-processor unit before generating the PIP video for displaying at display 440. Display 440 may also include more than one display screen for multi-display applications as discussed before.
Figure 5 shows a process for reading a media stream containing PIP information and rendering the stream in a PIP mode according to the present invention. PIP information data units such as pip_descriptors or PIP information boxes are extracted in step 510. PIP information is further identified in step 520 by analyzing the extracted PIP information data units, such as which streams are the master and slave layer component streams. With the PIP information, the process is able to extract the required component streams from the media stream in step 530. For each of the component streams, a decision step 550 determines whether the component stream requires pre-processing. If pre-processing is needed, the component stream content is processed according to the PIP information in step 560, such as resizing the content to a desired size and an aspect ratio. If pre-processing is determined to be unnecessary, the process moves directly to step 570 for rending. The loop for each of the component streams may be realized in parallel using multi-core processors or through multi- thread implementations.
Although preferred embodiments of the present invention have been described in detail herein, it is to be understood that this invention is not limited to these embodiments, and that other modifications and variations may be effected by one skilled in the art without departing from the scope of the invention as defined by the appended claims.

Claims

1. A method for preparing a media stream containing a plurality of component streams for picture-in-picture applications, the method comprising:
generating picture-in-picture information from the plurality of component streams contained in the media stream; and
inserting the picture-in-picture information into the media stream.
2. The method of claim 1, further comprising
updating the picture-in-picture information according to changes in the plurality of component streams.
3. The method of claim 2, wherein the media stream is one of a transport stream and a program stream.
4. The method of claim 2, wherein the inserting step comprises
building picture-in-picture data units according to the generated picture-in- picture information; and
inserting the picture-in-picture data units into the media stream.
5. The method of claim 4, wherein the picture-in-picture data units comprise picture- in-picture descriptors, and the inserting step further comprises
including the picture-in-picture descriptors into a program map table of the media stream.
6. The method of claim 4, wherein the picture-in-picture data units comprise picture- in-picture information boxes, and the inserting step further comprises
inserting the picture-in-picture information boxes into the plurality of component streams.
7. The method of claim 6, wherein a picture-in-picture information box is one of a picture-in-picture information box for live updating and a picture-in-picture information box for offline updating.
8. The method of claim 2, wherein the picture-in-picture information comprises information of layer, position, size, and aspect ratio of the component streams in a picture-in-picture mode.
9. The method of claim 8, wherein the picture-in-picture information for each component stream is defined with respect to one of a master layer component stream and an anchor component stream.
10. An apparatus for preparing a media stream containing a plurality of component streams for picture-in-picture applications, the apparatus comprising:
a correlator for generating picture-in-picture information from the plurality of component streams contained in the media stream; and
an embedder for inserting the picture-in-picture information into the media stream.
11. The apparatus of claim 10, wherein the media stream is one of a transport stream and a program stream.
12. The apparatus of claim 11, further comprising
a picture-in-picture information builder for generating picture-in-picture information data units from the identified picture-in-picture information, the picture- in-picture information data units being fed into the embedder for inserting into the media stream.
13. The apparatus of claim 12, wherein the picture-in-picture data units comprise picture-in-picture descriptors, and the embedder further performs steps of
including the picture-in-picture descriptors into a program map table of the media stream.
14. The apparatus of claim 12, wherein the picture-in-picture data units comprise picture-in-picture information boxes, and the embedder further performs the step of inserting the picture-in-picture information boxes into the plurality of component streams.
15. The apparatus of claim 14, wherein a picture-in-picture information box is one of a picture-in-picture information box for live updating and a picture-in-picture information box for offline updating.
16. A method for rendering a media stream containing a plurality of component streams in a picture-in-picture mode, the method comprising:
obtaining picture-in-picture information related to the media stream;
extracting a plurality of component streams from the media stream according to the obtained picture-in-picture information; and
rendering the plurality of component streams according to the obtained picture-in-picture information.
17. The method of claim 16, wherein the media stream is one of a transport stream and a program stream.
18. The method of claim 16, wherein the plurality of component streams comprises a master layer component stream and at least one slave layer component stream.
19. The method of claim 16, further comprising
if necessary, pre-processing at least one of the extracted plurality of component streams according to the picture-in-picture information prior to rendering.
20. The method of claim 19, wherein the obtaining step comprises
extracting picture-in-picture information data units from the media stream; and
analyzing the picture-in-picture information data units to identify the picture- in-picture information.
21. The method of claim 20, wherein a picture-in-picture data unit comprises at least one of a picture-in-picture descriptor, a picture-in-picture information box for live updating and a picture-in-picture information box for offline updating.
22. The method of claim 16, wherein the picture-in-picture information comprises information of layer, position, size, aspect ratio of the component streams in the picture-in-picture mode.
23. The method of claim 22, wherein the picture-in-picture information for each component stream is defined with respect to one of a master component stream and an anchor component stream.
24. The method of claim 22, wherein the pre-processing step comprises scaling at least one of the extracted component streams to a size specified in the picture-in- picture information.
25. An apparatus for rendering a media stream containing a plurality of component streams in a picture-in-picture mode, the apparatus comprising
an extractor for extracting, from the media stream, picture-in-picture information, and for extracting a plurality of component streams according to the picture-in-picture information;
a processor for processing the extracted plurality of component streams according to the picture-in-picture information.
26. The apparatus of claim 25, wherein the extractor extracts picture-in-picture information data units from the media stream, the apparatus further comprising
an analyzer for analyzing the picture-in-picture information data units to get picture-in-picture information, and for providing references to the plurality of component stream to the extractor.
27. The apparatus of claim 26, further comprising
a display for displaying the media stream in the picture-in-picture mode.
28. The method of claim 25, wherein the media stream is one of a transport stream and a program stream.
29. The method of claim 25, wherein the plurality of component streams comprises a master layer component stream and at least one slave layer component stream.
30. The apparatus of claim 25, wherein a picture-in-picture data unit comprises at least one of a picture-in-picture descriptor, a picture-in-picture information box for live updating and a picture-in-picture information box for offline updating.
31. The apparatus of claim 25, wherein the picture-in-picture information comprises at least information of layer, position, size, and aspect ratio of the component streams in the picture-in-picture mode.
32. The apparatus of claim 31, wherein the picture-in-picture information for each component stream is defined with respect to one of a master component stream and an anchor component stream.
33. The apparatus of claim 25, wherein the processor comprises a pre-processing unit for scaling the at least one component stream to a size specified in the picture-in- picture information.
34. The apparatus of claim 25, wherein the processor accepts a control signal and processes at least one of the extracted component streams according to the control signal.
35. The apparatus of claim 25, wherein the display comprises multiple screens for displaying the media stream in the picture-in-picture mode.
PCT/US2011/043494 2010-07-13 2011-07-11 Method of picture-in-picture for multimedia applications WO2012009245A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US13/809,812 US8767124B2 (en) 2010-07-13 2011-07-11 Method of picture-in-picture for multimedia applications
KR1020137003063A KR20130127423A (en) 2010-07-13 2011-07-11 Method of picture-in-picture for multimedia applications
JP2013519735A JP2013538482A (en) 2010-07-13 2011-07-11 Picture-in-picture method for multimedia applications
EP11733971.3A EP2594070A1 (en) 2010-07-13 2011-07-11 Method of picture-in-picture for multimedia applications
CN2011800347149A CN103004221A (en) 2010-07-13 2011-07-11 Method of picture-in-picture for multimedia applications

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US36369710P 2010-07-13 2010-07-13
US36369910P 2010-07-13 2010-07-13
US61/363,697 2010-07-13
US61/363,699 2010-07-13

Publications (1)

Publication Number Publication Date
WO2012009245A1 true WO2012009245A1 (en) 2012-01-19

Family

ID=44509614

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2011/043494 WO2012009245A1 (en) 2010-07-13 2011-07-11 Method of picture-in-picture for multimedia applications

Country Status (6)

Country Link
US (1) US8767124B2 (en)
EP (1) EP2594070A1 (en)
JP (1) JP2013538482A (en)
KR (1) KR20130127423A (en)
CN (1) CN103004221A (en)
WO (1) WO2012009245A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10643304B2 (en) 2016-11-03 2020-05-05 Hanwha Techwin Co., Ltd. Image providing apparatus and method

Families Citing this family (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10021431B2 (en) * 2013-01-04 2018-07-10 Omnivision Technologies, Inc. Mobile computing device having video-in-video real-time broadcasting capability
CN106031178A (en) * 2014-01-21 2016-10-12 Lg电子株式会社 Broadcast transmission device and operating method thereof, and broadcast reception device and operating method thereof
US9661339B2 (en) * 2014-01-21 2017-05-23 Intel Corporation Multi-core architecture for low latency video decoder
WO2015130022A1 (en) * 2014-02-26 2015-09-03 엘지전자 주식회사 Digital device and data processing method by digital device
KR102225946B1 (en) 2014-02-26 2021-03-10 엘지전자 주식회사 Digital device and method of processing application thereof
US10455257B1 (en) * 2015-09-24 2019-10-22 Tribune Broadcasting Company, Llc System and corresponding method for facilitating application of a digital video-effect to a temporal portion of a video segment
US9883212B2 (en) * 2015-09-24 2018-01-30 Tribune Broadcasting Company, Llc Video-broadcast system with DVE-related alert feature
US10979133B2 (en) 2017-06-12 2021-04-13 Ast & Science, Llc System and method for high throughput fractionated satellites (HTFS) for direct connectivity to and from end user devices and terminals using flight formations of small or very small satellites
US9973266B1 (en) * 2017-06-12 2018-05-15 Ast & Science, Llc System and method for high throughput fractionated satellites (HTFS) for direct connectivity to and from end user devices and terminals using flight formations of small or very small satellites
CN109819178B (en) * 2017-11-21 2022-07-08 虹软科技股份有限公司 Method and device for frame processing
TWI687863B (en) * 2018-07-04 2020-03-11 圓展科技股份有限公司 Display device and method for operating the same
CA3140057A1 (en) 2019-05-15 2020-11-19 Ast & Science, Llc Low earth orbit mechanical deployable structure
US11582463B2 (en) 2019-12-30 2023-02-14 Tencent America LLC Method for alignment across layers in coded video stream
EP4297418A1 (en) * 2022-06-24 2023-12-27 Beijing Xiaomi Mobile Software Co., Ltd. Signaling encapsulated data representing primary video sequence and associated auxiliary video sequence

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1008933A2 (en) * 1998-11-30 2000-06-14 Sun Microsystems, Inc. TV pip using java api classes and java implementation classes
EP1487212A1 (en) * 2002-03-15 2004-12-15 Fujitsu Limited Video information delivery apparatus and reception apparatus
WO2007013769A1 (en) * 2005-07-29 2007-02-01 Lg Electronics Inc. Recording medium, method and apparatus for reproducing data, and method and apparatus for recording data
US20070039032A1 (en) * 2005-08-11 2007-02-15 Echostar Technologies Corporation Method and system for toasted video distribution
WO2009095080A1 (en) * 2008-01-31 2009-08-06 Telefonaktiebolaget Lm Ericsson (Publ) Method and apparatus for obtaining media over a communications network
WO2009126686A1 (en) * 2008-04-09 2009-10-15 Harris Corporation Video multiviewer with scaling
EP2144248A2 (en) * 2005-08-09 2010-01-13 Panasonic Corporation Recording medium, playback apparatus, method and program

Family Cites Families (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CA2187796C (en) * 1994-04-15 2010-06-29 Jurgen Fritz Rosengren Arrangement and method for transmitting and receiving video signals
US6741617B2 (en) * 1995-04-14 2004-05-25 Koninklijke Philips Electronics N.V. Arrangement for decoding digital video signals
KR101049754B1 (en) * 2002-09-26 2011-07-19 코닌클리케 필립스 일렉트로닉스 엔.브이. Digital Information Signal Receiver
KR100689507B1 (en) * 2005-01-07 2007-03-02 삼성전자주식회사 Apparatus and method for adaptation of multimedia signal performing picture in picture function
CN102752639B (en) * 2005-07-18 2015-08-05 汤姆森许可贸易公司 Metadata is used to process the method and apparatus of multiple video flowing
JP4408845B2 (en) * 2005-07-27 2010-02-03 シャープ株式会社 Video composition apparatus and program
KR20070014968A (en) * 2005-07-29 2007-02-01 엘지전자 주식회사 Method and apparatus for reproducing data, recording medium and method and eapparatus for recording data
CN101283410A (en) * 2005-08-22 2008-10-08 Lg电子株式会社 Recording medium, method and apparatus for reproducing data and method and apparatus for recording data
CN101742173B (en) * 2008-11-12 2012-05-16 中兴通讯股份有限公司 Method for supporting watching two-path video and network interaction television system

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1008933A2 (en) * 1998-11-30 2000-06-14 Sun Microsystems, Inc. TV pip using java api classes and java implementation classes
EP1487212A1 (en) * 2002-03-15 2004-12-15 Fujitsu Limited Video information delivery apparatus and reception apparatus
WO2007013769A1 (en) * 2005-07-29 2007-02-01 Lg Electronics Inc. Recording medium, method and apparatus for reproducing data, and method and apparatus for recording data
EP2144248A2 (en) * 2005-08-09 2010-01-13 Panasonic Corporation Recording medium, playback apparatus, method and program
US20070039032A1 (en) * 2005-08-11 2007-02-15 Echostar Technologies Corporation Method and system for toasted video distribution
WO2009095080A1 (en) * 2008-01-31 2009-08-06 Telefonaktiebolaget Lm Ericsson (Publ) Method and apparatus for obtaining media over a communications network
WO2009126686A1 (en) * 2008-04-09 2009-10-15 Harris Corporation Video multiviewer with scaling

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10643304B2 (en) 2016-11-03 2020-05-05 Hanwha Techwin Co., Ltd. Image providing apparatus and method

Also Published As

Publication number Publication date
US8767124B2 (en) 2014-07-01
US20130113996A1 (en) 2013-05-09
CN103004221A (en) 2013-03-27
EP2594070A1 (en) 2013-05-22
JP2013538482A (en) 2013-10-10
KR20130127423A (en) 2013-11-22

Similar Documents

Publication Publication Date Title
US8767124B2 (en) Method of picture-in-picture for multimedia applications
KR101819736B1 (en) Auxiliary data in 3d video broadcast
US9271048B2 (en) Systems and methods for immersive viewing experience
US8773584B2 (en) Playback apparatus, playback method, integrated circuit, broadcast system, and broadcast method using a broadcast video and additional video
KR102528922B1 (en) A system for distributing metadata embedded in video
EP1547372B1 (en) Apparatus for receiving a digital information signal
US20160337706A1 (en) Method and apparatus for transreceiving broadcast signal for panorama service
US10547844B2 (en) Broadcasting signal transmission and reception method and device
US20160261927A1 (en) Method and System for Providing and Displaying Optional Overlays
US20170374313A1 (en) Method and device for transmitting and receiving broadcast signal
CA2838906C (en) Systems and methods for processing timed text in video programming
RU2633385C2 (en) Transmission device, transmission method, reception device, reception method and reception display method
US20160057488A1 (en) Method and System for Providing and Displaying Optional Overlays
US11936936B2 (en) Method and system for providing and displaying optional overlays
JP5957770B2 (en) Video processing apparatus, method, program, recording medium, and integrated circuit
EP2408211A1 (en) Auxiliary data in 3D video broadcast
US20160345078A1 (en) Carrier-based active text enhancement
EP3528505A1 (en) Apparatus and method for operating a media device to select from plural instances of content for play back
KR20120076625A (en) Method and apparatus for providing 3d contents
JP2005151101A (en) Information display device

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 11733971

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2013519735

Country of ref document: JP

Kind code of ref document: A

WWE Wipo information: entry into national phase

Ref document number: 13809812

Country of ref document: US

NENP Non-entry into the national phase

Ref country code: DE

WWE Wipo information: entry into national phase

Ref document number: 2011733971

Country of ref document: EP

ENP Entry into the national phase

Ref document number: 20137003063

Country of ref document: KR

Kind code of ref document: A