WO2012057164A1 - 符号化方法、表示装置、復号方法 - Google Patents
符号化方法、表示装置、復号方法 Download PDFInfo
- Publication number
- WO2012057164A1 WO2012057164A1 PCT/JP2011/074586 JP2011074586W WO2012057164A1 WO 2012057164 A1 WO2012057164 A1 WO 2012057164A1 JP 2011074586 W JP2011074586 W JP 2011074586W WO 2012057164 A1 WO2012057164 A1 WO 2012057164A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- video
- stream
- playback
- video stream
- information
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/2362—Generation or processing of Service Information [SI]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/178—Metadata, e.g. disparity information
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/46—Embedding additional information in the video signal during the compression process
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/2365—Multiplexing of several video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/434—Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
- H04N21/4345—Extraction or processing of SI, e.g. extracting service information from an MPEG stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/816—Monomedia components thereof involving special video data, e.g 3D video
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/84—Generation or processing of descriptive data, e.g. content descriptors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/83—Generation or processing of protective or descriptive data associated with content; Content structuring
- H04N21/845—Structuring of content, e.g. decomposing content into time segments
- H04N21/8451—Structuring of content, e.g. decomposing content into time segments using Advanced Video Coding [AVC]
Definitions
- the present invention relates to an encoding method technique, and particularly to an encoding method technique for a transport stream related to 3D video.
- 3D programs are broadcast by side-by-side video streams that enable 3D playback to be multiplexed into 1TS (one transport stream), and broadcast stations can use 1TS on TV display devices in each household. Made by supplying.
- the side-by-side method is a method that realizes 3D playback by horizontally arranging a left-eye image necessary for stereoscopic viewing and a right-eye image necessary for stereoscopic viewing and packaging them in one frame area (patent) Reference 1).
- the conventional 3D display device determines whether or not the input video stream is 3D video. When it is determined that the video stream is 3D video, the picture data of each frame constituting the video stream is always in a side-by-side format. The right eye image and the left eye image are decoded on the assumption that the right eye image is stored in the right half and the left eye image is stored in the left half.
- 3D TV broadcasting is a 1TS-1VS system (a system that transmits one video stream with one transport stream), and switching between 3D mode and 2D mode is not realized. Therefore, the user can only view the 3D television broadcast as 3D video, and it cannot be said that sufficient consideration is given to the user.
- 1TS-1VS system a system that transmits one video stream with one transport stream
- the transport stream storing the video stream for the right eye and the transport stream storing the video stream for the left eye are read from the BD-ROM and provided to the decoder, so that the 2D mode and 3D mode Free mode switching is realized.
- these two transport streams (2TS) are converted into interleaved files. Recorded on BD-ROM.
- the transport stream that can be used for one TV program is 1TS
- the video stream for the right eye and the video stream for the left eye are used by using two TSs. It cannot be transmitted.
- An object of the present invention is to provide an encoding method capable of realizing free mode switching between 2D mode and 3D mode on the assumption that a transport stream that can be used for one TV program is 1TS. It is.
- an encoding method includes an encoding step for compressing and encoding an image to generate a video stream, and a multiplexing process for a plurality of video streams generated by the encoding step. And a multiplexing step for obtaining one transport stream, and the plurality of video streams include a 2D video stream constituting a 2D video, and a 3D video is configured by combining the plurality of video streams,
- the combination of video streams constituting 3D video includes a combination of the 2D video stream and one other video stream, and a combination of two or more other video streams other than the 2D video stream, and the transport stream includes: 3D video that identifies the video streams that make up the 3D video It includes image specifying information.
- the transport stream is used when the display device is performing 2D playback and switching to 3D playback. By referring to the associated information, it is possible to know which video stream is necessary for 3D playback.
- the content table is stored when the content table is arranged at the beginning of the transport stream or arranged at predetermined intervals.
- the 2D video stream, the left-eye video stream constituting the left-eye video, and the video stream constituting the right-eye video each have information indicating a corresponding stream identifier in the transport stream, 2D playback, or In 3D playback, a video stream to be demultiplexed can be specified, and switching of the demultiplexing target between the 2D mode and the 3D mode can be performed at high speed.
- the content table has a flag indicating whether or not any of the 2D video stream and a plurality of video streams constituting the 3D video matches, so the packet storing the content table is transferred.
- the camera assignment in the stream descriptor indicates the channel configuration of the camera, the camera environment at the time of shooting can be reproduced during playback.
- information indicating which of the 2D video stream and the other video stream to use the closed caption subtitle data existing in the video stream is described in the content table.
- the video stream to be demultiplexed can be specified by referring to the stream information.
- Video can be played back.
- the stream identifier of the video stream that is the partner of the combination is described in the stream information corresponding to each video stream in the content table, it is necessary for 3D playback by referring to the stream information The other video stream can be identified.
- FIG. 38 is a diagram illustrating the structure of 3D_system_info_descriptor in the stream configuration of 2D + L + R. It is a figure which shows the value set to 3D_playback_type
- Fig. 10 is a diagram illustrating a structure of 3D_service_info_descriptor in a stream configuration of 2D + L + R.
- Fig. 11 is a diagram illustrating a structure of 3D_combi_info_descriptor in a stream configuration of 2D + L + R.
- FIG. 3 is a diagram illustrating a transstream configuration (2D + Side-by-Side) that stores Side-by-Side video in addition to 2D video.
- FIG. 5 is a diagram illustrating a transport stream configuration (2D + R1 + R2) that stores R video videos of multiple viewpoints in addition to video used as L video during 2D playback and 3D playback.
- FIG. 11 is a diagram illustrating a structure of 3D_system_info_descriptor in a stream configuration of 2D ⁇ ⁇ ⁇ ⁇ + R1 + R2. [Fig. 10] Fig. 10 is a diagram illustrating a structure of 3D_service_info_descriptor in a stream configuration of 2D + R1 + R2. [Fig. 11] Fig. 11 is a diagram illustrating a structure of 3D_combi_info_descriptor in a stream configuration of 2D + R1 + R2.
- FIG. 10 is a flowchart showing a flow of encoding processing of the data creation device 4000. It is a figure which shows the internal structure of 3D digital television 4200.
- FIG. 21 is a flowchart illustrating an example of a flow of program reproduction processing by the 3D digital television 4200. It is a flowchart which shows the flow of a process of the stream of 2D (+) + SBS. It is a flowchart which shows the flow of a 2D / SBS stream process. It is a flowchart which shows the flow of a process of the stream of 2D / L custom character + custom character R.
- 10 is a flowchart showing a flow of processing of a stream of 2D / L + R1 + R2.
- 10 is a flowchart showing a flow of MPEG2 ⁇ + AVC + AVC stream processing.
- 10 is a flowchart showing a flow of processing of a stream of MPEG2 + (MVC (Base) + (MVC (Dependent).
- the right eye and the left eye have a slight difference in appearance between the image seen from the right eye and the image seen from the left eye due to the difference in position. Using this difference, a human can recognize a visible image as a solid.
- a planar image is made to look like a three-dimensional image using human parallax.
- the time-separation method is a method in which left-eye video and right-eye video are alternately displayed in the time axis direction, and left and right scenes are superimposed in the brain by an afterimage reaction of the eyes to be recognized as a stereoscopic video.
- FIG. 24 schematically shows an example of generating parallax images of a left-eye video and a right-eye video from a 2D video and a depth map.
- the depth map has a depth value corresponding to each pixel in the 2D video.
- the circular object in the 2D video is assigned information indicating that the depth map has a high depth.
- the area is assigned information indicating that the depth is low.
- This information may be stored as a bit string for each pixel, or may be stored as an image (for example, “black” indicates that the depth is low and “white” indicates that the depth is high).
- the parallax image can be created by adjusting the parallax amount of the 2D video from the depth value of the depth map. In the example of FIG. 24, since the depth value of the circular object in the 2D video is high, when creating a parallax image, the amount of parallax of the pixel of the circular object is increased, and the depth value is low in regions other than the circular object.
- the left-eye image and the right-eye image are created by reducing the amount of parallax of the pixels of the circular object. If the left-eye image and the right-eye image are displayed using a time separation method or the like, stereoscopic viewing is possible.
- the playback device in the present embodiment is a device that decodes 2D video or 3D video and transfers the video to the display.
- a digital television will be described as an example.
- the digital television includes a playback device 100 that can view 3D video and a 2D digital television 300 that can play back only 2D video that does not support playback of 3D video.
- FIG. 2 (a) is a diagram showing a form of usage of the playback device. As shown in the figure, the digital television 100 and the 3D glasses 200 are configured and can be used by the user.
- the playback device 100 can display 2D video and 3D video, and displays video by playing back a stream included in the received broadcast wave.
- the playback apparatus 100 realizes stereoscopic viewing by wearing 3D glasses 200 by a user.
- the 3D glasses 200 include a liquid crystal shutter, and allow a user to view a parallax image by the continuous separation method.
- the parallax image is a set of videos composed of a video that enters the right eye and a video that enters the left eye, and performs stereoscopic viewing so that only pictures corresponding to each eye enter the user's eyes.
- FIG. 2B shows the display time of the left-eye video.
- the above-described 3D glasses 200 transmit the liquid crystal shutter corresponding to the left eye and shield the liquid crystal shutter corresponding to the right eye.
- FIG. 3C shows the time when the right-eye video is displayed.
- the liquid crystal shutter corresponding to the right eye is made transparent, and the liquid crystal shutter corresponding to the left eye is shielded from light.
- the left and right pictures are alternately output in the time axis direction in the previous time separation method, whereas the left-eye picture and right-eye picture in the vertical direction in one screen.
- the picture for the left eye is imaged only on the left eye and the picture for the right eye is only on the right eye through the lens on the eyelid called lenticular lens on the display surface.
- lenticular lens There is a method in which a picture with parallax can be shown and viewed as 3D by forming images.
- a device having a similar function for example, a liquid crystal element may be used.
- the left eye pixel is equipped with a vertically polarized filter
- the right eye pixel is equipped with a horizontally polarized filter
- the viewer is provided with polarized glasses with a vertically polarized filter for the left eye and a horizontally polarized filter for the right eye.
- polarized glasses with a vertically polarized filter for the left eye and a horizontally polarized filter for the right eye.
- the 2D digital television 300 cannot realize stereoscopic viewing, unlike playback device 100, as shown in FIG. 2 (d).
- the 2D digital television 300 can display only 2D video, and can reproduce the stream included in the received broadcast wave only as 2D video.
- the MPEG-2 transport stream is a standard for multiplexing and transmitting various streams such as video and audio. It is standardized in ISO / IEC13818-1 and ITU-T recommendation H222.0.
- FIG. 6 is a diagram showing the configuration of a digital stream in the MPEG-2 transport stream format.
- a transport stream is obtained by multiplexing a video stream, an audio stream, a subtitle stream, and the like.
- the video stream stores the main video of the program
- the audio stream stores the main audio portion and sub-audio of the program
- the subtitle stream stores the subtitle information of the program.
- the video stream is encoded and recorded using a method such as MPEG-2, MPEG-4 AVC.
- the audio stream is compressed and encoded and recorded by a method such as Dolby AC-3, MPEG-2 AAC, MPEG-4 AAC, HE-AAC.
- moving picture compression coding such as MPEG-2, MPEG-4 AVC, SMPTE VC-1, etc.
- data amount is compressed using redundancy in the spatial direction and temporal direction of moving images.
- inter-picture predictive coding is used as a method of using temporal redundancy.
- inter-picture predictive coding when a certain picture is coded, a picture that is forward or backward in display time order is used as a reference picture. Then, the amount of motion from the reference picture is detected, and the amount of data is compressed by removing the redundancy in the spatial direction from the difference value between the motion compensated picture and the picture to be coded.
- a picture that does not have a reference picture and performs intra-picture predictive coding using only a picture to be coded is called an I picture.
- a picture is a unit of encoding that includes both a frame and a field.
- a picture that is inter-picture prediction encoded with reference to one already processed picture is called a P picture
- a picture that is inter-picture predictively encoded with reference to two already processed pictures at the same time is called a B picture.
- a picture that is referred to by other pictures in the B picture is called a Br picture.
- a field having a frame structure and a field having a field structure are referred to as a video access unit here.
- the video stream has a hierarchical structure as shown in FIG.
- a video stream is composed of a plurality of GOPs (Group of Pictures). By using this as a basic unit for encoding processing, editing and random access of moving images are possible.
- a GOP is composed of one or more video access units.
- the video access unit is a unit for storing coded data of a picture, and stores data of one frame in the case of a frame structure and one field in the case of a field structure.
- Each video access unit includes an AU identification code, a sequence header, a picture header, supplementary data, compressed picture data, padding data, a sequence end code, a stream end code, and the like.
- Each data is stored in units called NAL units in the case of MPEG-4 AVC.
- AU identification code is a start code indicating the head of the access unit.
- the sequence header is a header that stores common information in a playback sequence composed of a plurality of video access units, and stores information such as resolution, frame rate, aspect ratio, and bit rate.
- the picture header is a header that stores information such as the coding method of the entire picture.
- the supplemental data is additional information that is not essential for decoding the compressed data, and stores, for example, closed caption character information or GOP structure information that is displayed on the TV in synchronization with the video.
- the compressed picture data stores compression-encoded picture data.
- the padding data stores meaningless data for formatting. For example, it is used as stuffing data for maintaining a predetermined bit rate.
- the sequence end code is data indicating the end of the reproduction sequence.
- the stream end code is data indicating the end of the bit stream.
- the configuration of the contents of the AU identification code, sequence header, picture header, supplemental data, compressed picture data, padding data, sequence end code, and stream end code differ depending on the video encoding method.
- the AU identification code is AU delimiter (Access Unit Delimiter)
- the sequence header is SPS (Sequence Parameter Set)
- the picture header is PPS (Picture Parameter Set)
- compressed picture data Indicates multiple slices
- supplemental data corresponds to SEI (Supplemental Enhancement Information)
- padding data corresponds to FillerData
- sequence end code corresponds to End of Sequence
- stream end code corresponds to End of Stream.
- the sequence header corresponds to sequence_Header, sequence_extension, group_of_picture_header, the picture header corresponds to picture_header, picture_coding_extension, the compressed picture data corresponds to a plurality of slices, the supplemental data corresponds to user_data, and the sequence end code corresponds to sequence_end_code.
- the sequence header may be necessary only in the video access unit at the head of the GOP and may not be present in other video access units.
- the picture header may refer to that of the previous video access unit, and there is no picture header in its own video access unit.
- the cropping area information and the scaling information will be described with reference to FIG.
- the area of the encoded frame and the area actually used for display can be changed.
- an area to be actually displayed out of the encoded frame area can be designated as a “cropping area”.
- the frame_cropping information is obtained by dividing the difference between the upper line / underline / left line / right line of the cropping area and the upper line / underline / left line / right line of the encoded frame area. Specify as an amount.
- frame_cropping_flag is set to 1, and the upper / lower / left / right crop amounts are specified in frame_crop_top_offset // frame_crop_bottom_offset // frame_crop_left_offset // frame_crop_right_offset.
- the vertical and horizontal sizes of the cropping area (display_horizontal_size of sequence_display_extension, display_vertical_size) and the difference information (frame_centre_horizontal_offset of picture_display_extension, frame_centre_vertical_offset) of the center of the encoded frame area and the center of the cropping area are used.
- a cropping area can be specified.
- scaling information indicating a scaling method when the cropping area is actually displayed on a television or the like. This is set as an aspect ratio, for example.
- the playback device uses the aspect ratio information to up-convert the cropping area for display.
- aspect ratio information (aspect_ratio_idc) is stored in the SPS as scaling information.
- the aspect ratio is specified as 4: 3.
- aspect ratio information (aspect_ratio_information) is stored in sequence_header.
- Each stream included in the transport stream is identified by a stream identification ID called PID.
- PID stream identification ID
- the composite apparatus can extract the target stream.
- the correspondence between PID and stream is stored in the descriptor of the PMT packet described later.
- FIG. 6 is a diagram schematically showing how the transport stream is multiplexed.
- a video stream 501 composed of a plurality of video frames and an audio stream 504 composed of a plurality of audio frames are converted into PES packet sequences 502 and 505, respectively, and converted into TS packets 503 and 506, respectively.
- the data of the subtitle stream 507 is converted into a PES packet sequence 508 and further converted into a TS packet 509.
- the MPEG-2 transport stream 513 is configured by multiplexing these TS packets into one stream.
- FIG. 8 shows in more detail how the video stream is stored in the PES packet sequence.
- the first level in the figure shows a video frame sequence of the video stream.
- the second level shows a PES packet sequence.
- multiple Video Presentation Unit I pictures, B pictures, and P pictures in the video stream are divided for each picture and stored in the payload of the PES packet.
- Each PES packet has a PES header, and a PTS (Presentation Time-Stamp) that is a picture display time and a DTS (Decoding Time-Stamp) that is a picture decoding time are stored in the PES header.
- PTS Presentation Time-Stamp
- DTS Decoding Time-Stamp
- FIG. 9 shows the data structure of TS packets constituting the transport stream.
- the TS packet is a 188-byte fixed-length packet composed of a 4-byte TS header, an adaptation field, and a TS payload.
- the TS header includes transport_priority, PID, adaptation_field_control, and the like.
- the PID is an ID for identifying a stream multiplexed in the transport stream as described above.
- the transport_priority is information for identifying the type of packet in TS packets with the same PID.
- adaptation_field_control is information for controlling the configuration of the adaptation field and the TS payload. There are cases where only one of the adaptation field and the TS payload exists or both, and adaptation_field_control indicates the presence / absence thereof. When adaptation_field_control is 1, only the TS payload exists, when adaptation_field_control is 2, only the adaptation field exists, and when adaptation_field_control is 3, it indicates that both the TS payload and the adaptation field exist.
- the adaptation field is a storage area for storing information such as PCR and stuffing data to make the TS packet a fixed length of 188 bytes. PTS packets are divided and stored in the TS payload.
- TS packets included in the transport stream include PAT (Program Association Table), PMT (Program Map Table), PCR (Program Clock Reference), etc. in addition to video, audio, and subtitle streams. These packets are called PSI (Program Specific Information).
- the PAT indicates what the PMT PID used in the transport stream is, and the PAT of the PAT itself is registered with 0.
- the PMT has PID of each stream such as video / audio / subtitles included in the transport stream and stream attribute information corresponding to each PID, and has various descriptors related to the transport stream.
- the descriptor includes copy control information for instructing permission / non-permission of copying of the AV stream.
- the PCR information on the STC time corresponding to the timing at which the PCR packet is transferred to the decoder have.
- FIG. 10 is a diagram for explaining the data structure of the PMT in detail.
- a PMT header describing the length of data included in the PMT is arranged at the head of the PMT. After that, a plurality of descriptors related to the transport stream are arranged.
- the copy control information described above is described as a descriptor.
- a plurality of pieces of stream information regarding each stream included in the transport stream are arranged after the descriptor.
- the stream information is composed of a stream descriptor in which a stream type, a stream PID, and stream attribute information (frame rate, aspect ratio, etc.) are described to identify the compression codec of the stream. *
- FIG. 3 shows the user's face on the left side, and the right side shows an example when the dinosaur skeleton as the object is viewed from the left eye and the example when the dinosaur skeleton as the object is viewed from the right eye. ing. If it repeats from the translucency and shading of the right eye and the left eye, the left and right scenes are overlapped by the afterimage reaction of the eyes in the user's brain, and it can be recognized that there is a stereoscopic image on the extension line in the center of the face. .
- an image entering the left eye is referred to as a left eye image (L image), and an image entering the right eye is referred to as a right eye image (R image).
- L image left eye image
- R image right eye image
- a moving image in which each picture is an L image is referred to as a left view video
- a moving image in which each picture is an R image is referred to as a right view video.
- the 3D video format that compresses and encodes the left-view video and the right-view video includes a frame compatible method and a service compatible method.
- the first frame compatible method is a method of performing normal moving image compression coding by thinning out or reducing the corresponding pictures of the left-view video and right-view video and combining them into one picture. .
- each corresponding picture of the left-view video and the right-view video is compressed in half in the horizontal direction and then combined in one picture by arranging them side by side.
- a moving image based on the combined picture is streamed by performing normal moving image compression encoding.
- the stream is decoded into a moving image based on a normal moving image compression encoding method.
- Each picture of the decoded moving image is divided into left and right images, and each picture corresponding to left-view video and right-view video is obtained by extending the picture in the horizontal direction twice.
- a stereoscopic image as shown in FIG. 2 can be obtained by alternately displaying the obtained left-view video picture (L image) and right-view video picture (R image).
- the frame compatible method includes the Top and Bottom method in which left and right images are arranged vertically, and the Line Alternative method in which left and right images are alternately arranged for each line in a picture.
- FIG. 1 is a diagram for explaining frame packing information.
- the lower part of FIG. 1 shows a video frame sequence.
- the section (A) is a section where Side-by-Side video is played back
- the section (B) is a section where 2D video is played back
- the section (C) is a section where TopBottom video is played back.
- An example of frame packing information in such a playback section is shown in the upper part of FIG.
- the frame packing information includes a frame storage type, a cancel flag, and a repeat flag.
- the frame storage type indicates the type of method for storing the left and right images for stereoscopic viewing in the frame.
- the methods such as “Side-by-Side”, “TopBottom”, “Checkerboard”, and “Line-by-line” described above are used. It is information to identify.
- the Frame_packing_arrangement of MPEG-4 AVC corresponds to the Frame_packing_arrangement_type.
- the repeat flag indicates the validity period of the frame packing information. If it is 0, it indicates that the frame packing information is valid only for the corresponding frame. If it is 1, the next video sequence comes as the corresponding frame packing information. Or until a frame having frame packing information later than the corresponding frame in the display order comes.
- MPEG-4 AVC Frame_packing_arrangement corresponds to Frame_packing_arrangement_repetition_period.
- the cancel flag is a flag for canceling the valid period of the previous frame packing information. When the cancel flag is 1, the previously transmitted frame packing information is canceled, and when it is 0, the corresponding frame packing information is valid.
- MPEG-4 AVC Frame_packing_arrangement corresponds to Frame_packing_arrangement_cancel_flag.
- the frame packing information (A) stored at the beginning of the Side-by-Side playback section the frame storage type is Side-by-Side, the repeat flag is 1, and the cancel flag is 0. Since no frame packing information is stored at the head of the Side-by-Side playback section and the repeat flag is 1, the frame sequence of this section has a head at the head of the Side-by-Side playback section.
- the stored frame packing information (A) is valid.
- the cancel flag is 1, and the frame storage type and repeat flag are not stored. Since frame packing information is unnecessary in the 2D section, no frame packing information is stored after cancellation at the head.
- frame packing information (C) is stored in all frames.
- the frame storage type of the frame packing information (C) is TopBottom, the repeat flag is 0, and the cancel flag is 0. Since the repeat flag is 0, it is necessary to store frame packing information in all frames in order to indicate that all frames are TopBottom.
- the playback apparatus can realize stereoscopic display processing according to the method by referring to the information.
- a left-view video stream and a right-view video stream which are video streams obtained by digitizing and compression-coding left-view video and right-view video, are used.
- the left-view video and the right-view video that are compression-encoded by the inter-picture predictive encoding technique using the correlation characteristic between the viewpoints is particularly called a multi-view encoding system.
- FIG. 5 is a diagram illustrating an example of an internal configuration of a left-view video stream and a right-view video stream for stereoscopic viewing using a multi-view encoding method.
- the second row in the figure shows the internal structure of the left-view video stream.
- This stream includes picture data of picture data I1, P2, Br3, Br4, P5, Br6, Br7, and P9. These picture data are decoded according to Decode Time Stamp (DTS).
- DTS Decode Time Stamp
- the first row shows a left eye image.
- the decoded picture data I1, P2, Br3, Br4, P5, Br6, Br7, and P9 are reproduced in the order of I1, Br3, Br4, P2, Br6, Br7, and P5 in accordance with PTS. Will be played.
- a picture that does not have a reference picture and performs intra-picture predictive coding using only a picture to be coded is called an I picture.
- a picture is a unit of encoding that includes both a frame and a field.
- a picture that is inter-picture prediction encoded with reference to one already processed picture is referred to as a P picture
- a picture that is inter-picture predictively encoded while simultaneously referring to two already processed pictures is referred to as a B picture.
- B picture pictures referenced from other pictures are called Br pictures.
- the fourth row shows the internal structure of the left-view video stream.
- This left-view video stream includes picture data of P1, P2, B3, B4, P5, B6, B7, and P8. These picture data are decoded according to DTS.
- the third row shows a right eye image.
- the right-eye image is reproduced by reproducing the decoded picture data P1, P2, B3, B4, P5, B6, B7, and P8 in the order of P1, B3, B4, P2, B6, B7, and P5 according to PTS. Will be played.
- the display of one of the left-eye image and right-eye images with the same PTS is displayed for half the PTS interval (hereinafter referred to as “3D display delay”). Just display with a delay.
- the fifth row shows how the state of the 3D glasses 200 is changed. As shown in the fifth row, the right-eye shutter is closed when the left-eye image is viewed, and the left-eye shutter is closed when the right-eye image is viewed.
- left-view video stream and right-view video stream are compressed by inter-picture prediction encoding using correlation characteristics between viewpoints in addition to inter-picture prediction encoding using temporal correlation characteristics.
- Pictures in the right-view video stream are compressed with reference to pictures at the same display time in the left-view video stream.
- the first P picture of the right-view video stream refers to the I picture of the left-view video stream
- the B picture of the right-view video stream refers to the Br picture of the left-view video stream
- two of the right-view video streams The P picture of the eye refers to the P picture of the left view video stream.
- the left-view video stream and right-view video stream that have been compression-encoded, one that can be decoded alone is called a “base-view video stream”.
- the left-view video stream and the right-view video stream are compressed and encoded based on the inter-frame correlation characteristics with the individual picture data constituting the base-view video stream, and the base-view video stream is decoded.
- a video stream that can be decoded is referred to as a “dependent view stream”.
- the base-view video stream and the dependent-view stream may be stored and transmitted as separate streams, or may be multiplexed into the same stream such as MPEG2-TS.
- MVC Multiview Video Coding
- encoding methods for storing frame packing information include a method of arranging only at the beginning of a playback section, such as frame packing information (A) and frame packing information (B), and frame packing ( There are cases where methods of storing in all the frames are mixed as shown in C).
- frame packing information A
- frame packing information B
- frame packing There are cases where methods of storing in all the frames are mixed as shown in C).
- processing is inefficient in the reproduction and editing apparatus. For example, when performing jump playback from a video frame other than the head of the Side-by-Side playback section (A) in FIG. 1, the frame packing information stored in the Side-By-Side head frame is analyzed. Need to get.
- the following video format structure is adopted so that the encoding method for storing the frame packing information can be specified in advance and the efficiency of the reproduction process can be improved.
- the structure of the video format will be described with reference to FIG.
- the example of FIG. 13 is an example in the case of storing 3D video of a side-by-side frame compatible method.
- the video stream stored in the transport stream is a video stream compressed by a video encoding method such as MPEG-4 AVC or MPEG-2.
- the frame packing information is stored in the supplementary data of the video stream.
- the frame packing information is the information described with reference to FIG.
- the frame packing information stores a frame storage type, a repeat flag, and a cancel flag.
- the frame packing information is not stored in the supplementary data of all the video access units, but the repeat flag is set to 1 and stored only at the head of the GOP for other video accesses. It is possible not to store in the unit.
- the frame packing information descriptor is stored in the PMT packet.
- the frame packing information descriptor is prepared for each video stream stored in the transport stream, and stores attribute information of the frame packing information included in the supplementary data of the corresponding video stream.
- the frame packing information descriptor stores “frame storage type”, “frame packing information storage type”, and “start PTS”.
- the frame storage type has the same meaning as the frame storage type of the frame packing information, and indicates the frame storage method (Side-by-Side method, etc.) of the stereoscopic video of the corresponding video stream. This information matches the frame storage type of the frame packing information included in the supplementary data of the corresponding video stream.
- the playback device can determine the frame storage method of the stereoscopic video without analyzing the video stream. Thereby, for example, the playback device can determine the 3D display method in advance, so that it is possible to perform processing necessary for 3D display, such as generation processing of OSD for 3D display, before decoding the video stream.
- the frame packing information storage type indicates a method of inserting frame packing information included in the corresponding video stream.
- the frame packing information can be stored only in the GOP head by setting the repeat flag to 1 and not stored in other video access units. Conversely, the repeat flag can be set to 0 and stored in all frames.
- the frame packing information storage type is information for specifying the storage method of the frame packing information. That is, if the frame packing information type is “GOP unit”, the frame is only included in the supplementary data of the video access unit at the head of the GOP. Packing information is stored, and “access unit unit” indicates that frame packing information is stored only in supplementary data of all video access units.
- the playback apparatus can determine the storage method of the frame packing information without analyzing the video stream, so that the playback and editing processes can be made more efficient. For example, in the case of a playback device that performs jump playback even in a frame other than the GOP head, by referring to the frame packing information storage type, if it indicates “GOP head”, it always plays back from the GOP head only. Can be controlled to start.
- the frame packing information descriptor may store information indicating whether the attribute changes in GOP units.
- the frame packing information storage type is “frame unit” and the attribute does not change in the GOP, the same frame packing information can be stored in all frames. Therefore, analysis of frame packing information other than the GOP head included in the video stream can be skipped.
- the start PTS indicates the time when the corresponding frame packing information descriptor becomes valid. Since the position of the PMT packet is generally not synchronized with the multiplexing position of the video stream, the time when the corresponding frame packing information descriptor is valid cannot be known on the display time of the video stream. Therefore, the playback device can know the time when the frame packing information descriptor becomes valid by referring to the start PTS.
- the starting PTS may be restricted to indicate the PTS attached to the video, so that the playback device can be clearly instructed to synchronize with the video.
- the packets may be arranged in the order of multiplexing (code). When there are a plurality of PMT packets including the start PTS, only the top packet may be arranged forward in the multiplexing (code) order.
- FIG. 14 is a diagram showing an example of the relationship between the frame packing information descriptor and the frame packing information.
- the lower part of FIG. 14 is a diagram in which video frame sequences are arranged in display order.
- the section (A) is a section where Side-by-Side video is played back
- the section (B) is a section where 2D video is played back
- the section (C) is a section where TopBottom video is played back.
- An example of the frame packing information in such a playback section is shown in the middle part of FIG. This is the same as the configuration shown in FIG.
- the upper part of FIG. 14 shows the structure of the frame packing information descriptor in this data structure.
- the frame packing information descriptor (A) includes information corresponding to the frame packing information in the Side-by-Side playback section (A).
- Each value of the frame packing information descriptor (A) is set as follows.
- the frame storage type is “Side-by-Side”, which is the same as the frame storage type of the frame packing information.
- the frame packing information storage type is “GOP head” because the frame packing information is stored only at the head of the section.
- the first “video PTS value (in the example, 180000)” in the section (A) is set.
- the frame packing information descriptor (B) includes information corresponding to the frame packing information in the 2D playback section (B).
- Each value of the frame packing information descriptor (B) is set as follows.
- the frame storage type is the same as the frame storage type of the frame packing information and is not set. Alternatively, if the frame storage type “2D” is defined, the value is set.
- the frame packing information storage type is set to “GOP head” because the frame packing information is stored only at the head of the section. In the start PTS, the first “video PTS value (5580000 in the example)” in the playback section (B) is set.
- the frame packing information descriptor (C) includes information corresponding to the frame packing information in the TopBottom playback section (C).
- Each value of the frame packing information descriptor (C) is set as follows.
- the frame storage type is “TopBottom”, which is the same as the frame storage type of the frame packing information, and the frame packing information storage type is “access unit unit” because the frame packing information is stored in all video access units in the section.
- the first “video PTS value (10980000 in the example)” in the section (C) is set.
- the playback device is specifically a plasma television or a liquid crystal television that supports 3D video display, and receives a transport stream for sending a video stream.
- the television is a 3D system that uses shutter glasses in a continuous separation system.
- the playback device is connected to the IP network and the playback device, and also decodes and displays the video stream output from these.
- the playback apparatus includes a tuner 1501, a NIC 1502, a demultiplexing unit 1503, a video decoding unit 1504, a display determination unit 1505, a display processing unit 1506, a display unit 1507, a frame buffer (1) 1510, a frame buffer ( 2) 1511 and switch 1512.
- the tuner 1501 has a function of receiving a digital broadcast wave transport stream and demodulating the received signal.
- the NIC 1502 is connected to the IP network and has a function of receiving a transport stream output from the outside.
- the demultiplexing unit 1503 separates the received transport stream into a video stream and other audio streams, and outputs the video stream to the video decoding unit 1504.
- the demultiplexing unit extracts a system packet such as PSI from the received transport stream, acquires a “frame packing information descriptor” from the PMT packet, and notifies the display determination unit and the video decoding unit.
- the demultiplexing unit 1503 can read the transport stream from the recording medium in addition to the input from the tuner 1501 and the NIC 1502.
- the video decoding unit 1504 When receiving a video stream from the demultiplexing unit 1503, the video decoding unit 1504 has a function of decoding the received stream and extracting “frame packing information” in the video stream.
- the video decoding unit 1504 decodes video in units of frames.
- the “frame packing information storage type” of the frame packing information descriptor notified from the demultiplexing unit 1503 is a GOP unit, the extraction of “frame packing information” can be skipped except for the video access unit at the head of the GOP.
- the video decoding unit 1504 writes the decoded frame into the frame buffer (1) 1508 and outputs “frame packing information” to the display determination unit 1506.
- Frame buffer (1) 1508 has an area for storing frames decoded by the video decoding unit 1504.
- the display determination unit 1505 determines a display method based on “frame packing information descriptor” and “frame packing information”.
- the 3D video storage method is determined according to the frame storage type stored in the “frame packing information descriptor” or “frame packing information”, and the “start PTS” in the “frame packing information descriptor” or the frame packing information is stored.
- the display processing unit is notified at the timing of the video PTS.
- the display determination unit 1505 determines the display method in this way, and notifies the display processing unit 1506 of the content.
- the display processing unit 1506 transforms the decoded frame data stored in the frame buffer (1) according to an instruction from the display determination unit 1505, and writes it to the frame buffer (L) or the frame buffer (R).
- the display processing unit 1506 crops the HalfHD left-eye image from the left half of the frame and writes it to the frame buffer (L), and the display processing unit 1506 displays the HalfHD right-eye image from the right half of the frame.
- the frame buffer (R) In the case of the TopBottom method, the display processing unit 1506 crops the HalfHD left-eye image from the upper half of the frame and writes it to the frame buffer (L), and the display processing unit 1506 crops the HalfHD right-eye image from the lower half of the frame.
- frame buffer (R) In the case of 2D, the video in the frame buffer (1) is written to both the frame buffer (L) and the frame buffer (R).
- the frame buffer (L) 1510 and the frame buffer (R) 1511 have areas for storing frames output from the display processing unit 1506.
- the switch 1512 selects the frame images written in the frame buffer (L) 1510 and the frame buffer (R) 1511 and transfers them to the display unit.
- the frame buffer (L) 1510 and the frame buffer (R) 1511 are alternately selected and displayed according to the frame to be displayed.
- the display unit 1506 displays the frame transferred from the switch 1512.
- the display unit 1506 communicates with the 3D glasses.
- the left eye image is displayed, the left eye side of the 3D glasses is opened, and when the right eye image is displayed, the liquid crystal shutter of the 3D glasses is opened so that the right eye side of the 3D glasses is opened.
- Control Note that when displaying 2D video, control of 3D glasses is not performed.
- the frame packing information descriptor may be stored in a “SI (Service Information)” descriptor including program information, a TS packet header, a PES header, or the like.
- SI Service Information
- the frame packing information storage type of the frame packing information descriptor includes “GOP unit” and “access unit unit”, but “PES packet unit” indicating that one exists in the PES packet, There may be other types such as “I picture unit” indicating the presence of one, and “attribute switching unit” indicating that one exists every time the value included in the frame packing information is switched.
- the frame packing information descriptor may include an identifier that indicates whether there is a change in the value of the frame packing information descriptor stored in the previous PMT packet. By referring to this identifier, when there is no change, the analysis processing of the frame packing information descriptor, the notification to the display determination unit, and the processing of the display determination unit can be skipped.
- a repeat flag may be stored as the frame packing information storage type of the frame packing information descriptor. For example, if the repeat flag of the frame packing information descriptor is 1, the playback device can be determined as a GOP unit, and if the repeat flag of the frame packing information descriptor is 0, the playback unit can be determined.
- the frame packing information storage type of the frame packing information descriptor may be set for each frame storage type.
- “GOP unit” may be set for the Side-by-Side method
- “Frame unit” may be set for the TopBottom method.
- the frame packing information storage type of the frame packing information descriptor may be set for each ID of the frame packing information.
- a plurality of pieces of frame packing information can be set with IDs. This ID corresponds to Frame_packing_arrangement_id in the Frame_packing_arrangement SEI of MPEG-4 AVC.
- a frame packing information storage type may be set for each ID.
- FIG. 16 shows the relationship between the TS packet sequence and the video frame sequence to be played back.
- the video frame sequence is a Side-by-Side 3D video playback section up to PTS 5580000, and a 2D video playback section from PTS 5580000.
- the structure of the frame packing information descriptor included in the PMT packet in the TS packet is described in (1) to (4) above.
- (1) is a descriptor indicating a Side-by-Side section
- (2), (3), and (4) are descriptors indicating a 2D section.
- the video display time is still a section displaying Side-By-Side. Therefore, when the playback device performs display processing with reference to the frame packing information descriptor in the PMT packet at the time when the PMT packet arrives, display switching processing is performed in the time of the gap (A), The gap (A) cannot be correctly played back as 3D video.
- processing priority is stored in the frame packing information descriptor.
- This “processing priority” includes “descriptor priority” indicating that the PMT frame packing information descriptor is processed with priority, and “video priority” indicating that the frame packing information stored in the video stream is processed with priority. "Are prepared.
- the processing priority is “descriptor priority”
- the playback device performs display switching processing with priority on the frame packing information descriptor included in the PMT.
- display switching processing is performed in the gap (A).
- the reproduction transition in this case is indicated by a reproduction transition X in the lower part of the lower part of FIG. In this way, the end of the Side-by-Side playback section cannot be correctly played back by the display switching process, but instead, the 2D playback section can be played back correctly from the beginning.
- the playback device When the processing priority is “video priority”, the playback device performs display switching processing by giving priority to the frame packing information included in the video. Even if the PMT packet arrives, the display switching process is not performed, and the display switching process is performed from the timing of the display time of the video stream. In this case, the data is correctly reproduced in the gap (A), and the display switching process is performed using the section (B) from the time of transition to the 2D video of PTS 5580000.
- the reproduction transition in this case is indicated by reproduction transition Y in the lower part of the lower part of FIG. In this way, the top part of the 2D playback section cannot be played back correctly by the display switching process, but the end part of the Side-by-Side playback section can be played back correctly.
- processing priority is set to “descriptor priority” and priority is given to side-by-side 3D video playback.
- the processing priority may be set to “video priority”.
- a meaningless video such as a black video may be stored in a section where the display switching process is performed according to the processing priority.
- the section is a gap (A) when the processing priority is “descriptor priority”, and is a section (B) when the processing priority is “video priority”. In this way, it is not necessary to generate a section in which the user cannot enjoy the content.
- display switching start time may be set in the frame packing information descriptor instead of processing priority as shown in FIG. With this configuration, the display processing start time can be more accurately controlled.
- the left-eye video and the right-eye video are stored in one transport stream as separate video streams as shown in FIG. Either of the above images may be played back, and the 3D video may be played back with both the left eye and the right eye.
- FIG. 19 is a diagram in which the frames of the video stream of the left eye / right eye video in the configuration of FIG. 18 are arranged in display order.
- 2D video and 3D video as shown in the upper part of FIG. 19
- one video frame sequence is redundant data.
- a 3D playback information descriptor is prepared as shown in FIG. 20 so that the 2D playback section and the 3D playback section of the video stream multiplexed in the transport stream can be discriminated.
- the 3D playback information descriptor is stored in the PMT packet.
- a playback method and a start PTS are prepared.
- the playback method is an identifier indicating whether 2D playback or 3D playback
- the start PTS is time information for indicating from which frame the corresponding playback section starts.
- the 3D playback information descriptor (A) indicates that the 3D playback section starts from PTS 180000
- the 3D playback information descriptor (B) indicates that the 2D playback section starts from PTS 5580000
- the 3D playback information descriptor (C) indicates PTS 109800000. Indicates that the 3D playback section begins.
- the 3D playback device can determine where the 3D playback is or 2D playback, so only the left-eye video frame sequence can be decoded and displayed in the 2D video playback section. During this time, it is not necessary to store data in the right-eye video frame sequence, and a large bit rate can be secured for encoding the left-eye video frame sequence.
- the 3D playback information descriptor may specify the PID of the video to be played back as 2D video.
- a video stream that is played back as 2D video is hereinafter referred to as a base video stream
- a video stream that is played back only as 3D video is referred to as an extended video stream.
- the base video may use a normal stream type
- the extended video may have a special stream type.
- the 3D playback information descriptor may be stored in supplementary data or an extension area of the base video stream. In order to prepare for display switching by the playback device in advance, it may be stored not in the video stream of the corresponding 2D playback section (B) but in the 3D playback section (A) in front of it.
- information indicating that there is no video frame may be stored in the final video frame of the 2D playback section (B) in which no extended video exists. For example, EndOfSequence.
- the playback device receives this signal during decoding, it knows that there is no extended video and can transition to 2D video playback.
- 2D video is stored in the base video stream and an extended video stream is also prepared, but it can be realized at a low bit rate such as black video instead of 2D video
- a video for displaying a message prompting 2D playback may be stored, and a 3D playback information descriptor may be stored in supplementary data or an extended area of the extended video stream.
- the playback device refers to the 3D playback information descriptor stored in the extended video stream and, when it can be determined that the playback is 2D playback, plays back 2D video using only the base video stream.
- the playback device cannot process the 3D playback information descriptor, a message prompting the user to perform 2D playback is displayed, so that the user can be prompted to perform 2D video playback processing. Since the bit rate of the extended video in the 2D playback section can be low, the bit rate can be assigned to the base video accordingly.
- the playback method of the 3D playback information descriptor is 2D
- duplicate the frame in the corresponding 2D playback section so that it has the same format (frame rate, etc.) as 3D.
- 2D frames may be doubled for playback.
- the PMT packet descriptor stores information indicating which video stream is paired with 3D video.
- the left-eye video has a PID of 0x1011 and the right-eye video has a PID of 0x1015.
- the PID of the opposite view may be described in the stream descriptor corresponding to the video stream. For example, in the example of FIG.
- 0x1015 is stored as the PID of the right-eye video stream in the stream descriptor corresponding to the video stream of the left-eye video, and the PID of the left-eye video stream is stored in the stream descriptor corresponding to the video stream of the right-eye video. Is stored as 0x1011.
- the PIDs of the other party that becomes the pair may be stored in the descriptor added to each video stream.
- a descriptor for identifying the left and right pairs a hierarchy descriptor defined in the MPEG-2 system standard may be used. In this case, a new hierarchy type may be prepared.
- the picture type is restricted as shown in the lower part of FIG. 21 in order to improve the efficiency of special playback such as fast-forwarding in 3D playback.
- the video access unit of the base video stream is an I picture
- the video access unit of an extended video stream having the same PTS is also configured by an I picture.
- the video access unit of the base video stream is a P picture
- the video access unit of the extended video stream having the same PTS is also composed of the P picture.
- the upper part of FIG. 21 shows a case where the restriction is not applied.
- the base video access unit is a P picture (P3).
- the video access unit of the extended video at the same time is a B picture (B3), in this case, the extended video has to be decoded up to the previously existing P picture (P2), which increases the load.
- B # NStart is a GOP # N head base video TS packet
- E # NStart is a GOP # N head extension video TS packet
- B # N + 1 Start is a GOP # N GOP # N TS packet.
- N + 1 is the first base video TS packet
- E # NEnd is a GOP # N-terminated extended video TS packet.
- the TS packet of the GOP # N first base video is configured to come before the GOP # N first extended video TS packet, and the GOP # N + 1 first base video is configured.
- the TS packet is configured to come after the GOP # N-terminated extended video TS packet. In this way, dive playback and editing can be performed in base video units.
- the extended video indicates the left-eye or right-eye video, but may be a depth map that visualizes the depth of the 2D video.
- the 3D playback method may be specified by a descriptor.
- the data creation apparatus includes a video encoder 2301, a multiplexing processing unit 2302, and a data storage method determination unit 2303.
- the data storage method determination unit 2303 specifies the data format of the transport stream to be created. For example, in the case of the video format shown in the example of FIG. 14, PTS 180000 to 5580000 are designated as Side-by-Side playback sections, PTS 5580000 to 109800000 are designated as 2D playback sections, and PTS 109800000 and later are designated as TopBottom playback sections.
- the playback system information, time information, and frame packing information storage type are designated to the video encoder 2301.
- the video encoder 2301 uses an image image such as an uncompressed bitmap of the left-eye image and an image image such as an uncompressed bitmap of the right-eye image according to a compression method such as MPEG4-AVC or MPEG2, and the data storage method determination unit 2303 Encode as specified.
- a compression method such as MPEG4-AVC or MPEG2
- the data storage method determination unit 2303 designates “Side-by-Side 3D video”
- the full HD left-eye video image and the full HD right-eye video image are down-converted to half HD respectively.
- compression coding is performed.
- image data compression encoding of full HD 2D video is performed.
- the data storage method determination unit 2303 designates “TopBottom 3D video”
- the full HD left-eye video image and the full HD right-eye video image are down-converted to half HD, respectively.
- the image is stored in one frame up and down by the side-by-side method, and then compression coding is performed.
- frame packing information according to each method is stored in the supplemental data.
- the storage method follows the frame packing information storage type specified from the data storage method determination unit 2303.
- the compressed video stream is output as a video stream.
- the multiplexing processing unit 1702 performs multiplexing according to the instruction of the data storage method determination unit 2303 together with the video stream output from the video encoder 2301, audio, subtitles, and the like, and creates and outputs a transport stream.
- the data storage method determination unit 1703 designates “Side-by-Side 3D video”
- the data storage method determination unit 1703 performs multiplexing on the transport stream, and in accordance with the video format described in the present embodiment,
- the “frame packing information descriptor” is stored in and a transport stream is output.
- Broadcasting of 3D programs is performed by multiplexing a plurality of video streams into one transport stream and supplying the transport stream to a television display device of each household by a broadcasting station.
- various patterns exist for combinations of video streams stored in the transport stream.
- the descriptor according to the present embodiment can realize 2D playback, compatible playback of 3D playback, and seamless switching between 2D playback and 3D playback in such various stream configurations.
- FIG. 25 is a diagram showing a transport stream configuration (2D / L + R) for storing the video of the right eye (R) video in addition to the video used as the left eye (L) video during 2D playback and 3D playback. is there.
- the transport stream stores a video stream (base video stream) used as video for the left eye during 2D playback and 3D playback, and a video stream for the right eye (extended video stream # 1). ing.
- the base video stream and the extended video stream each define a unique stream type for the PMT. Also, the base video stream is compressed and encoded by MPEG-2, and the extended video stream is encoded and encoded by AVC.
- 2D playback is performed using a 2D / L video stream.
- 3D playback is performed using a 2D / L video stream and an R video stream.
- a transport stream configuration (2D +) that stores two left-eye (L) videos and two right-eye (R) videos separately from 2D video L + R).
- FIG. 26 is a diagram showing a stream configuration of a 2D + L + R transport stream.
- the transport stream includes a 2D video stream (base video stream), a left-eye video stream (extended video stream # 1), and a right-eye video stream (extended video stream # 2). Storing.
- the base video stream is moving image compression encoded by MPEG-2
- the extended video stream is moving image compression encoded by AVC.
- 2D playback is performed using a 2D video stream.
- 3D playback is performed using a left-eye video stream and a right-eye video stream.
- the playback device identifies the video streams that make up 2D video and 3D video, enabling 2D playback, compatible playback of 3D playback, and seamless switching between 2D playback and 3D playback. Therefore, in the present embodiment, the descriptor shown below is stored in the transport stream.
- This descriptor includes 3D_system_info_descriptor that notifies the 3D system, 3D_service_info_descriptor that is supplementary information for realizing 3D playback, and 3D_combi_info_descriptor that indicates the correspondence between video streams used for 2D playback and 3D playback.
- 3D_system_info_descriptor is stored in a descriptor field (program loop) following the program information length (program_info_length) field in the PMT packet. That is, it is stored in descriptors # 1 to #N in FIG.
- 3D_system_info_descriptor indicates the 3D system provided by the transport stream. Specifically, any one of 2D playback, 3D playback using a frame compatible system, or 3D playback using a service compatible system is shown. Also, 3D_system_info_descriptor indicates whether or not a video stream used for 2D playback and a video stream used for 3D playback are shared in the case of 3D playback using a frame compatible method or 3D playback using a service compatible method.
- FIG. 27 shows the structure of 3D_system_info_descriptor.
- 3D_playback_type is an identifier indicating a playback method provided by the transport stream.
- FIG. 28 is a diagram illustrating values set in 3D_playback_type. As shown in this figure, when the value is 0, it indicates that 2D playback is provided by this transport stream, and when the value is 01, it indicates that 3D playback by the frame compatible method is provided. A value of 10 indicates that 3D playback by a service compatible method is provided. In the case of 2D + L + R or 2D / L + R stream configuration, a value of 10 is set.
- the playback device can identify the playback method provided by the transport stream by referring to 3D_playback_type.
- 2D_independent_flag is an identifier indicating whether a video stream used for 2D playback and a video stream used for 3D playback are shared.
- a value of 0 indicates that a video stream used for 2D playback and a video stream used for 3D playback are shared.
- a value of 1 indicates that a video stream used for 2D playback and a video stream used for 3D playback exist independently.
- a value of 0 is set.
- a value of 1 is set.
- the playback device refers to the 2D_independent_flag so that 2D playback is performed in the case of 3D playback by the frame compatibility method or 3D playback by the service compatibility method (when the value set to 3D_playback_type is 01 or 10). It is possible to identify whether the video stream used for the video stream and the video stream used for 3D playback are shared.
- 2D_view_flag is an identifier indicating which of 3D video streams is used for 2D playback.
- 2D_view_flag indicates which of the left-eye image and the right-eye image is used for 2D playback.
- 2D_view_flag indicates which of the base video stream and the extended video stream is used for 2D playback.
- 3D_service_info_descriptor is stored in a descriptor field (ES loop) following the ES information length (ES_info_length) field in the PMT packet. That is, it is stored in stream descriptors # 1 to #N in FIG.
- 3D_service_info_descriptor indicates supplementary information for realizing 3D playback. Specifically, it indicates whether the video stream is a left-eye video or a right-eye video. However, 3D_service_info_descriptor is not stored for a video stream used only for 2D playback. This is because the video stream is not used for 3D playback.
- FIG. 29 shows the structure of 3D_service_info_descriptor.
- Is_base_video is an identifier indicating whether the video stream is a base video stream or an extended video stream. A value of 1 indicates that the video stream is a base video stream. A value of 0 indicates that the video stream is an extended video stream.
- Leftview_flag is an identifier indicating whether the video stream is a left-eye video or a right-eye video. A value of 1 indicates that the video stream is a left-eye video. A value of 0 indicates that the video stream is a right-eye video.
- the playback device can identify whether the video stream is to be output as the left-eye or right-eye viewpoint video when displayed on a television as a 3D video.
- leftview_flag is provided in both cases where the video stream is a base video stream and an extended video stream.
- 3D_combi_info_descriptor is stored in a descriptor field (program loop) following the program information length (program_info_length) field in the PMT packet. That is, it is stored in descriptors # 1 to #N in FIG.
- 3D_combi_info_descriptor indicates the correspondence between video streams used for 2D playback and 3D playback. Specifically, it indicates the PID of the video stream that constitutes the transport stream.
- FIG. 30 shows the structure of 3D_combi_info_descriptor.
- 2D_view_PID / tag indicates the PID of the video stream used for 2D playback.
- Left_view_PID / tag indicates the PID of the video stream of the left-eye video.
- Light_view_PID / tag indicates the PID of the video stream of the video for the right eye.
- the playback device can specify a pair of video streams used for 3D playback and a video stream used for 2D playback by referring to these descriptors. Since the packet identifier that should be used for demultiplexing in each of 2D mode / 3D mode is described in one descriptor, the switching of the demultiplexing target between 2D mode and 3D mode can be made faster, Seamless switching between 2D and 3D playback is possible.
- FIG. 31 is a diagram showing a stream structure of a 2D + Side-by-Side transport stream.
- the transport stream stores a 2D video stream (base video stream) and a side-by-side video stream (extended video stream # 1).
- the base video stream is moving image compression encoded by MPEG-2
- the extended video stream is moving image compression encoded by AVC.
- the above transport stream includes 3D_system_info_descriptor that notifies the 3D system, 3D_service_info_descriptor that is supplementary information for realizing 3D playback, video stream used for 2D playback and 3D playback 3D_combi_info_descriptor indicating the corresponding relationship is stored.
- 2D playback is performed using a 2D base video stream in 2D TV or 2D mode of 3D TV.
- 3D playback is performed using the side-by-side extended video stream # 1.
- 3D_system_info_descriptor has the same structure as that in the case of the 2D + L + ⁇ R stream configuration shown in FIG.
- the playback device can identify the playback method provided by the transport stream by referring to this descriptor.
- FIG. 32 shows the structure of 3D_service_info_descriptor.
- frame_packing_arrangement_type is provided.
- “Frame_packing_arrangement_type” is an identifier indicating whether or not the video stream is a Side-by-Side format. A value of 1 indicates that the video stream is a side-by-side format. A value of 0 indicates the TopBottom method.
- the playback device can identify whether or not the extended video stream is a Side-by-Side format by referring to this identifier, and can perform 3D playback corresponding to the storage mode.
- frame_packing_arrangement_type in the case of the Side-by-Side method and the case of the TopBottom method has been set, but in addition, the left-eye video is interleaved on the odd lines and the right-eye video is interleaved on the even lines.
- the value in the checkerboard method in which the arranged line-by-line method and the left and right images are combined and the images are stored alternately in the vertical and horizontal directions like a checkerboard pattern in one frame may be set in frame_packing_arrangement_type.
- 3D_service_info_descriptor is not stored.
- FIG. 33 is a diagram showing the structure of 3D_combi_info_descriptor.
- 2D_view_PID / tag indicates the PID of the video stream used for 2D playback.
- “Frame_compatible_3D_PID / tag” indicates the PID of a video stream in a frame compatible format.
- the playback device can specify the frame compatible video stream used for 3D playback and the video stream used for 2D playback by referring to these descriptors. This enables seamless switching between 2D and 3D playback.
- FIG. 34 is a diagram showing the structure of a 2D + MVC transport stream.
- the transport stream includes a 2D video stream (base video stream), an MVC base view stream (extended video # 1), and an MVC dependent view stream (extended video # 2). Storing.
- the base video stream is compressed and encoded with moving picture by MPEG-2
- the extended video stream # 1 and extended video stream # 2 are compressed and encoded with moving picture by MVC.
- the above transport stream includes 3D_system_info_descriptor that notifies the 3D system, 3D_service_info_descriptor that is supplementary information for realizing 3D playback, video stream used for 2D playback and 3D playback 3D_combi_info_descriptor indicating the corresponding relationship is stored.
- a playback device such as a TV refers to these descriptors, and performs 2D playback using a 2D base video stream in the 2D TV or 2D mode of 3D TV.
- 3D playback is performed using the extended video stream # 1 and the extended video stream # 2 compression-coded by MVC.
- 3D_system_info_descriptor and 3D_service_info_descriptor have the same structure as that in the case of the 2D + L + R stream configuration shown in FIGS. 27 and 29 and will not be described. Note that 3D_service_info_descriptor is not stored for a video stream used only for 2D playback, as in the case of the stream configuration of 2D + L + R.
- FIG. 35 shows the structure of 3D_combi_info_descriptor.
- 2D_view_PID / tag indicates the PID of the video stream used for 2D playback.
- MVC_base_view_PID / tag indicates the PID of the MVC base view stream.
- MVC_dept_view_PID / tag indicates the PID of the MVC dependent view stream.
- the playback device can identify a pair of MVC video streams used for 3D playback and a video stream used for 2D playback. This enables seamless switching between 2D and 3D playback.
- FIG. 36 is a diagram showing the stream structure of the 2D + R1 + R2 transport stream.
- the transport stream includes a video stream (base video stream) used as L video during 2D playback and 3D playback, a video stream of first R video (extended video stream # 1), And a video stream of the second R video.
- the base video stream is video compression encoded by MPEG-2
- the extended video stream # 1 and the extended video stream # 2 are video compression encoded by AVC.
- the transport stream stores 3D_system_info_descriptor for notifying the 3D system, 3D_service_info_descriptor which is supplementary information for realizing 3D playback, and 3D_combi_info_descriptor indicating the correspondence between video streams used for 2D playback and 3D playback.
- a playback device such as a TV refers to these descriptors and performs 2D playback using a base video stream in 2D TV or 2D mode of 3D TV.
- 3D playback is performed using the base video stream and the extended video stream # 1, or the base video stream and the extended video stream # 2.
- FIG. 37 shows the structure of 3D_system_info_descriptor.
- camera_assingment_type is provided instead of 2D_independent_flag.
- camera_assingment_type is an identifier indicating the camera arrangement type for the video stream stored in the transport stream.
- a value of 1 indicates that the transport stream is composed of a video stream from the camera viewpoint of the center (C).
- a value of 2 indicates that the transport stream is composed of left-view (L) and right-view (R) camera-view video streams.
- a value of 3 indicates that the transport stream is composed of a video stream of the camera viewpoint of the center (C), the left viewpoint (L), and the right viewpoint (R).
- a value of 4 indicates that the transport stream is composed of video streams of camera viewpoints of the left viewpoint (L), the first right viewpoint (R1), and the second right viewpoint (R2).
- the playback device can identify the camera assignment of the video stream constituting the transport stream.
- FIG. 38 is a diagram illustrating the structure of 3D_service_info_descriptor. In the descriptor in the 2D + L + R stream configuration shown in FIG. 31, camera_assingment is newly provided.
- Camera_assingment is an identifier indicating camera position information such as the left eye, center, and right eye of the video stream.
- the playback device can identify the camera arrangement corresponding to the video stream by referring to this identifier.
- FIG. 39 shows the structure of 3D_combi_info_descriptor.
- 2D_view_PID / tag indicates the PID of the video stream used as the L video during 2D playback and 3D playback.
- Right1_view_PID / tag indicates the PID of the video stream of the first R video.
- Light2_view_PID / tag indicates the PID of the video stream of the second R video.
- the playback device can specify a video stream used as an L video during 2D playback and 3D playback, and a plurality of R viewpoint video streams. This enables seamless switching between 2D and 3D playback.
- the playback device identifies the stream configuration of the transport stream and can perform seamless switching between 2D playback and 3D playback.
- An extended video stream that is compression-encoded by a moving image compression encoding technique other than AVC may be stored.
- the extended video stream may be compression-encoded using H.265, which is a next-generation moving image compression encoding technique.
- the stream descriptor corresponding to the L and R video streams includes the video stream of the viewpoint that is the other side of the combination in 3D playback. PID may be described.
- closed caption (Closed Caption) caption data when closed caption (Closed Caption) caption data is included in the base stream and the extended stream, an identifier indicating which data is used in 2D playback or 3D playback may be stored in the PMT of the transport stream.
- the playback device can identify the closed caption data to be used in 2D playback or 3D playback by referring to this identifier.
- 3D_system_info_descriptor, 3D_service_info_descriptor, and 3D_combi_info_descriptor are stored in the PMT packet, the storage position is not limited to this. It may be stored in any area of the transport stream. For example, it may be stored in supplementary data of each video stream other than the PMT packet.
- the video stream used for 2D playback and the video stream used for 3D playback are specified by setting the PID of the video stream in 3D_combi_info_descriptor.
- the present invention is not limited to this.
- the 3D_combi_info_descriptor only needs to include information that can specify the video stream to be multiplexed.
- each video stream may be specified by setting a hierarchy descriptor defined in the MPEG-2 system standard. Specifically, by defining a new hierarchy_type in hierarchy_descriptor and specifying a video stream from 3D_combi_info_descriptor using hierarchy_layer_index, each video stream pair used for 3D playback and each video stream used for 2D playback Identify the stream.
- FIG. 40 is a diagram showing an internal configuration of the data creation device 4000 according to the present embodiment.
- the data creation device 4000 includes a video encoder 4001, a multiplexing processing unit 4002, a data storage method determination unit 4003, and a user interface unit 4004.
- the user interface unit 4004 provides the data creator with data input using a keyboard, mouse, other controller, and the like.
- the data creator uses the user interface unit 4004 to specify the stream configuration pattern and compression encoding method of the transport stream to be created.
- the data storage method determination unit 4003 determines the stream configuration of the transport stream and the compression encoding method of the video stream according to the user designation by the user interface unit 4004.
- the video encoder 4001 compresses and encodes a 3D video original image using a compression encoding method such as MPEG-2, AVC, MVC, or H.265, thereby converting the video stream specified by the data storage method determination unit 4003 into a video stream. create.
- a compression encoding method such as MPEG-2, AVC, MVC, or H.265
- the multiplexing processing unit 4002 creates 3D_system_info_descriptor, 3D_service_info_descriptor, and 3D_combi_info_descriptor descriptors according to the stream structure of the transport stream to be created in accordance with the instruction of the data storage method determination unit 3903. Then, a transport stream is created by multiplexing each descriptor together with a stream such as a video stream, audio, and caption output from the video encoder 4001 in accordance with an instruction from the data storage method determination unit 4003.
- the created transport stream is recorded on an external recording medium.
- the data transmission apparatus is described above with respect to the configuration in which data is transmitted via a broadcast or network by an external transmission unit. Next, the operation of this data creation device will be described.
- FIG. 41 is a flowchart showing the flow of the encoding process of the data creation device 4000.
- the data storage method determination unit 4003 determines the pattern of the video stream constituting the transport stream (step S4101). Specifically, the stream configuration of the transport stream and the compression encoding method of the video stream stored in the transport stream are determined.
- the stream configuration patterns shown in FIG. 25, FIG. 26, FIG. 31, FIG. 34, and FIG. 37 can be considered, but the stream configuration (2D / SBS) that stores only the side-by-side video stream, and others
- the stream configuration may be as follows.
- the video encoder 4001 compresses and encodes the 3D original image to create a video stream (step S4102). At this time, the video encoder 4001 determines a compression encoding method based on the designation of the stream configuration of the transport stream and the compression encoding method of the video stream stored in the transport stream by the data storage method determination unit 4003, and the 3D Perform compression encoding of the original video.
- the multiplexing processing unit 4002 stores the video stream in each frame based on the stream configuration of the transport stream specified by the data storage method determination unit 4003 (step S4103).
- the multiplexing processing unit 4002 creates 3D_system_info_descriptor, 3D_service_info_descriptor, and 3D_combi_info_descriptor descriptors and stores them in the PMT of the transport stream (step S4104). At this time, the multiplexing processing unit 4002 creates each descriptor based on the stream configuration of the transport stream designated by the data storage method determining unit 4003.
- FIG. 42 is a diagram showing an internal configuration of the 3D digital television 4200 according to the present embodiment.
- the 3D digital television 4200 includes a tuner 4201, a NIC 4202, a user interface unit 4203, a mode storage unit 4204, a demultiplexing unit 4205, a display determination unit 4206, a video decoding unit 4207, a frame buffer (1) 4208.
- the tuner 4201 has a function of receiving a digital broadcast wave transport stream and demodulating the received signal.
- the network interface card (NIC) 4202 is connected to the IP network, and has a function of receiving a transport stream input from the outside.
- the user interface unit 4203 accepts a channel selection operation, a 2D mode / 3D mode selection operation, and the like from the user.
- the mode storage unit 4204 stores a flag indicating whether the current display mode is the 2D mode or the 3D mode.
- the demultiplexing unit 4205 has a function of separating the received transport stream into a video stream and other audio streams, graphics streams, and the like, and outputting the video stream to the video decoding unit 4207.
- the demultiplexing unit 4205 extracts system packets such as PSI from the received transport stream, acquires 3D_system_info_descriptor, 3D_service_info_descriptor, and 3D_combi_info_descriptor descriptors from the PMT packet, and notifies the display determining unit 4206 of them.
- the designation of the TS packet PID to be demuxed in the current display mode is received from the display determination unit 4206.
- the demultiplexing unit 4205 acquires a video stream by demultiplexing a TS packet of a designated PID.
- the demultiplexing unit 4205 can read the transport stream from the recording medium in addition to reading the transport stream from the tuner 4201 and the NIC 4202.
- the display determination unit 4206 refers to each descriptor of 3D_system_info_descriptor, 3D_service_info_descriptor, and 3D_combi_info_descriptor notified from the demultiplexing unit 4205 to grasp the stream configuration of the transport stream.
- the PID of the TS packet to be demultiplexed is notified to the demultiplexing unit 4205.
- the display determination unit 4206 refers to the 2D_view_flag of the 3D_system_info_descriptor and the frame_packing_arrangement_type of the 3D_service_info_descriptor to determine whether the left eye image or the right eye image is displayed. Is used for 2D playback, or whether the video stream is a Side-by-Side format.
- the video decoding unit 4207 Upon receiving the video stream from the demultiplexing unit 4205, the video decoding unit 4207 decodes the received stream. The video decoding unit 4207 writes the decoded frame to the frame buffer (1) 4208.
- Frame buffer (1) 4108 has an area for storing the frame decoded by the video decoding unit 4207.
- the display processing unit 4209 executes the cropping control according to the cropping information and the scaling control according to the scaling information when the video stream stored in the frame buffer (1) 4208 is in the Side-by-Side format.
- the left-eye frame and the right-eye frame obtained by the cropping process are stored in the frame buffer (L) and the frame buffer (R).
- the display processing unit 4209 converts the video stream into a frame buffer (L) 4210, a frame buffer, (R) Sort to 4212.
- the frame buffer (L) 4210 and the frame buffer (R) 4212 have areas for storing frames output from the display processing unit 4209.
- the switch 4211 selects a frame image written in the frame buffer (L) 4210 and the frame buffer (R) 4212 and transfers it to the display unit.
- the display unit 4213 displays the frame transferred from the switch 4211.
- the display unit 4213 communicates with the 3D glasses, and when the left eye image is displayed, the left eye side of the 3D glasses is opened, and when the right eye image is displayed, the liquid crystal shutter of the 3D glasses is opened so that the right eye side of the 3D glasses is opened. Control. Note that when displaying 2D video, control of 3D glasses is not performed.
- FIG. 43 is a flowchart showing an exemplary flow of a program playback process performed by the 3D digital television 4200.
- the demultiplexing unit 4205 analyzes the PMT of the transport stream and extracts a descriptor (step S4301).
- the display determination unit 4206 refers to the 3D_playback_type of the 3D_system_info_descriptor extracted by the demultiplexing unit 4205 to determine the playback method of the received transport stream (step S4302).
- the display determination unit 4206 refers to 2D_independent_flag of 3D_system_info_descriptor, and whether or not the video stream used for 2D playback and the video stream used for 3D playback are shared. Is determined (step S4303).
- the display determination unit 4206 refers to the 3D_combi_info_descriptor and identifies the stream configuration (step S4304).
- the 3D digital television 4200 performs the processing of the 2D / L + R1 + R2 stream described later (step S4306).
- the 3D digital television 4200 performs processing of the 2D / L + R stream described later (step S4307).
- the display determination unit 4206 refers to the 3D_combi_info_descriptor and identifies the stream configuration (step S4308).
- the 3D digital television 4200 transmits the MPEG2 + MVC (Base) + MVC (Dependent) stream described later. Processing is performed (step S4311).
- step S4309 When the stream configuration of the transport stream is MPEG2 + AVC + AVC (step S4309, YES), the 3D digital television 4200 performs MPEG2 + AVC + AVC stream processing described later (step S4312).
- the display determination unit 4206 refers to 2D_independent_flag of 3D_system_info_descriptor, and whether or not the video stream used for 2D playback and the video stream used for 3D playback are shared. Is determined (step S4313).
- step S4313 If the value of 2D_independent_flag is 0 (step S4313, NO), the 3D digital television 4200 performs 2D / SBS stream processing (to be described later) (step S4314).
- step S4313 When the value of 2D_independent_flag is 1 (step S4313, YES), the 3D digital television 4200 performs processing of a 2D + SBS stream to be described later (step S4315).
- FIG. 44 is a flowchart showing the flow of processing of a 2D + SBS stream.
- the display determination unit 4206 determines whether the current mode is the 2D mode or the 3D mode with reference to the flag of the mode storage unit 4204 (step S4401).
- the display determination unit 4206 separates the TS packet indicated by 2D_view_PID / tag of 3D_combi_info_descriptor, and extracts a 2D video stream (step S4402).
- the 3D digital television 4200 performs 2D playback by decoding the extracted MPEG2 (2D) video stream by the video decoding unit 4207 and outputting the video signal to the display unit 4213 (step S4403).
- the display determination unit 4206 separates the TS packet indicated by frame_compatible_3D_PID / tag of 3D_combi_info_descriptor and extracts a video stream (step S4404).
- the display determination unit 4206 refers to the frame_packing_arrangement_type of 3D_service_info_descriptor and determines whether it is stored in the Side-by-Side format (step S4405).
- the display processing unit 4209 performs 3D playback by cropping out the left-eye image and the right-eye image that exist on the left and right (step S4406).
- step S4305 When the frame_packing_arrangement_type is not the Side-by-Side format (step S4305, NO), the TopBottom method is specified, and the display processing unit 4209 performs 3D playback by cropping out the left-eye image and the right-eye image existing above and below. This is performed (step S4407).
- step S4315 This completes the description of the details of the processing of the 2D + BS SBS stream in step S4315. Next, details of the 2D / SBS stream processing in step S4314 will be described.
- FIG. 45 is a flowchart showing the flow of processing of a 2D / SBS stream.
- the demultiplexing unit 4205 separates the TS packet indicated by frame_compatible_3D_PID / tag of 3D_combi_info_descriptor, and extracts a 2D / SBS video stream (step S4501).
- the display determination unit 4206 refers to the flag of the mode storage unit 4204 to determine whether the current mode is the 2D mode or the 3D mode (step S4502).
- the display determination unit 4206 refers to the 2D_view_flag of 3D_system_info_descriptor_ and determines which of the side-by-side frames is used for 2D playback.
- step S4503 When 2D_view_flag indicates a left-eye image (step S4503, YES), the display processing unit 4209 performs 2D playback by cropping out the left-eye image region in the side-by-side frame (step S4505). .
- step S4503 When 2D_view_flag indicates a right-eye image (step S4503, NO), the display processing unit 4209 performs 2D playback by cropping out the region of the right-eye image in the side-by-side frame (step S4504). .
- the display processing unit 4209 crops out the image area for the right eye in the side-by-side frame (step S4506), and the side-by-side frame. Among them, the area of the image for the left eye is cropped out (step S4507).
- the 3D digital television 4200 performs 3D reproduction by alternately outputting the cropped left-eye image and right-eye image to the display unit 4213 (step S4508).
- FIG. 46 is a flowchart showing a processing flow of 2D / L / + R stream. As shown in this figure, the display determination unit 4206 determines whether the current mode is the 2D mode or the 3D mode with reference to the flag of the mode storage unit 4204 (step S4601).
- the display determination unit 4206 separates the TS packet indicated by Left_view_PID / tag of 3D_combi_info_descriptor and the TS packet indicated by Right_view_PID / tag, and 2D / L video stream, R Are extracted (step S4602).
- the 3D digital television 4200 performs 3D playback by decoding the extracted 2D / L video stream and R video stream by the video decoding unit 4207 and outputting the video signal to the display unit 4213 (step S4603).
- the demultiplexing unit 4205 separates the TS packet indicated by 2D_view_PID / tag of 3D_combi_info_descriptor, and extracts a 2D / L video stream (step S4604).
- the 3D digital television 4200 performs 2D playback by decoding the extracted 2D / L video stream by the video decoding unit 4207 and outputting the video signal to the display unit 4213 (step S4605).
- FIG. 47 is a flowchart showing a flow of processing of 2D / L + R1 + R2. As shown in this figure, the display determination unit 4206 determines whether the current mode is the 2D mode or the 3D mode with reference to the flag of the mode storage unit 4204 (step S4601).
- the display determination unit 4206 separates the TS packet indicated by Left_view_PID / tag of 3D_combi_info_descriptor, the TS packet indicated by Right1_view_PID / tag, and the TS packet indicated by Right2_view_PID / tag. Then, the 2D / L video stream, the R1 video stream, and the R2 video stream are extracted (step S4701).
- the 3D digital television 4200 performs 3D playback by decoding the extracted 2D / L video stream, the R1 video stream, or the R2 video stream by the video decoding unit 4207 and outputting the video signal to the display unit 4213. (Step S4702).
- Steps S4604 and 4605 are the same processing as the processing of the 2D / L + R processing shown in FIG. 46, and will not be described.
- step S4306 The above is the details of the processing of the 2D / L + R1 + R2 stream in step S4306. Next, details of the MPEG2 + AVC + AVC stream processing in step 4312 will be described.
- FIG. 48 is a flowchart showing the flow of MPEG2 + AVC + AVC stream processing.
- the display determination unit 4206 determines whether the current mode is the 2D mode or the 3D mode with reference to the flag of the mode storage unit 4204 (step S4801).
- the display determination unit 4206 separates the TS packet indicated by 2D_view_PID / tag of 3D_combi_info_descriptor, and extracts an MPEG2 (2D) video stream (step S4802).
- the 3D digital television 4200 performs 2D playback by decoding the extracted MPEG2 (2D) video stream by the video decoding unit 4207 and outputting the video signal to the display unit 4213 (step S4803).
- the display determination unit 4206 separates the TS packet indicated by Left_view_PID / tag of 3D_combi_info_descriptor and the TS packet indicated by Right_view_PID / tag, and generates a video stream for the left eye and a video stream for the right eye. Extract (step S4804).
- the 3D digital television 4200 performs 3D playback by decoding the extracted left-eye video stream and right-eye video stream by the video decoding unit 4207 and outputting the video signal to the display unit 4213 (step S4805).
- FIG. 49 is a flowchart showing the flow of MPEG2 + MVC (Base) + MVC (Dependent) stream processing.
- the display determination unit 4206 determines whether the current mode is the 2D mode or the 3D mode with reference to the flag of the mode storage unit 4204 (step S4901).
- the display determination unit 4206 separates the TS packet indicated by 2D_view_PID / tag of 3D_combi_info_descriptor, and extracts an MPEG2 (2D) video stream (step S4902).
- the 3D digital television 4200 performs 2D playback by decoding the extracted MPEG2 (2D) video stream by the video decoding unit 4207 and outputting the video signal to the display unit 4213 (step S4903).
- the display determination unit 4206 separates the TS packet indicated by MVC_base_view_PID / tag of 3D_combi_info_descriptor and the TS packet indicated by MVC_dept_view_PID / tag, and extracts the base view stream and the dependent view stream. (Step S4904).
- the extracted base view stream and dependent view stream are decoded by the video decoding unit 4207, and the video signal is output to the display unit 4213, whereby the 3D digital television 4200 performs 3D playback (step S4905).
- the stream structure of the transport stream can be specified by referring to the descriptor multiplexed in the transport stream, so that 2D playback, compatible playback of 3D playback, and 2D Playback and 3D playback can be switched seamlessly.
- the present invention may be an application execution method disclosed by the processing procedure described in each embodiment. Further, the present invention may be a computer program including program code that causes a computer to operate according to the processing procedure.
- the present invention can also be implemented as an LSI that controls the image processing apparatus described in each of the above embodiments.
- Such an LSI can be realized by integrating each functional block. These functional blocks may be individually made into one chip, or may be made into one chip so as to include a part or all of them.
- LSI is used, but depending on the degree of integration, it may be called IC, system LSI, super LSI, or ultra LSI.
- the method of circuit integration is not limited to LSI, and implementation with a dedicated circuit or a general-purpose processor is also possible.
- An FPGA Field Programmable Gate Array
- a reconfigurable processor that can reconfigure the connection and setting of circuit cells inside the LSI may be used.
- the encoding method of the present invention it is possible to store a video stream constituting 2D video and a descriptor specifying the video stream constituting 3D video in the transport stream.
- the stream structure of the transport stream can be specified, so that 2D playback, compatible playback of 3D playback, and seamless switching between 2D playback and 3D playback can be performed. Be beneficial.
- Playback apparatus 100 Playback apparatus 200 3D glasses 300 2D digital television 501 Video frame sequence 502 Video PES packet 503 Video TS packet 504 Audio frame sequence 505 Audio PES packet 506 Audio TS packet 507 Subtitle stream 508 Subtitle stream PES packet 509 Subtitle stream TS packet 513 Transport stream 1501 Playback device tuner 1502 Playback device NIC 1503 Demultiplexing unit of playback device 1504 Video decoding unit of playback device 1505 Display determination unit of playback device 1506 Display processing unit of playback device 1507 Display unit of playback device 1508 Frame buffer (1) of playback device 1510 Playback device frame buffer (L) 1511 Playback device frame buffer (R) 1512 Playback Device Switch 2301 Video Encoder 2302 Multiplexing Processing Unit 2303 Data Storage Method Determination Unit 4000 Data Creation Device 4001 Video Encoder 4002 Multiplexing Processing Unit 4003 Data Storage Method Determination Unit 4004 User Interface Unit 4004 4200 3D Digital TV 4201 Tuner 4202 NIC 4203 user interface unit 4204
Abstract
Description
本実施の形態に係る映像フォーマットと、その映像フォーマットにおけるデータ作成方法、データ作成装置、再生方法、再生装置について説明する。
次に、本実施の形態に係る3D映像を格納するデータフォーマットについて、図面を参照しながら説明を行う。
次に本実施の携帯に係る3D映像を再生する再生装置の構成について図15を用いて説明する。
次に、本実施の形態に係る3D映像を格納するデータフォーマットの変形例について図面を参照しながら説明を行う。
次に、本実施の形態に係る3D映像を2本のビデオストリームで構成する場合のデータフォーマットについて図面を参照しながら説明を行う。
次に本実施の形態に係るデータ作成装置およびデータ作成方法について図23を参照しながら説明を行う。
実施の形態2として、上述のディスクリプタの具体的な形態について説明する。
が左視点(L)と右視点(R)のカメラ視点のビデオストリームから構成されることを示す。値が3の場合は、トランスポートストリームがセンター(C)、左視点(L)、および右視点(R)のカメラ視点のビデオストリームから構成されることを示す。値が4の場合は、トランスポートストリームが左視点(L)、第1右視点(R1)、および第2右視点(R2)のカメラ視点のビデオストリームから構成されることを示す。
おいて、camera_assingmentが新たに設けられている。
以上がデータ作成装置の構成についての説明である。続いて、このデータ作成装置の動作について説明する。
なお、上記の実施の形態に基づいて説明してきたが、本発明は上記の実施の形態に限定されないことはもちろんである。以下のような場合も本発明に含まれる。
200 3D眼鏡
300 2Dデジタルテレビ
501 ビデオフレーム列
502 ビデオのPESパケット
503 ビデオのTSパケット
504 オーディオフレーム列
505 オーディオのPESパケット
506 オーディオのTSパケット
507 字幕ストリーム
508 字幕ストリームのPESパケット
509 字幕ストリームのTSパケット
513 トランスポートストリーム
1501 再生装置のチューナ
1502 再生装置のNIC
1503 再生装置の多重分離部
1504 再生装置の映像デコード部
1505 再生装置の表示判定部
1506 再生装置の表示処理部
1507 再生装置の表示部
1508 再生装置のフレームバッファ(1)
1510 再生装置のフレームバッファ(L)
1511 再生装置のフレームバッファ(R)
1512 再生装置のスイッチ
2301 ビデオエンコーダ
2302 多重化処理部
2303 データ格納方法決定部
4000 データ作成装置
4001 ビデオエンコーダ
4002 多重化処理部
4003 データ格納方法決定部
4004 ユーザーインターフェイス部4004
4200 3Dデジタルテレビ
4201 チューナ
4202 NIC
4203 ユーザーインターフェイス部
4204 モード記憶部
4205 多重分離部
4206 表示判定部
4207 映像デコード部
4208 フレームバッファ(1)
4209 表示処理部4209、
4210 フレームバッファ(L)
4211 スイッチ
4212 フレームバッファ(R)
4213 表示部
Claims (19)
- 符号化方法であって、
画像を圧縮符号化してビデオストリームを生成するエンコードステップと、
前記エンコードステップによって生成される複数のビデオストリームに対して多重化処理を施すことで、1つのトランスポートストリームを得る多重化ステップとを有し、
前記複数のビデオストリームは、2D映像を構成する2Dビデオストリームを含むとともに、複数のビデオストリームを組み合わせることで3D映像を構成し、
3D映像を構成するビデオストリームの組み合わせには、前記2Dビデオストリームと1つの他のビデオストリームの組み合わせと、2Dビデオストリーム以外の2以上の他のビデオストリームの組み合わせがあり、
前記トランスポートストリームは、3D映像を構成するビデオストリームを特定する3D映像特定情報を含む
ことを特徴とする符号化方法。 - 前記符号化方法は、コンテンツテーブルを作成するコンテンツテーブル作成ステップを有し、
前記多重化ステップによる多重化は、
前記複数のビデオストリームとコンテンツテーブルとを多重化することでなされ、
前記コンテンツテーブルは、1つ以上のテーブルディスクリプタと、各ビデオストリームに対応するストリーム情報とを含み、
前記各ストリーム情報は、ストリームタイプと、ストリーム識別子と、ストリームディスクリプタとを含み、
前記3D映像特定情報は、テーブルディスクリプタ及びストリームディスクリプタの何れかに存在する
ことを特徴とする請求項1に記載の符号化方法。 - 前記3D映像特定情報は、2Dビデオストリームを特定する2D映像特定情報を含むことを特徴とする請求項1に記載の符号化方法。
- 2D映像特定情報及び3D映像特定情報によるビデオストリームの特定は、2Dビデオストリーム、左目映像を構成する左目ビデオストリーム、右目映像を構成するビデオストリームに、それぞれ対応するストリーム識別子を記載することでなされる
ことを特徴とする請求項3に記載の符号化方法。 - 前記コンテンツテーブルには2D/3D共用フラグを含み、
前記2D/3D共用フラグは、2Dビデオストリームと、3D映像を構成する複数のビデオストリームのいずれかが一致するか否かを示すことを特徴とする請求項2に記載の符号化方法。 - 前記3D特定情報によって指定される3D映像を構成するビデオストリームが1つの場合、
前記ビデオストリームは、LRパック映像を構成し、
前記LRパック映像は、動画像を構成する個々のフレームに、左目画像と、右目画像とを格納した映像であり、
前記コンテンツテーブルは、LRパック情報を含み、
LRパック情報は、左目画像及び右目画像が、どのようなパッキング方式でビデオストリームの各フレームに格納されているかを示す
ことを特徴とする請求項2に記載の符号化方法。 - 前記コンテンツテーブルは、カメラのチャネル構成を示すカメラアサインメントを含み、
前記カメラのチャネル構成には、
Cチャネル、Lチャネル+Rチャネル、Cチャネル+Lチャネル+Rチャネル、Cチャネル+R1チャネル+R2チャネルがあり
前記3D特定情報によって特定されるビデオストリームは、前記カメラのチャネル構成のうち、何れかのものに従って撮影されたものであるかを示す
ことを特徴とする請求項2に記載の符号化方法。 - 前記2Dビデオストリーム、他のビデオストリームのそれぞれには、文字表示制御情報が存在しており、
前記作成ステップは、
2D再生モード、3D再生モードのそれぞれにおいて、2Dビデオストリーム、他のビデオストリームのうち、どちらのビデオストリームに存在する文字表示制御情報を用いて文字表示を実現すべきかを示す情報をコンテンツテーブルに記述する
ことを特徴とする請求項2に記載の符号化方法。 - ストリーム情報のストリームディスクリプタは、
ビデオストリームが左目画像及び右目画像のうち、何れに該当するかを示すフラグを含むことを特徴とする請求項2に記載の符号化方法。 - 前記3D映像特定情報を、コンテンツテーブルにおけるストリーム情報内のストリームディスクリプタ内に記述する
ことを特徴とする請求項2に記載の符号化方法。 - 前記作成ステップは、
前記コンテンツテーブルにおける各ビデオストリームに対応するストリーム情報に、組合せの相手側となるビデオストリームのストリーム識別子を記述することで、組合せの対象となる2以上のビデオストリームを示す
ことを特徴とする請求項2に記載の符号化方法。 - 前記作成ステップは、
ディスクリプタを作成して、作成されたディスクリプタをビデオストリームの内部に組込む
ことを特徴とする請求項1に記載の符号化方法。 - 表示装置であって、
装置外部からのトランスポートストリームの入力を受け付ける受付手段と、
2Dモード及び3Dモードの何れか1つをカレントモードとして記憶する記憶手段と、
カレントモードが2Dモードである場合、トランスポートストリームに含まれる1つのビデオストリームを用いて2D映像の再生を行う再生手段とを備え、
トランスポートストリームは3D映像特定情報を備え、
3D映像特定情報は、3D映像を構成するビデオストリームを特定し、
カレントモードが3Dモードである場合、トランスポートストリームに含まれる3D映像を構成するビデオストリームを用いて3D映像の再生を行い、
3D映像の構成するビデオストリームは、前記2Dビデオストリームと1つの他のビデオストリームとの組み合わせからなる場合と、2Dビデオストリーム以外の他のビデオストリームの2以上組み合わせからなる場合とがある
ことを特徴とする表示装置。 - 前記トランスポートストリームは、複数のビデオストリームと、コンテンツテーブルとをトランスポートストリームパケット列に変換することで得られ、
前記表示装置は、
トランスポートストリームから所定のトランスポートストリームパケットを分離する多重分離手段を備え、コンテンツテーブルを格納したトランスポートストリームパケットを分離することで、3D映像特定情報の取得を行う
ことを特徴とする請求項13に記載の表示装置。 - 前記3D映像特定情報には、2Dビデオストリームを特定する2D映像特定情報を含み、
カレントモードが2Dモードである場合、多重分離部は、2D映像特定情報に基づき、2D映像を構成するビデオストリームをトランスポートストリームから分離し、
カレントモードが3Dモードである場合、多重分離部は、3D映像特定情報に基づき、3D映像を構成するビデオストリームを格納した複数のトランスポートストリームパケットをトランスポートストリームから分離する
ことを特徴とする請求項14に記載の表示装置。 - 2D映像特定情報及び3D映像特定情報によるビデオストリームの特定は、2Dビデオストリーム、左目映像を構成する左目ビデオストリーム、右目映像を構成するビデオストリームに、それぞれ対応するストリーム識別子を記載することでなされる
ことを特徴とする請求項15に記載の表示装置。 - トランスポートストリームは、2Dビデオストリームと、3D映像を構成する複数のビデオストリームのいずれかが一致するか否かを示す2D/3D共用フラグを含み、
前記多重分離部は、2D映像を構成するビデオストリームが複数のビデオストリームの何れかと一致しない場合、2Dモード、3Dモードのそれぞれにおいて、多重分離の対象となるビデオストリームを変更し、
再生手段は、3D映像を構成するビデオストリームが1つである場合、1つのビデオストリームにおける個々のフレームから、左目画像及び右目画像を切り出して、左目画像及び右目画像のそれぞれを表示に供することで3D画像の再生を行い、
3D映像を構成するビデオストリームが2つ以上である場合、多重分離部によって分離された2以上のビデオストリームをデコードすることで左目画像及び右目画像を得て、左目画像及び右目画像のそれぞれを表示に供することで3D画像の再生を行う
ことを特徴とする請求項16に記載の表示装置。 - 前記3D特定情報によって指定される3D映像を構成するビデオストリームが1つの場合、
前記ビデオストリームはLRパック映像を構成し、前記LRパック映像は、動画像を構成する個々のフレームに、左目画像と、右目画像とを格納した映像であり、
前記コンテンツテーブルは、LRパック情報を含み、
LRパック情報は、左目画像及び右目画像が、どのようなパッキング方式でビデオストリームの各フレームに格納されているかを示し、
再生手段は、LRパック情報に従って、フレームのうち、左目画像及び右目画像を切り出すべき領域を特定する
ことを特徴とする請求項17に記載の表示装置。 - 復号方法であって、
トランスポートストリームの入力を受け付ける受付ステップと、
2Dモード及び3Dモードの何れか1つをカレントモードとして記憶する記憶ステップと、
カレントモードが2Dモードである場合、トランスポートストリームに含まれる1つのビデオストリームを用いて2D映像の再生を行う再生ステップとを含み、
トランスポートストリームは3D映像特定情報を備え、
3D映像特定情報は、3D映像を構成するビデオストリームを特定し、
カレントモードが3Dモードである場合、トランスポートストリームに含まれる3D映像を構成するビデオストリームを用いて3D映像の再生を行い、
3D映像の構成するビデオストリームは、前記2Dビデオストリームと1つの他のビデオストリームとの組み合わせからなる場合と、2Dビデオストリーム以外の他のビデオストリームの2以上組み合わせからなる場合とがある
ことを特徴とする復号方法。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020137008616A KR20130135244A (ko) | 2010-10-25 | 2011-10-25 | 부호화방법, 표시장치, 복호방법 |
EP11836297.9A EP2635033A1 (en) | 2010-10-25 | 2011-10-25 | Encoding method, display device, decoding method |
CN2011800512729A CN103202023A (zh) | 2010-10-25 | 2011-10-25 | 编码方法、显示装置、解码方法 |
JP2012540892A JP5336666B2 (ja) | 2010-10-25 | 2011-10-25 | 符号化方法、表示装置、復号方法 |
MX2013004068A MX2013004068A (es) | 2010-10-25 | 2011-10-25 | Metodo de codificacion, dispositivo de visualizacion, metodo de codificacion. |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US40634710P | 2010-10-25 | 2010-10-25 | |
US61/406,347 | 2010-10-25 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012057164A1 true WO2012057164A1 (ja) | 2012-05-03 |
Family
ID=45993870
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2011/074586 WO2012057164A1 (ja) | 2010-10-25 | 2011-10-25 | 符号化方法、表示装置、復号方法 |
Country Status (8)
Country | Link |
---|---|
US (1) | US20120106921A1 (ja) |
EP (1) | EP2635033A1 (ja) |
JP (1) | JP5336666B2 (ja) |
KR (1) | KR20130135244A (ja) |
CN (1) | CN103202023A (ja) |
MX (1) | MX2013004068A (ja) |
TW (1) | TW201234833A (ja) |
WO (1) | WO2012057164A1 (ja) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014042034A1 (ja) * | 2012-09-12 | 2014-03-20 | ソニー株式会社 | 送信装置、送信方法、受信装置および受信方法 |
WO2014050447A1 (ja) * | 2012-09-27 | 2014-04-03 | ソニー株式会社 | 送信装置、送信方法、受信装置および受信方法 |
JP2016519480A (ja) * | 2014-02-21 | 2016-06-30 | エルジー エレクトロニクス インコーポレイティド | 3d放送信号を処理する方法及び装置 |
JP2017509278A (ja) * | 2014-02-10 | 2017-03-30 | エルジー エレクトロニクス インコーポレイティド | 3d放送サービスのための放送信号の送受信方法及び装置 |
JP2019103099A (ja) * | 2017-12-08 | 2019-06-24 | 日本放送協会 | 映像処理装置及びプログラム |
JP2020526116A (ja) * | 2017-07-05 | 2020-08-27 | オランジュ | 画像をエンコーディング及びデコーディングする方法、エンコーディング及びデコーディング装置、並びに、対応するコンピュータプログラム |
Families Citing this family (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP2356820B1 (en) | 2008-12-02 | 2017-07-19 | LG Electronics Inc. | 3d caption display method and 3d display apparatus for implementing the same |
US8670070B2 (en) * | 2010-07-15 | 2014-03-11 | Broadcom Corporation | Method and system for achieving better picture quality in various zoom modes |
JP2012100181A (ja) * | 2010-11-05 | 2012-05-24 | Hitachi Consumer Electronics Co Ltd | 映像出力装置、映像出力方法、受信装置および受信方法 |
KR20120058702A (ko) | 2010-11-27 | 2012-06-08 | 전자부품연구원 | 디지털 방송에서 서비스 호환 방식 전송 방법 |
KR20120058700A (ko) | 2010-11-27 | 2012-06-08 | 전자부품연구원 | 디지털 방송의 전송 모드 제공 및 인지 방법 |
US9185386B2 (en) * | 2011-06-01 | 2015-11-10 | Panasonic Intellectual Property Management Co., Ltd. | Video processing device, transmission device, video processing system, video processing method, transmission method, computer program and integrated circuit |
JP2013090016A (ja) * | 2011-10-13 | 2013-05-13 | Sony Corp | 送信装置、送信方法、受信装置および受信方法 |
JP6192902B2 (ja) * | 2011-11-11 | 2017-09-06 | サターン ライセンシング エルエルシーSaturn Licensing LLC | 画像データ送信装置、画像データ送信方法、画像データ受信装置および画像データ受信方法 |
EP2597876A1 (en) * | 2011-11-24 | 2013-05-29 | Koninklijke Philips Electronics N.V. | Interlaced 3D video |
US9106894B1 (en) | 2012-02-07 | 2015-08-11 | Google Inc. | Detection of 3-D videos |
US20130243079A1 (en) * | 2012-03-19 | 2013-09-19 | Nokia Siemens Networks Oy | Storage and processing savings when adapting video bit rate to link speed |
US9998734B2 (en) * | 2012-04-06 | 2018-06-12 | Sony Corporation | Decoding device and decoding method, and encoding device and encoding method |
JP5871332B2 (ja) * | 2012-05-10 | 2016-03-01 | ソニー株式会社 | 受信装置、受信方法、送信装置、及び送信方法 |
US9565452B2 (en) * | 2012-09-28 | 2017-02-07 | Qualcomm Incorporated | Error resilient decoding unit association |
ITTO20120901A1 (it) * | 2012-10-15 | 2014-04-16 | Rai Radiotelevisione Italiana | Procedimento di codifica e decodifica di un video digitale e relativi dispositivi di codifica e decodifica |
US20140165125A1 (en) * | 2012-12-11 | 2014-06-12 | Electronics And Telecommunications Research Institute | Broadcasting system and broadcasting method of providing multi-picture |
US10567804B2 (en) * | 2014-01-08 | 2020-02-18 | Qualcomm Incorporated | Carriage of HEVC extension bitstreams and buffer model with MPEG-2 systems |
WO2015126213A1 (ko) * | 2014-02-21 | 2015-08-27 | 엘지전자 주식회사 | 방송 신호 송신 장치 및 방송 신호 수신 장치 |
US20150253974A1 (en) | 2014-03-07 | 2015-09-10 | Sony Corporation | Control of large screen display using wireless portable computer interfacing with display controller |
WO2018169176A1 (ko) * | 2017-03-17 | 2018-09-20 | 엘지전자 주식회사 | 퀄리티 기반 360도 비디오를 송수신하는 방법 및 그 장치 |
US10594758B2 (en) * | 2017-12-15 | 2020-03-17 | Cisco Technology, Inc. | Latency reduction by sending audio and metadata ahead of time |
KR102007432B1 (ko) * | 2018-05-03 | 2019-08-06 | (주)블루어 | 3차원 콘텐츠 영상 생성 및 제공 시스템 |
US11743440B2 (en) | 2021-04-19 | 2023-08-29 | Apple Inc. | Transmission and consumption of multiple image subframes via superframe |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2002218502A (ja) * | 2001-01-22 | 2002-08-02 | Nippon Television Network Corp | 立体映像信号の伝送方法及び、そのシステム |
JP3789794B2 (ja) | 2001-09-26 | 2006-06-28 | 三洋電機株式会社 | 立体画像処理方法、装置、およびシステム |
JP2010130690A (ja) * | 2008-11-28 | 2010-06-10 | Korea Electronics Telecommun | 多視点映像送受信装置及びその方法 |
WO2010067810A1 (ja) * | 2008-12-09 | 2010-06-17 | ソニー株式会社 | 画像処理装置および方法 |
JP2011234336A (ja) * | 2010-04-06 | 2011-11-17 | Sony Corp | 画像データ送信装置、画像データ送信方法および画像データ受信装置 |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101472332B1 (ko) * | 2008-04-15 | 2014-12-15 | 삼성전자주식회사 | 3차원 디지털 컨텐츠의 제공 방법, 수신 방법 및 그 장치 |
CN101911713B (zh) * | 2008-09-30 | 2014-01-08 | 松下电器产业株式会社 | 再现装置、集成电路、再现方法、记录方法、记录介质再现系统 |
US8704874B2 (en) * | 2009-01-08 | 2014-04-22 | Lg Electronics Inc. | 3D caption signal transmission method and 3D caption display method |
US20120212579A1 (en) * | 2009-10-20 | 2012-08-23 | Telefonaktiebolaget Lm Ericsson (Publ) | Method and Arrangement for Multi-View Video Compression |
-
2011
- 2011-10-25 MX MX2013004068A patent/MX2013004068A/es active IP Right Grant
- 2011-10-25 EP EP11836297.9A patent/EP2635033A1/en not_active Withdrawn
- 2011-10-25 TW TW100138798A patent/TW201234833A/zh unknown
- 2011-10-25 JP JP2012540892A patent/JP5336666B2/ja not_active Expired - Fee Related
- 2011-10-25 CN CN2011800512729A patent/CN103202023A/zh active Pending
- 2011-10-25 KR KR1020137008616A patent/KR20130135244A/ko not_active Application Discontinuation
- 2011-10-25 WO PCT/JP2011/074586 patent/WO2012057164A1/ja active Application Filing
- 2011-10-25 US US13/280,764 patent/US20120106921A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2002218502A (ja) * | 2001-01-22 | 2002-08-02 | Nippon Television Network Corp | 立体映像信号の伝送方法及び、そのシステム |
JP3789794B2 (ja) | 2001-09-26 | 2006-06-28 | 三洋電機株式会社 | 立体画像処理方法、装置、およびシステム |
JP2010130690A (ja) * | 2008-11-28 | 2010-06-10 | Korea Electronics Telecommun | 多視点映像送受信装置及びその方法 |
WO2010067810A1 (ja) * | 2008-12-09 | 2010-06-17 | ソニー株式会社 | 画像処理装置および方法 |
JP2011234336A (ja) * | 2010-04-06 | 2011-11-17 | Sony Corp | 画像データ送信装置、画像データ送信方法および画像データ受信装置 |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2014042034A1 (ja) * | 2012-09-12 | 2014-03-20 | ソニー株式会社 | 送信装置、送信方法、受信装置および受信方法 |
WO2014050447A1 (ja) * | 2012-09-27 | 2014-04-03 | ソニー株式会社 | 送信装置、送信方法、受信装置および受信方法 |
JP2017509278A (ja) * | 2014-02-10 | 2017-03-30 | エルジー エレクトロニクス インコーポレイティド | 3d放送サービスのための放送信号の送受信方法及び装置 |
US9930382B2 (en) | 2014-02-10 | 2018-03-27 | Lg Electronics Inc. | Method and apparatus for transmitting/receiving broadcast signal for 3-dimensional (3D) broadcast service |
JP2016519480A (ja) * | 2014-02-21 | 2016-06-30 | エルジー エレクトロニクス インコーポレイティド | 3d放送信号を処理する方法及び装置 |
JP2020526116A (ja) * | 2017-07-05 | 2020-08-27 | オランジュ | 画像をエンコーディング及びデコーディングする方法、エンコーディング及びデコーディング装置、並びに、対応するコンピュータプログラム |
JP7242571B2 (ja) | 2017-07-05 | 2023-03-20 | オランジュ | 画像をエンコーディング及びデコーディングする方法、エンコーディング及びデコーディング装置、並びに、対応するコンピュータプログラム |
JP2019103099A (ja) * | 2017-12-08 | 2019-06-24 | 日本放送協会 | 映像処理装置及びプログラム |
JP7018752B2 (ja) | 2017-12-08 | 2022-02-14 | 日本放送協会 | 映像処理装置及びプログラム |
Also Published As
Publication number | Publication date |
---|---|
US20120106921A1 (en) | 2012-05-03 |
JPWO2012057164A1 (ja) | 2014-05-12 |
TW201234833A (en) | 2012-08-16 |
EP2635033A1 (en) | 2013-09-04 |
JP5336666B2 (ja) | 2013-11-06 |
CN103202023A (zh) | 2013-07-10 |
KR20130135244A (ko) | 2013-12-10 |
MX2013004068A (es) | 2013-05-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5336666B2 (ja) | 符号化方法、表示装置、復号方法 | |
JP6229962B2 (ja) | 符号化装置及び符号化方法 | |
JP5906462B2 (ja) | 映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム | |
WO2012111325A1 (ja) | 映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム | |
US20120050476A1 (en) | Video processing device | |
US20140089962A1 (en) | Image playback device, image playback method, image playback program, image transmission device, image transmission method and image transmission program | |
WO2012169204A1 (ja) | 送信装置、受信装置、送信方法及び受信方法 | |
WO2013099289A1 (ja) | 再生装置、送信装置、再生方法及び送信方法 | |
US20140002621A1 (en) | Video processing device and video processing method | |
JP6008292B2 (ja) | ビデオストリームの映像のデータ作成装置及び再生装置 | |
US8934757B2 (en) | Content distribution system, playback device, distribution server, playback method, and distribution method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 11836297 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2012540892 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20137008616 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011836297 Country of ref document: EP Ref document number: MX/A/2013/004068 Country of ref document: MX |
|
NENP | Non-entry into the national phase |
Ref country code: DE |