WO2012108210A1 - ビデオストリームの映像のデータ作成装置及び再生装置 - Google Patents
ビデオストリームの映像のデータ作成装置及び再生装置 Download PDFInfo
- Publication number
- WO2012108210A1 WO2012108210A1 PCT/JP2012/000899 JP2012000899W WO2012108210A1 WO 2012108210 A1 WO2012108210 A1 WO 2012108210A1 JP 2012000899 W JP2012000899 W JP 2012000899W WO 2012108210 A1 WO2012108210 A1 WO 2012108210A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- stream
- video stream
- view video
- frame
- picture
- Prior art date
Links
- 230000006835 compression Effects 0.000 claims abstract description 85
- 238000007906 compression Methods 0.000 claims abstract description 85
- 230000001419 dependent effect Effects 0.000 claims abstract description 57
- 238000000034 method Methods 0.000 claims description 137
- 230000008569 process Effects 0.000 claims description 47
- 238000012545 processing Methods 0.000 claims description 24
- 230000002123 temporal effect Effects 0.000 claims description 4
- 239000000872 buffer Substances 0.000 description 48
- 230000000875 corresponding effect Effects 0.000 description 38
- 238000001824 photoionisation detection Methods 0.000 description 35
- 230000015654 memory Effects 0.000 description 27
- 238000010586 diagram Methods 0.000 description 25
- 210000003128 head Anatomy 0.000 description 19
- 230000006870 function Effects 0.000 description 16
- 238000004590 computer program Methods 0.000 description 9
- 230000008929 regeneration Effects 0.000 description 9
- 238000011069 regeneration method Methods 0.000 description 9
- 238000005516 engineering process Methods 0.000 description 8
- 239000011521 glass Substances 0.000 description 8
- 102100037812 Medium-wave-sensitive opsin 1 Human genes 0.000 description 7
- 230000006978 adaptation Effects 0.000 description 7
- 238000012546 transfer Methods 0.000 description 7
- 239000004973 liquid crystal related substance Substances 0.000 description 6
- 238000012986 modification Methods 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 5
- 230000008859 change Effects 0.000 description 5
- 230000004048 modification Effects 0.000 description 5
- 230000000750 progressive effect Effects 0.000 description 5
- 230000007704 transition Effects 0.000 description 5
- 230000001172 regenerating effect Effects 0.000 description 4
- 230000011664 signaling Effects 0.000 description 4
- 238000006243 chemical reaction Methods 0.000 description 3
- 230000010354 integration Effects 0.000 description 3
- 241000086550 Dinosauria Species 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 239000000470 constituent Substances 0.000 description 2
- 239000000284 extract Substances 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000000153 supplemental effect Effects 0.000 description 2
- 230000002194 synthesizing effect Effects 0.000 description 2
- 241000872198 Serjania polyphylla Species 0.000 description 1
- 206010047571 Visual impairment Diseases 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 230000001276 controlling effect Effects 0.000 description 1
- 230000002596 correlated effect Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000009191 jumping Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000001094 photothermal spectroscopy Methods 0.000 description 1
- 230000010287 polarization Effects 0.000 description 1
- 238000002360 preparation method Methods 0.000 description 1
- 238000007430 reference method Methods 0.000 description 1
- 238000000926 separation method Methods 0.000 description 1
- 238000003860 storage Methods 0.000 description 1
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/30—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using hierarchical techniques, e.g. scalability
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/70—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals characterised by syntax aspects related to video coding, e.g. related to compression standards
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/503—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving temporal prediction
- H04N19/51—Motion estimation or motion compensation
- H04N19/577—Motion compensation with bidirectional frame interpolation, i.e. using B-pictures
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/59—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding involving spatial sub-sampling or interpolation, e.g. alteration of picture size or resolution
Definitions
- the present invention relates to a technique for encoding and decoding for video recording or transmission, and reproduction and reproduction, and more particularly to a data creation apparatus for 3D video data and a reproduction apparatus thereof.
- MPEG-4 AVC / H. MPEG-4 MVC (Multiview Video Coding) is a method for compressing and encoding 3D video.
- this MPEG-4 MVC format compression encoding method it is possible to generate a 3D video stream that realizes compression encoding with a high compression rate.
- the present invention has been made in view of the above problems, and while maintaining reproduction on an existing reproduction apparatus that supports MPEG-2 format video streams, the video stream corresponding to 3D video is reduced in data amount. It is an object of the present invention to provide a data creation device that generates a reduced amount of data, and a playback device for such a video stream.
- a data creation device includes a first frame image group that shows a view over time at a first viewpoint, and a second frame image group that shows a view over time at a second viewpoint.
- a data generation apparatus for compression encoding wherein the first encoding means for generating an MPEG-2 format stream by compressing and encoding the first frame image group, and a base conforming to the MPEG-4 MVC format
- a second encoding unit that generates a view video stream and a dependent-view video stream; and a transmission unit that transmits the streams generated by the first and second encoding units, and the second encoding unit.
- a stream composed of pseudo data with a reduced total data amount is generated.
- a frame of the MPEG-2 format stream at the same time as a corresponding frame of the base-view video stream is referred to as the dependent-view stream.
- a stream obtained by compression-encoding each frame constituting the second frame image group is generated.
- the playback device includes an acquisition unit that acquires a stream transmitted by the data creation device, and a decoder that decodes the MPEG-2 format stream included in the stream acquired by the acquisition unit.
- a first decoding unit that obtains a first frame image group; a base-view video stream included in the stream acquired by the acquiring unit; and a first frame at the same time as a corresponding frame of the base-view video stream
- a second decoding unit that obtains the second frame image group by decoding a dependent-view video stream included in the stream using the frame decoded by one decoding unit as a reference image; and the first and second frames Reproduction means for reproducing the first and second frame image groups obtained by the decoding means.
- the Rukoto wherein the Rukoto.
- the data creation device reduces the data amount of the video stream corresponding to 3D video while maintaining the playback on the existing playback device that supports the MPEG-2 format video stream. Can be generated.
- the playback device can play back a video stream corresponding to the 3D video.
- the figure which shows the reference relationship of the picture of a video stream The figure which shows the compression encoding system by the MPEG-4 MVC format.
- the conceptual diagram in the case of producing
- the conceptual block diagram of a PES packet The figure which shows the data structure of TS packet which comprises a transport stream.
- the conceptual diagram of the Side-by-Side system The figure which shows an example of an internal structure of the view video stream of MPEG-4 MVC format.
- the conceptual diagram of the internal structure of the access unit of a video stream The figure which shows the structure of the video access unit of each picture of a base view video stream, and each picture of the image video stream for right eyes.
- the figure which shows the GOP structure of a base view video stream and a dependent view video stream The figure which shows the structure of the video access unit contained in dependent GOP.
- a 2D compatible video stream is an MPEG-2 video
- a multi-view video stream is an MPEG-4 MVC format compression encoding system
- the video attributes to be matched and fields for indicating the video attributes are shown.
- the figure which shows a name The figure which shows the example of the relationship of PTS, DTS, and picture type allocated to each video access unit of a 2D compatibility video stream in a transport stream, a base view video stream, and a dependent view video stream.
- regenerating apparatus which concerns on this Embodiment.
- the 2D playback portion of the playback device decodes and plays back the MPEG-2 format stream using an existing decoding method
- the 3D playback portion includes a base-view video stream and a video stream conforming to MPEG-4 MVC.
- the pent-view video stream is decoded and reproduced in a decoding format corresponding to the new encoding.
- FIG. 21 shows the data structure of the transport stream created by the broadcast system in the present embodiment. As shown in the figure, the transport stream is composed of a 2D compatible video stream A and a multi-view video stream B.
- the latter multi-view video stream B is composed of a base-view video stream B1 and a dependent-view video stream B2.
- the 2D compatible video stream A is created by compressing and coding the left-eye image
- the base-view video stream B1 is created by compressing and coding a monochrome image such as black (hereinafter referred to as “black image”).
- the dependent-view video stream B2 is generated by compressing and encoding the difference between the left-eye image and the right-eye image.
- the base-view video stream B1 is a compression-encoded black image, it cannot be used as a reference image when generating the dependent-view video stream B2. This point is different from the existing MPEG-4 MVC format, and the reference image is set to the frame image of the 2D compatible video stream A at the same time.
- both 2D video and 3D video can be transmitted, and the base view video stream B1 compresses and encodes a black image, so that the bit rate is extremely low.
- both 2D video and 3D video can be transmitted in the range of the existing allocated frequency band.
- the dependent-view video stream is decoded with reference to the frame image of the base-view video stream.
- the MPEG-2 compatible stream is decoded. That is, the dependent-view video stream can be decoded using the frame image of the left-eye image as a reference image.
- FIG. 26 is a block diagram of a functional configuration of the data creation device 2601 according to the present embodiment.
- the data creation device 2601 receives a left-eye image, a right-eye image, and a black image that constitute a 3D video as inputs, and includes a transformer including a 2D-compatible video stream, a base-view video stream, and a dependent-view video stream in a data format described later. Output the port stream.
- the data creation device 2601 includes a 2D compatible video encoder 2602, a Dec (2D compatible video decoder) 2603, an extended multi-view video encoder 2604, and a multiplexer 2610.
- the extended multi-view video encoder 2604 includes a base-view video encoder 2605, a 2D compatible video frame memory 2608, and a dependent-view video encoder 2609.
- the 2D compatible video encoder 2602 receives the left-eye image as an input, compresses and encodes it in the MPEG-2 format, and generates and outputs a 2D compatible video stream.
- Dec 2603 decodes the compression-encoded picture in the 2D compatible video stream, and outputs a decoded picture obtained as a result and 2D compatible video encoding information 2606.
- a picture is an image constituting a frame or a field, and is a unit of encoding.
- the decoded picture is stored in the 2D compatible video frame memory 2608 of the extended multiview video encoder 2604.
- the 2D compatible video encoding information 2606 is input to the base view video encoder 2605.
- the 2D compatible video encoding information 2606 includes attribute information (resolution, aspect ratio, frame rate, progressive / interlaced distinction, etc.), picture attribute information (picture type, etc.) of the corresponding picture, GOP of the decoded 2D compatible video stream. (Group of Pictures) structure and configuration including 2D compatible video frame memory management information.
- the 2D compatible video frame memory management information includes the decoded picture memory address stored in the 2D compatible video frame memory 2608, the display order information (PTS (Presentation Time Stamp) and temporal_reference) of the corresponding picture, and the code order information (code order of the file). Or DTS (Decoding Time Stamp).
- the extended multi-view video encoder 2604 receives the decoded picture output from the Dec 2603, the 2D-compatible video encoding information, the right-eye image, and the black image, performs compression encoding, and performs the base-view video stream, the dependent Dent-view video stream is output.
- the base view video encoder 2605 has a function of outputting the data compressed and encoded in accordance with the MPEG-4 MVC format as a base view video stream, and compresses and encodes a black image according to the 2D compatible video encoding information 2606. , A base-view video stream and base-view video encoding information 2607 are output.
- Base view video encoding information 2607 includes base view video stream attribute information (resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.), picture attribute information (picture type, etc.) of the corresponding picture, GOP structure, and base
- the configuration includes view video frame memory management information.
- the base-view video encoder 2605 sets the attribute information of the base-view video stream to the same value as the video attribute information included in the 2D compatible video encoding information 2606. Furthermore, the base-view video encoder 2605 determines a picture type for compression coding of pictures at the same display time in accordance with picture attribute information (picture type, etc.) and GOP structure included in the 2D compatible video coding information 2606. Thus, the black image is compression encoded. For example, if the picture type indicated by the 2D compatible video encoding information 2606 of the picture at time a is an I picture and the picture is the first picture of the GOP, the base view video encoder 2605 displays a black image indicating the same display time. The video access unit is compressed and encoded so as to be an I picture, and is used as a video access unit at the head of the GOP of the base-view video stream.
- the base-view video encoder 2605 compresses and encodes a black image indicating the same display time so as to become a B picture. To do. At this time, the DTS and PTS of the base-view video stream are matched with the DTS and PTS of the picture corresponding to the view showing the same time of the 2D compatible video stream, respectively.
- the base-view video frame memory management information includes a memory address of the frame memory 2608 that stores a decoded picture obtained by decoding a 2D-compatible video stream based on the 2D-compatible video frame memory management information, and the decoded picture
- This is information obtained by converting and associating a syntax element indicating display order information and code order information into a syntax element conforming to the rules of the compression encoding method of the base-view video stream.
- the syntax element is an element that defines attribute information necessary for encoding in the MPEG-2 or MPEG-4 MVC format compression encoding system. For example, header information such as macroblock type, motion vector, conversion coefficient, etc. It is an element to show.
- the dependent-view video encoder 2609 has a function of compressing and encoding in accordance with the MPEG-4 MVC format to generate a dependent-view video stream, and is based on information included in the base-view video encoding information 2607
- the right-eye image is compressed and encoded, and a dependent-view video stream is output.
- the dependent-view video encoder 2609 performs compression encoding using the decoded picture in the 2D compatible video frame memory as an inter-view reference.
- inter-view reference refers to referring to a picture showing a view from different viewpoints.
- the dependent-view video encoder 2609 determines a reference picture ID for performing inter-view reference based on the base-view video frame memory management information of the base-view video encoding information 2607.
- the dependent-view video encoder 2609 sets the same value as the value of the attribute information of the base-view video stream included in the base-view video encoding information 2607 as the video attribute information of the dependent-view video stream.
- the dependent-view video encoder 2609 determines the picture type of the image to be encoded based on the attribute information (for example, picture type) of the picture stored in the base-view video encoding information 2607 and the GOP structure. Then, the right-eye image is compression-encoded. For example, if the picture type indicated by the base-view video encoding information 2607 of the picture at time a is an I picture and the head of the GOP, the dependent-view video encoder 2609 sets the picture type of the picture at the same time a as an anchor picture. Then, the right-eye image is compression-encoded to obtain a video access unit at the head of the dependent GOP.
- the attribute information for example, picture type
- the dependent-view video encoder 2609 sets the picture type of the picture at the same time a as an anchor picture. Then, the right-eye image is compression-encoded to obtain a video access unit at the head of the dependent GOP.
- An anchor picture is a picture that does not refer to a picture temporally prior to the own picture and can be reproduced by jumping from the own picture. If the picture type indicated by the base-view video encoding information 2607 of the picture at time b is a B picture, the dependent-view video encoder 2609 sets the picture type of the picture at the same time b to B picture, and The image is compressed and encoded.
- the DTS and PTS of the dependent-view video stream are compression-encoded to match the DTS and PTS of the picture corresponding to the view to be displayed at the same time of the base-view video stream.
- the multiplexer 2610 divides the output 2D compatible video stream, base-view video stream, and dependent-view video stream into PES (Packetized Elementary Stream) packets, divides them in units of TS packets, and multiplexes them as a transport stream Output.
- PES Packetized Elementary Stream
- the playback device In order to allow the playback device to identify each video stream from the stream data of the multiplexed transport stream, the 2D compatible video stream, the base view video stream, and the dependent view video stream are: Separate PIDs are set.
- FIG. 22 shows video attributes to be matched in each compression encoding method in the MPEG-2 format and MPEG-4 MVC format compression encoding, and field names for indicating the video attributes.
- the resolution, aspect ratio, The frame rate and the value of the video attribute indicating progressive or interlace are configured to be the same between pictures of the respective encoding formats.
- FIG. 25 shows a GOP configuration of a 2D compatible video stream, a base view video stream, and a dependent view video stream in the present embodiment.
- the 2D compatible video stream, the base view video stream, and the dependent view video stream are all configured to have the same number of GOPs. That is, when the picture of the 2D compatible video stream is the head of the GOP, the picture of the base view video stream having the same value of the PTS and the picture of the dependent view video stream having the same value of the PTS are respectively the GOP head, Must be at the beginning of the dependent GOP.
- entry map information indicating where the GOP head picture is on the file may be stored as management information.
- the entry map information is stored as a separate information file as a management information file.
- both the base view position and the dependent view position at the same time are entered. Try to register on the map. In this way, the 3D video jump-in reproduction is facilitated by referring to the entry map.
- FIG. 36 shows the relationship between the structure of the transport stream and the PMT (Program Map Table) packet.
- a transport stream including a 3D video stream signaling information for performing 3D video decoding processing is included in a system packet such as a PMT packet.
- the descriptor includes a 3D information descriptor for performing signaling such as the relationship between each video stream and the start / end of 3D video playback of the present system, and a 3D stream descriptor set for each video stream. It is the composition which includes.
- FIG. 37 shows the structure of the 3D information descriptor.
- the 3D information descriptor includes a playback method, a left-eye video type, a 2D compatible video PID, a base view video PID, and a dependent view video PID.
- the playback method is information for signaling the playback method of the playback device.
- the playback device plays back 2D video of only the 2D compatible video stream.
- the playback apparatus When the value of the playback method is “1”, it indicates 3D video playback using 2D compatible video and dependent view video (that is, the 3D video playback method described in this embodiment).
- the playback apparatus performs 3D video playback of the 2D compatible video stream, base-view video stream, and dependent-view video stream by the playback method described in this embodiment. Note that a 3D video playback method in the present embodiment will be described later.
- the playback device plays back the 3D video of the video stream as a video stream compressed and encoded in the normal MPEG-4 MVC format.
- the playback device When the value of the playback method is “3”, this indicates that a 2D compatible video stream or a base-view video stream is double-played, and the playback device performs double playback.
- the doubling reproduction is to output one of the left and right views at a certain time a to both the L and R planes.
- Playback by this playback method is equivalent to 2D video playback as a user viewing screen, but since the frame rate does not change during 3D video playback, the playback device is connected to a display or the like via HDMI (High-Definition Multimedia Interface) or the like. In this case, re-authentication does not occur and seamless connection playback between the 2D video playback section and the 3D video playback section can be realized.
- HDMI High-Definition Multimedia Interface
- the left-eye video type is information indicating which of the multi-view video streams the left-eye image is compression-coded (the other video stream becomes the right-eye image).
- the value of the reproduction method is “0”, the value of this field has no meaning to refer to.
- the value of the reproduction method is “1”, it indicates which of the 2D compatible video and the dependent view video is the image for the left eye. That is, when the value of the playback method is “1” and the value of the left-eye video type is “0”, it indicates that the 2D compatible video stream is a video stream corresponding to the left-eye image.
- the playback method value is “2” or “3”
- the playback apparatus refers to the value of the left-eye video type to determine which video stream corresponds to the left-eye image. Judgment can be made.
- the 2D compatible video PID, the base view video PID, and the dependent view video PID indicate the PID of each video stream in the transport stream. With this information, the decoding target stream can be identified.
- FIG. 39 shows a 3D stream descriptor.
- the field name of the 3D descriptor includes a base view video type, a reference destination type, and a referenced type.
- the base view video type indicates which video is compressed and encoded in the base view video stream.
- the value of the base view video type is “0”, it indicates whether one of the left-eye image or the right-eye image data of the 3D video is compression-encoded, and when it is “1”, A black image is compressed and encoded as a dummy video that is replaced by a 2D compatible video stream and is not output to a plane.
- the reference destination type indicates the type of the video stream to which the dependent-view video stream is referred to as an inter-view reference.
- the value of the reference destination type is “0”, it indicates that the picture of the base view video stream is the inter-view reference destination, and when the value is “1”, the picture of the 2D compatible video stream is the inter-view reference destination. It shows that.
- the case where the value of the reference destination type is “1” indicates that the reference method is based on the 3D video method in the present embodiment.
- the referenced type indicates whether the corresponding video stream is referenced between views. If the reference is not made, the inter-view reference process can be skipped, so that the load of the decoding process can be reduced. Note that all or a part of the information of the 3D information descriptor and the 3D stream descriptor may be stored in supplementary data of each video stream instead of the PMT packet.
- FIG. 23 shows an example of the relationship among PTS, DTS, and picture type assigned to each video access unit of the 2D compatible video stream, base-view video stream, and dependent-view video stream in the transport stream.
- the data creation device 2601 converts the DTS and PTS values of the 2D compatible video stream picture and the dependent view video stream picture generated by compressing and encoding the left-eye image at the same time into the same DTS and PTS values, respectively. Set. Also, the same values as the PTS, DTS, and POC of the picture of the dependent-view video stream are set for the PTS, DTS, and POC of the picture of the base-view video stream to be reproduced at the same time.
- the picture of the base-view video stream having the same value of PTS, DTS, and POC is referred to.
- the picture reference ID (ref_idx_l0 or ref_idx_l1) specified from each macroblock of the picture of the dependent-view video stream has the same POC value.
- a value indicating the base view picture that is held is set.
- FIG. 27 is a diagram showing a data creation flow of the data creation device 2601. The data creation flow will be described below.
- the variable N is a variable for storing the frame number of the frame image to be compressed and encoded.
- step S2701 it is checked whether or not the Nth frame exists in the left-eye image. If it does not exist (step S2701: NO), it is determined that there is no more data to be compressed and encoded, and the process ends.
- step S2701 the number of images to be compression encoded (hereinafter referred to as “one encoded number”) is determined in one compression encoding flow (steps S2702 to S2706) (step “1 encoded number” hereinafter).
- the maximum number of video access units (the maximum number of GOPs, for example, 30 frames) set as one GOP is set as one encoded number.
- the last GOP of the video stream it is assumed that the number of frames to be encoded is less than the maximum number of GOPs depending on the length of the input video stream. In this case, the remaining number of frames is set as one encoded number. To do.
- the 2D compatible video encoder 2602 generates 2D compatible video streams for one encoded number (step S2703).
- the 2D compatible video stream is generated by the compression encoding according to the compression encoding method of the 2D compatible video stream from the Nth frame of the left-eye image by one encoded number, and output.
- the 2D compatible video decoder 2603 performs decoding processing of 2D compatible video streams for one encoded number (step S2704).
- a decoded picture and 2D compatible video encoding information obtained by decoding compressed picture data for one encoded number from the Nth frame are output.
- the base view video encoder 2605 generates base view video streams for one encoded number (step S2705). Specifically, based on 2D compatible video encoding information, base view video stream attribute information (resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.) as base view encoding information 2607, Set picture attribute information (picture type, etc.), GOP (Group of Pictures) structure, and 2D compatible video frame memory management information for each picture. Generate a video stream. Also, the set base-view encoding information 2607 is output.
- base view video stream attribute information resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.
- Set picture attribute information picture type, etc.
- GOP Group of Pictures
- the dependent view video encoder 2609 generates a dependent view video stream for one encoded number (step S2706). Specifically, based on the base-view video encoding information output in step S2705, the attribute information of the dependent-view video stream (resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.), Set picture attribute information (picture type, etc.), GOP (Group of Pictures) structure, and 2D compatible video frame memory management information for each picture.
- the attribute information of the dependent-view video stream resolution, aspect ratio, frame rate, progressive or interlaced distinction, etc.
- Set picture attribute information picture type, etc.
- GOP Group of Pictures
- the dependent-view video stream encoder 2609 does not refer to the picture of the base-view video stream, but performs the same display in the 2D-compatible video frame memory 2608 when encoding using inter-picture predictive encoding.
- the dependent-view video stream is compressed and encoded for one encoded number from the Nth frame of the right-eye image using inter-picture predictive encoding while referring to the picture obtained by decoding the 2D compatible video stream indicating the time. Is generated.
- the multiplexer 2610 converts the 2D compatible video stream, base view video stream, and dependent view video stream into PES packets. Next, the PES packet is divided into TS packets, and a multiplexed transport stream is generated. Thereafter, one encoded number is added to N (step S2707).
- step S2707 When the process of step S2707 is completed, the process returns to step S2701 and the process is repeated.
- the number of encoded sheets in one flow can be changed.
- the value of one encoded sheet in step S2702 may be set to be small. For example, when the number of reordering at the time of video encoding is two, if compression encoding is executed in units of four, it will not be affected by reordering. It is assumed that the number of reordering is a compression encoding method with two pictures and the picture types are I1, P4, B2, B3, P7, B5, and B6 (numbers are in display order). When the number of encoded images is 3, since the picture of P4 cannot be processed, the compression encoding process of B2 and B3 cannot be performed.
- the number of encoded images is set to 4, the P4 picture can be processed, so that the B2 and B3 compression encoding processes can be performed.
- the number of encoded sheets may be set to an optimal number according to the characteristics of the image within the maximum number of GOPs for each compression encoding flow.
- FIG. 28 is a block diagram showing a functional configuration of the playback device 2823.
- the playback device 2823 includes a PID filter 2801, a 2D compatible video decoder 2821, an extended multi-view video decoder 2822, a first plane 2808, and a second plane 2820.
- the PID filter 2801 filters the input transport stream.
- the PID filter 2801 determines the TS packet whose PID value matches the PID value required for reproduction, from the TS packet, to the 2D compatible video decoder 2821 or the extended multi-view video decoder 2822 according to the PID value. Forward.
- the PID of the 2D compatible video stream is 0x1011
- the PID of the base-view video stream of the multi-view video stream is 0x1012
- the PID of the dependent-view video stream of the multi-view video stream is 0x1013
- the PID value of the TS packet is When the PID value included in the TS packet matches the predetermined PID value, the TS packet is transferred to the corresponding decoder.
- the first plane 2808 is a plane memory that holds a picture decoded by the 2D compatible video decoder 2821 and output according to the value of the PTS.
- the second plane 2820 is a plane memory that holds a picture decoded by the extended multi-view video decoder 2822 and output according to the value of the PTS.
- the 2D compatible video decoder 2821 has basically the same decoding function as the MPEG-2 format decoder which is a compression encoding method of 2D video
- the extended multi-view video decoder 2822 is a 3D video that realizes inter-view reference.
- the decoder has basically the same decoding function as the MPEG-4 MVC format decoder, which is a compression encoding method.
- a general decoder of MPEG-2 format compression encoding system is a video decoder 2901
- a general decoder of MPEG-4 MVC format compression encoding system is a multi-view video decoder 2902.
- the video decoder 2901 and the multi-view video decoder 2902 will be described with reference to FIG. Thereafter, different portions between the 2D compatible video decoder 2821 and the video decoder 2901 and different portions between the extended multi-view video decoder 2822 and the multi-view video decoder 2902 will be mainly described.
- the video decoder 2901 includes TB (Transport Stream Buffer) (1) 2802, MB (Multiplexing Buffer) (1) 2803, EB (Elementary Stream Buffer) (1) 2804, D1 (2D compatible video compression video decoder) 2805. , And O (Re-ordering Buffer) 2806.
- TB (1) 2802 is a buffer that temporarily accumulates TS packets as they are when TS packets including a video stream are output from the PID filter 2801.
- MB (1) 2803 is a buffer for temporarily storing PES packets when outputting a video stream from TB (1) 2802 to EB (1) 2804.
- TB (1) 2802 to MB (1) 2803 When data is transferred from TB (1) 2802 to MB (1) 2803, the TS header and adaptation field of the TS packet are removed.
- EB (1) 2804 is a buffer for storing pictures (I picture, B picture, P picture) in a compression-encoded state.
- pictures I picture, B picture, P picture
- the PES header is removed.
- D1 (2805) creates a picture of the frame image by decoding each video access unit of the video elementary stream at a predetermined DTS time.
- the picture decoded in D1 (2805) is output to the plane 2808 or O2806.
- DTS and PTS are different, such as P picture and I picture, they are output to O2806, and when the values of DTS and PTS are the same as B picture, they are output to plane 2808 as they are.
- O2806 is a buffer for performing reordering when the values of DTS and PTS of the decoded picture are different, that is, when the decoding order and display order of the pictures are different.
- D1 (2805) performs decoding processing with reference to the picture data stored in O2806.
- Switch 2807 switches between outputting an image buffered in O2806 or outputting directly from D1 (2805) when outputting the decoded picture to plane 2808.
- the multi-view video decoder 2902 includes TB (2) 2809, MB (2) 2810, EB (2) 2811, TB (3) 2812, MB (3) 2813, EB (3) 2814, A decoding switch 2815, an inter-view buffer 2816, a D2 (multiview video compression video decoder) 2817, a DPB (decoded picture buffer) 2818, and an output plane switch 2819 are included.
- TB (2) 2809, MB (2) 2810, and EB (2) 2811 have the same functions as TB (1) 2802, MB (1) 2803, and EB (1) 2804, respectively, but the data to be buffered Is the base view video stream.
- TB (3) 2812, MB (3) 2813, and EB (3) 2814 have the same functions as TB (1) 2802, MB (1) 2803, and EB (1) 2804, respectively, but the data to be buffered Is a dependent-view video stream.
- the switch 2815 extracts the data of the video access unit to which the DTS is assigned according to the value of the DTS from the EB (2) 2811 and the EB (3) 2814, and configures the 3D video access unit. ).
- D2 (2817) performs a decoding process on the 3D video access unit transferred via the switch 2815, and creates a picture of the frame image.
- the decoded picture of the base-view video decoded by D2 (2817) is temporarily stored in the Inter-view buffer 2816.
- D2 (2817) performs a decoding process on the picture of the dependent-view video stream with reference to the decoded picture of the base-view video stream having the same PTS value stored in the Inter-view buffer 2816.
- the multi-view video decoder 2902 creates a reference picture list for designating a picture for performing inter-view reference based on the picture type and the syntax element of the picture of the base-view video stream and the picture of the dependent-view video stream.
- D2 transfers the decoded picture of the base view and the decoded picture of the dependent view stored in the inter-view buffer 2816 to the DPB 2818, and outputs it via the output plane switch 2819 according to the value of the PTS.
- the DPB 2818 is a buffer that temporarily stores decoded pictures.
- D2 (2817) is used to refer to a picture that has already been decoded when decoding a video access unit such as a P picture or a B picture using the inter picture predictive coding mode.
- the output plane switch 2819 performs a process of outputting the decoded picture to an appropriate plane. For example, when the base-view video stream indicates the left-eye image and the dependent-view video stream indicates the right-eye image, the base-view video stream picture is set to the left-eye image plane and the dependent-view video The stream picture is output to the right-eye image plane.
- the 2D compatible video decoder 2821 shown in FIG. 28 not only transfers the picture decoded in D1 (2805) to the O2806 or the switch 2807, but also in the inter-view buffer 2816 of the extended multi-view video decoder 2822 based on the DTS value. Forward.
- the extended multi-view video decoder 2822 has the same basic structure as that of the multi-view video decoder 2902. Therefore, the description of the common parts of functions will be omitted and the different parts will be described.
- the extended multi-view video decoder 2822 converts the picture transferred according to the DTS value from the 2D compatible video decoder 2821 to the same value in the PTS and DTS of the base-view video stream that has been decoded previously in the inter-view buffer 2816. Overwrite the area where the decoded picture is stored. Therefore, the extended multi-view decoder 2822 can refer to the decoded picture of the 2D-compatible video stream as if it were the decoded picture of the base-view video stream when decoding the picture of the dependent-view video stream.
- the address management on the Inter-view buffer 2816 need not be changed from the management of the decoded picture of the conventional base-view video stream.
- the extended multi-view video decoder 2822 outputs to the second plane 2820 out of the video stored in the DPB 2818, so that only the picture of the dependent-view video stream is output according to the value of the PTS. To control. The pictures of the base view video stream are not output to the plane because they are not related to display.
- the 2D-compatible video decoder 2821 outputs the picture of the 2D-compatible video stream to the first plane 2808 according to the value of the PTS
- the extended multi-view video decoder 2822 outputs the dependent view of the multi-view video stream.
- the picture of the video stream is output to the second plane 2820 according to the value of PTS.
- FIG. 30 is a diagram illustrating a flow of decoding processing and output processing in 3D video by the playback device 2823.
- the playback device 2823 determines whether there is a picture in EB (1) 2804 (step S3001). If there is no picture (step S3001: No), it is determined that the transfer of the video stream has ended, and the processing is ended.
- the playback device 2823 uses the extended multi-view video decoder 2822 to perform base-view video stream decoding processing (step S3002). Specifically, in accordance with the value of DTS, a picture attached to the corresponding DTS is extracted from EB (2), decoded, and stored in Inter-view buffer 2816.
- the management of pictures in Inter-view buffer 2816 is the same as the management in the conventional MPEG-4 MVC format and will not be described in detail. However, as the management information for creating the reference picture list, the reference destination of PTS and POC and the decoded picture
- the table information that associates the data address of Inter-view buffer 2816 indicating the internal address is managed internally.
- the playback device 2823 uses the 2D compatible video decoder 2821 to perform decoding processing of the 2D compatible video stream (step S3003). Specifically, according to the value of DTS, a picture assigned to the corresponding DTS is extracted from EB (1), and decoding processing is performed. At this time, the decoded picture is transferred to O2806 or switch 2807. Further, the decoded picture is also transferred to Inter-view buffer 2816.
- the extended multi-view video decoder overwrites the transferred picture with the base-view picture to which the same value of DTS and PTS in the Inter-view buffer 2816 is assigned.
- the picture management of the inter-view buffer 2816 is managed by the memory address of the PTS and the inter-view buffer 2816, for example.
- the decoded picture is overwritten.
- management information eg, PTS
- PTS management information
- D2 2817
- D2 can be decoded with reference to the picture obtained by decoding the 2D compatible video stream in the same process as the decoding process of the conventional MPEG-4 MVC dependent view video stream. it can.
- the extended multi-view video decoder 2822 performs a dependent-view video stream decoding process (step S3004). Specifically, the extended multi-view video decoder 2822 takes out the picture attached to the corresponding DTS from the EB (3) according to the value of the DTS, and stores the picture of the dependent-view video stream in the Inter-view buffer 2816. The decoding process is performed while referring to the picture.
- the picture referred to at this time is not a picture of the base-view video stream but a picture of the 2D compatible video stream overwritten in step S3003.
- the playback device 2823 outputs the decoded picture of the 2D-compatible video stream to the first plane 2808 according to the PTS value, and outputs the decoded picture of the dependent-view video stream to the second plane 2820 according to the PTS value (step).
- D1 decoding process performed by D1 (2805) of the playback apparatus 2823
- the LSI Large Scale Integration
- the MPEG-4 MVC format decoding process performed in D2 (2817) is the same as the existing MPEG-4 MVC format process
- the LSI or software of the existing MPEG-4 MVC format video playback apparatus Can be used.
- ⁇ Example of usage mode of playback device 2823> Referring to FIG. 5, a 3D digital television 100 that can play back 3D video of a video stream created by the data creation device 2823 and a 2D digital television 300 that can play back only conventional 2D video that does not support playback of 3D video are given as examples. I will explain.
- the user views 3D video using the 3D digital television 100 and the 3D glasses 200.
- the 3D digital television 100 can display 2D video and 3D video, and displays video by playing back a stream included in the received broadcast wave. Specifically, a 2D compatible video stream compression-encoded in the MPEG-2 format, a base-view video stream and a dependent-view video stream compression-encoded in accordance with the MPEG-4 MVC format are reproduced.
- the 3D digital television 100 alternately displays the left-eye image obtained by decoding the 2D-compatible video stream and the right-eye image obtained by decoding the dependent-view video stream.
- the user can view the image reproduced in this way as a stereoscopic image by wearing the 3D glasses 200 and viewing the image.
- FIG. 5B shows the time when the left-eye image of the 3D glasses 200 is displayed.
- the 3D glasses 200 make the liquid crystal shutter corresponding to the left eye transparent and the liquid crystal shutter corresponding to the right eye shielded.
- (C) in the figure shows when the right eye image is displayed.
- the liquid crystal shutter corresponding to the right eye is made transparent, and the liquid crystal shutter corresponding to the left eye is shielded from light.
- the 2D digital television 300 in FIG. 5D corresponds to the playback of 2D video, and 2D video obtained by decoding the 2D compatible video stream among the video streams included in the transport stream created by the data creation device 2601 is displayed. Can be played.
- the embodiments of the data creation device and the playback device according to the present invention have been described above. However, the illustrated data creation device and the playback device can be modified as follows, and the present invention is described in the above embodiment. Of course, the present invention is not limited to the data creation device and the playback device as shown in FIG.
- the decoded picture of the 2D compatible video stream having the same PTS value is overwritten on the decoded picture of the base-view video stream of Inter-view buffer 2816 in step S3003.
- the address reference destination may be changed without performing the overwriting process.
- the decoded picture of the base view is stored in the DPB 2818.
- the decoded picture of the base view video stream is not referred to, and may not be stored in the DPB 2818. In this way, it is possible to reduce the amount of memory for storing the pictures of the base-view video stream from the DPB 2818.
- the base-view video stream is generated by including it in the transport stream, and the decoding process of the pictures of the base-view video stream is performed in the decoding process. Processing may be omitted.
- the extended multi-view video decoder 2822 analyzes header information (eg, POC acquisition, picture type, ViewID, information acquisition with or without reference, etc.) and Inter-view buffer 2816 without decoding pictures of the base-view video stream. An area where one picture can be stored is secured.
- the extended multi-view video decoder 2822 stores the decoded picture output from the 2D compatible video decoder having the same value as the PTS and DTS obtained by header information analysis in the reserved area.
- an extended multi-view video decoder can manage an inter-view buffer 2816.
- a 2D compatible video stream including the above information may be generated.
- all or part of the syntax element of the base-view video stream is configured to be stored in the supplementary data of the 2D compatible video stream. That is, information for managing pictures in the Inter-view buffer buffer 2816 (the POC indicating the display order of the MPEG-4 MVC format, the slice_type indicating the picture type, the nal_ref_idc indicating the reference / non-reference of the picture, the base reference picture list) Ref_pic_list_mvc_modification, View ID of base-view video stream, MMCO command) and the like, which are information for creation, are included in the supplementary data of each picture of the 2D compatible video stream.
- the base-view video stream may not be multiplexed on the transport stream.
- the picture of the MPEG-4 MVC dependent view video stream directly refers to the picture of the MPEG-2 video stream.
- the data format is almost the same as the conventional one, so it is compatible with the encoding device and playback device compatible with the conventional MPEG-4 MVC format.
- the encoding device and the playback device corresponding to the video stream data of the present embodiment can be realized with good performance and a small improvement.
- the O2806 and DPB2818 memories are handled as separate areas, but the memory space may be shared as shown in FIG.
- the overwriting process can be omitted.
- Inter-view buffer 2816 and DPB 2818 are handled as separate buffers, but these may be the same buffer.
- the decoded picture of the base-view video stream having the same PTS and the same ViewID in DPB2818 may be replaced with the decoded picture of the 2D compatible video stream.
- the 2D compatible video stream and the base view video stream are P pictures (P3), but the picture of the dependent view video stream is a B picture (B3). ing.
- the pictures of the 2D compatible video stream, the base view video stream, and the dependent view video stream are all P pictures, and only the I picture and the P picture are included in all the video streams. Therefore, special reproduction processing with the I picture and P picture selected is facilitated.
- separate PIDs are set as the PID of each video stream in the multiplexing of the transport stream, but the base view video stream and the dependent view video stream are The same PID may be assigned.
- the access units of the video streams can be merged and transferred in accordance with the specifications of the compression encoding method of the multi-view video stream.
- the base-view video stream and the dependent video stream are merged according to the specifications of the compression coding system, and the playback device has a configuration as shown in FIG. 45, and the data transfer line of the extended multi-view video decoder is one. What is necessary is just to comprise.
- the base view video stream and the dependent view video stream may share header (for example, sequence header, picture header) information of each access unit storing pictures at the same time. That is, when header information is provided only in the base-view video stream and the dependent-view video stream is decoded, the header information necessary for decoding is decoded with reference to the header information of the base-view video stream. Also good. Therefore, addition of header information necessary for decoding can be omitted from the dependent-view video stream.
- header for example, sequence header, picture header
- the DTS values of the pictures of the 2D compatible video stream and the dependent view video stream at the same display time are equal to each other.
- the DTS of the picture of the pendant view video stream and the picture of the base view video stream are also set to the same value.
- the DTS values of pictures in each video stream at the same display time may not be the same.
- the DTS value of the 2D compatible video stream may be set to a value that is decoded earlier (for example, one frame before) than the base-view / dependent-view video stream.
- the inter-view buffer overwrite process and the dependent-view video stream picture decoding process can be performed with a margin. it can.
- the PTS of the picture of the 2D compatible video stream storing the parallax images at the same display time is the same as the PTS of the picture of the dependent view.
- the decoding process of the 2D compatible video stream is performed in advance. Therefore, the PTS of the pictures of the 2D compatible video stream at the same display time is set to be ahead of the base-view / dependent-view video stream (for example, one frame before).
- the PTS of the picture of the 2D compatible video stream is set to one frame before the PTS of the picture of the dependent view video stream.
- the pictures of the base-view video stream of the Inter-view buffer are exchanged, the pictures of the base-view video stream are exchanged with the pictures of the 2D compatible video stream of the PTS having a value indicating the previous frame.
- the 2D compatible video decoder 2821 outputs the decoded picture of the 2D compatible video stream to the first plane 2808 in accordance with the PTS value in step S3005.
- the extended multi-view video decoder 2822 may be configured to output both videos using the output plane switch 2819.
- the multiplexing method is the transport stream, but the present invention is not limited to this.
- the MP4 system format can be used as a multiplexing method.
- the input in FIG. 34 is separated into a 2D compatible video stream, a base-view video stream, and a dependent-view video stream as a file multiplexed by MP4 and decoded.
- the picture of the dependent-view video stream is decoded while referring to the picture in which the picture of the base-view video stream of Inter-view buffer 2816 is overwritten with the picture of the 2D-compatible video stream.
- time information in each access unit is specified based on the header information (stts, stsz, etc.) of the MP4 system format.
- the reference picture of the dependent-view video stream is a decoded picture of the 2D compatible video stream, and normal multi-view video Since it is different from the stream structure, the stream type and stream_id given to the PES packet header may be set to be different from the values in the case of the conventional multi-view video stream.
- the playback device can determine that the playback method is the 3D video playback method according to the present embodiment by referring to the stream type and stream_id, and can change the playback method.
- step S3003 When the inter-codec reference switch 2824 is ON as shown in FIG. 40, the data transfer described in step S3003 is performed from the 2D compatible video decoder to the inter-view buffer in the extended multi-view video decoder. In this case, data transfer is not performed.
- the plane selector 2825 includes a first plane 2808 in which picturer is output from the 2D compatible video decoder, a second plane 2820 in which a picture of the base-view video stream of the extended multi-view video decoder is output, and an extended multi-view video decoder dependent. Which plane is output as 2D video, 3D video left-eye image, or 3D video right-eye image to the third plane 2826 in which the view video stream picture is output select.
- the playback device 2823b changes the playback mode by switching the output from the inter-codec reference switch 2824 and the plane selector 2825 according to the playback method.
- the playback device 2823b turns off the inter-codec reference switch 2824. Then, the plane selector 2825 selects the first plane 2808 as 2D video.
- the playback device 2823b turns on the inter-codec reference switch 2824. Then, the plane selector 2825 selects the first plane 2808 or the second plane 2820 as the left-eye image, and selects the third plane 2826 as the right-eye image.
- the playback device 2823b turns off the inter-codec reference switch 2824. Then, the plane selector 2825 selects the second plane 2820 as the left-eye image and selects the third plane 2826 as the right-eye image.
- the playback device 2823b turns off the inter-codec reference switch 2824. Then, the plane selector 2825 selects the first plane 2808 as the left-eye image and selects the first plane 2808 as the right-eye image.
- the playback system when the playback system generates a transport stream that switches from 3D video playback by 2D compatible video stream and dependent-view video stream to 2D video playback of 2D compatible video stream Considering the delay of the decoding process, as shown in FIG. 42, the same video as the 2D compatible video stream may be compressed and encoded into the dependent-view video stream when the playback method changes. Good.
- a section in which the same video as the 2D compatible stream is compression-coded into a dependent-view video stream is a 2D transition period shown in the upper part of FIG. In this 2D transition section, since playback is performed as 2D video regardless of which method is used, smooth video transition is achieved for the user.
- the POC value of the picture of the dependent-view video stream indicating the same time display may be set to the value of temporal_reference indicating the display order included in each picture in MPEG-2 format compression encoding.
- the POC of the dependent-view video stream indicating the same time display may be included in the user data of each picture of the 2D compatible video stream.
- the value of temporal_reference can also be set to a unique value, so that the degree of freedom of setting in the compression encoding process increases.
- a high-quality filter 4301 may be applied to the decoding result of the 2D compatible video stream.
- the high image quality filter 4301 is a filter that reduces block noise like a deblocking filter defined by MPEG-4 AVC.
- a flag indicating whether to apply the high image quality filter 4301 is prepared. For example, the high image quality filter 430 is applied when the flag indicates ON, and is not applied when the flag is OFF.
- This flag can be included in the PMT descriptor and supplementary data of the stream.
- the playback device applies a filter to the inter-view buffer 2816 in the decoding result before transferring the data.
- This configuration realizes high quality 2D video for 2D compatible video streams.
- the dependent-view video stream can be decoded while referring to the picture that has been subjected to the high image quality processing, it is possible to improve the quality of the resulting 3D video.
- the extended multi-view video stream is configured so that a plurality of dependent-view video streams can be processed, and the PTS is the same in the replacement process with the picture of the 2D compatible video stream in the Inter-view buffer 2816.
- the base view picture having a value may be replaced.
- the ViewID to be replaced can be specified in the 2D compatible video stream, and instead of always replacing the picture of the base view, it may be selected and replaced from the pictures of the plurality of views.
- the 2D compatible video stream is described as an MPEG-2 format video stream
- the multi-view video stream base view video stream and dependent view video stream
- the type of compression encoding method is not limited to this.
- the configuration may be changed as appropriate according to the specifications of the compression encoding method.
- the 2D compatible video stream is MPEG-4 AVC and the multi-view video stream is a new compression encoding method
- O2806 of the playback device described in FIG. 34 is replaced with DPB, and the picture in the Inter-view buffer 2816 is managed. May be performed in accordance with the configuration of a novel compression encoding method.
- the left-eye picture and right-eye picture are arranged alternately and alternately in the vertical direction in the screen displayed on the display, and the pixels constituting the left-eye picture are passed through the upper lens called a lenticular lens on the display surface. Only the right eye and the pixels that form the picture for the right eye may be imaged only to the right eye, so that the left and right eyes can see a picture with parallax and can be displayed as a 3D image. Further, instead of the lenticular lens, a device having the same function, for example, a liquid crystal element may be used.
- a vertically polarized filter is installed on the left-eye pixel on the display side
- a horizontally polarized filter is installed on the right-eye pixel
- viewers install a vertically polarized filter on the left eye and a horizontally polarized filter on the right eye. You may use the system called the polarization system using the installed polarized glasses.
- a depth map in which a depth value is given in pixel units to the 2D video is prepared separately. You may make it produce
- FIG. 4 schematically shows an example of generating a parallax image of a left-eye image and a right-eye image from a 2D video image and a depth map.
- the depth map has a depth value corresponding to each pixel in the 2D image.
- the circular object in the 2D image is information indicating that the circular object is nearby in the depth map (the depth value is High) is assigned, and information indicating that the other area is farther from the circular object (depth value is Low).
- This information may be represented by a bit string for each pixel, or an image image (for example, “black” indicates that the depth is Low and “white” indicates that the depth is High). Also good.
- the parallax image can be created by adjusting the parallax amount of the 2D video from the depth value of the depth map. In the example of FIG.
- the parallax amount of the pixel of the circular object is increased when creating the parallax image.
- the depth value of the region other than the circular object is Low, the parallax amount of the pixels other than the circular object is reduced, and the left-eye image and the right-eye image are created. If the left-eye image and the right-eye image are displayed using a sequential separation method or the like, stereoscopic viewing is possible.
- a part or all of the constituent elements constituting each of the above-described devices may be constituted by one system LSI.
- a system LSI is an ultra-multifunctional LSI manufactured by integrating a plurality of components on a single chip. Specifically, a microprocessor, a ROM (Read Only Memory), a RAM (Random Access Memory), etc. It is a computer system comprised including. A computer program is stored in the RAM. The system LSI achieves its functions by the microprocessor operating according to the computer program.
- LSI Integrated Circuit
- IC Integrated Circuit
- the method of circuit integration is not limited to LSI, and may be realized by a dedicated circuit or a general-purpose processor.
- An FPGA Field Programmable Gate Array
- a reconfigurable processor that can reconfigure the connection and setting of the circuit cells inside the LSI may be used.
- the data creation device and the playback device may be a computer system including a microprocessor, a ROM, a RAM, a hard disk unit, and the like.
- a computer program is stored in the RAM or the hard disk unit. Each device achieves its function by the microprocessor operating according to the computer program.
- the computer program is configured by combining a plurality of instruction codes indicating instructions for the computer in order to achieve a predetermined function.
- the present invention may be the method described above. Further, the present invention may be a computer program that realizes these methods by a computer, or may be a digital signal composed of the computer program.
- the present invention also provides a computer-readable recording medium such as a flexible disk, hard disk, CD-ROM, MO, DVD, DVD-ROM, DVD-RAM, BD (Blu-ray Disc). ), Recorded in a semiconductor memory or the like. Further, the present invention may be the computer program or the digital signal recorded on these recording media.
- the computer program or the digital signal may be transmitted via an electric communication line, a wireless or wired communication line, a network represented by the Internet, a data broadcast, or the like.
- a data creation device includes a first frame image group that shows a view over time at a first viewpoint, and a second frame image group that shows a view over time at a second viewpoint.
- a data creation device for compression encoding comprising: a first encoding means for generating a stream in MPEG-2 format by compressing and encoding the first frame image group; a base conforming to the MPEG-4 MVC format; A second encoding unit that generates a view video stream and a dependent-view video stream; and a transmission unit that transmits the streams generated by the first and second encoding units, and the second encoding unit.
- each frame constituting the second frame image group is compressed and encoded using a frame of the MPEG-2 format stream at the same time as the corresponding frame of the base view video stream as a reference image. Generate stream.
- the data creation apparatus having this configuration can perform inter-view reference between video streams that have been compression-encoded with different encoding methods.
- the extended video stream is decoded while referring to the view of the 2D compatible video stream without referring to the picture of the base view stream.
- Possible video streams can be generated. Therefore, it is possible to generate a video stream with a reduced data amount while maintaining compatibility with the conventional compression encoding method.
- the second encoding means of the data creation device may generate a stream composed of the pseudo data by compressing and encoding a frame image group composed of the same frame image.
- the compression encoding of images in the MPEG-2 format and the MPEG-4 MVC format is very high in compression encoding efficiency for images with no difference between the preceding and succeeding images, so that compression encoding with reduced data amount is possible. . Since the data creation apparatus having this configuration uses, as dummy data, data obtained by compressing and encoding an image group that does not change over time, it is possible to generate a video stream that can reduce the data amount of the base-view video stream.
- the same frame image may be a frame image of a single color image.
- a monochrome (for example, black with no luminance) image is compression-encoded, a video stream with a reduced data amount can be generated.
- the second encoding means of the data creation device may generate the base-view video stream at a bit rate lower than the bit rate of the stream obtained by compressing and encoding the first frame image group. Good.
- the data creation apparatus having this configuration uses a bit rate lower than the bit rate used for compression encoding when generating the dependent-view video stream for compression encoding when generating the base-view video stream, the base-view video stream It is possible to generate a video stream that can reduce the amount of data.
- the second encoding unit of the data creation device may display a PTS (Presentation Time Stamp) of a picture constituting the frame of the dependent-view video stream at the same time as the corresponding frame of the base-view video stream.
- the compression encoding may be performed with the same value as the PTS of the picture constituting the frame of the MPEG-2 format stream.
- the data creation apparatus having this configuration can make the setting values of the PTSs of the compression-encoded pictures of the images corresponding to the first viewpoint frame image and the second viewpoint frame image the same. As a result, it is possible to easily know the correspondence relationship of pictures to be referenced in different compression coding systems, so that a video stream that facilitates inter-view reference between video streams compressed and coded by different compression coding systems is generated. can do.
- the first encoding unit and the second encoding unit of the data creation device perform compression encoding of the first frame image group and the second frame image group with the same aspect ratio, and generate The attribute information indicating the aspect ratio may be included in each stream.
- the data creation device having this configuration can make the aspect ratios of video streams generated by different compression encoding methods the same. Accordingly, it is possible to generate a video stream that can simplify the reference image conversion process at the time of decoding processing by inter-view reference between different compression-encoded video streams.
- the second encoding means of the data creation device may include the MPEG-2 format at the same time as the corresponding frame of the base-view video stream in the DTS of a picture constituting the frame of the dependent-view video stream.
- the compression coding may be performed with a value indicating a time after N frames (an integer value of N ⁇ 1) from the DTS value of the picture constituting the frame of the stream.
- the data creation apparatus uses the DTS of the picture obtained by compression-encoding the second frame as the DTS value of each picture when the frame images corresponding to the first viewpoint and the second viewpoint are compression-encoded. It can be set to indicate the time N frames after the DTS of a picture in which one frame is compression-encoded.
- the reference picture in the first stream data is decoded in advance, so that the writing process to the Inter-view buffer and the decoding process of the picture in the second stream data are performed. Can be performed with a margin.
- the second encoding unit of the data creation device may determine that the number of frames in the GOP (Group of Pictures) in the base-view video stream and the dependent-view video stream is the same time as the frame.
- the compression encoding may be performed with the same number as the number of frames in the GOP in the MPEG-2 format stream included.
- the data creating apparatus having this configuration can match the number of frame images in the GOP of each stream data containing the number of corresponding frame images of the first stream data and the second stream data. Thereby, since the heads of the GOPs of the pictures at the same display time of the first viewpoint and the second viewpoint coincide with each other, it is possible to generate a video stream that can be easily used for special reproduction such as jump reproduction.
- the second encoding means of the data creation device may, when the picture type of the picture constituting the frame of the MPEG-2 format stream is an I picture, the MPEG-2 format frame
- the pictures constituting the frame of the base-view video stream at the same time as the constituent pictures may be compression-coded as I pictures.
- the data creation device having this configuration can set each picture having the same value as the PTS of the picture in the first stream data and the PTS of the picture in the second stream data as an I picture. Therefore, since both the first stream and the second stream can be decoded from the PTS time of the same I picture, it is possible to generate a video stream that can simplify the decoding process during special reproduction such as jump reproduction.
- a playback device is included in the stream acquired by the data creation device according to the embodiment of the present invention and the stream acquired by the acquisition device.
- the second frame image group is obtained by decoding the dependent-view video stream included in the stream by using the frame decoded by the first decoding means at the same time as the corresponding frame of the video stream as a reference image.
- a reproduction means for reproducing an image group is included in the stream acquired by the data creation device according to the embodiment of the present invention and the stream acquired by the acquisition device.
- the playback device When decoding a dependent-view video stream compressed and encoded in the MPEG-4 MVC format, the playback device having this configuration compresses and encodes the MPEG-2 format instead of the frame image group obtained by decoding the base-view video stream.
- the decoding process can be performed with reference to the frame image group obtained by decoding the converted first stream data. Accordingly, the dependent-view video stream can be decoded while referring to the image group obtained by decoding different compression-encoded video stream data, so that the video stream is maintained while maintaining data compatibility with the conventional compression encoding method. As a whole, it is possible to reproduce a video stream with a reduced amount of data.
- the second decoding unit of the playback device may include a PTS picture having the same value as a PTS of a picture constituting a frame of the second frame image group among the frames decoded by the first decoding unit.
- the decoding process may be performed using a frame constituted by the reference image as a reference image.
- the playback device configured as described above is a picture in the video stream of the first viewpoint that is compression-encoded in the MPEG-2 format, and is a decoding target of the dependent-view video stream A frame image obtained by decoding a picture having the same value as the PTS of the picture is set as a reference image. Therefore, it is possible to reproduce the frame image of the second viewpoint with reference to the first frame image displayed at the same time. ⁇ 3.
- data amount compression encoding is performed using redundancy in the spatial direction and temporal direction of moving images.
- Inter-picture prediction coding is used as a compression coding method using redundancy.
- inter-picture predictive coding when a certain picture is coded, a picture that is forward or backward in display time order is used as a reference picture. Then, the amount of motion from the reference picture is detected, and the difference between the picture subjected to motion compensation and the picture to be coded is compression-coded.
- FIG. 1 shows a reference relationship between pictures in a video stream.
- picture P3 is compression-encoded with reference to picture I0
- picture B1 and picture B2 are compression-encoded with reference to picture I0 and picture P3, respectively.
- a right-eye image (R image) and a left-eye image (L image) are prepared, and only the image corresponding to each eye for the user is displayed for each eye.
- the stereoscopic display is realized by performing display controlled so as to be reflected on the screen.
- a moving image composed of right-eye images is called a left-view video
- a moving image composed of left-eye images is called a right-view video.
- FIG. 13 is a diagram showing an example of a stereoscopic image display.
- the example which displayed the image for left eyes of the skeleton of the dinosaur which is a target object, and the image for right eyes of the skeleton of the dinosaur which is a target object is shown. If 3D glasses are used to repeat the light and left eye light transmission and shading, the left and right scenes are overlapped by the afterimage reaction of the eyes in the user's brain, and a stereoscopic image exists on the extension line in the center of the face Then you can recognize.
- 3D video systems that compress and encode left-view video and right-view video include a frame compatible system and a multi-view encoding system.
- the frame compatible method is a method of performing compression coding by synthesizing each picture corresponding to an image showing a view at the same time of the left view video and the right view video and then synthesizing them into one picture.
- each picture corresponding to an image showing a view at the same time of the left-view video and the right-view video is compressed in half in the horizontal direction and then arranged in the left and right to form one picture.
- a video stream is generated by compressing and encoding a moving image based on a synthesized picture by a 2D video compression encoding method (for example, MPEG-2).
- the video stream is decoded based on the same compression encoding method as that for generating the video stream.
- Each decoded picture is divided into left and right images, and each of the divided images is expanded in the horizontal direction twice to obtain corresponding pictures of the left view video and the right view video.
- L image left-view video picture
- R image right-view video picture
- the multi-view encoding method is a method for compressing and encoding the left-view video and the right-view video picture as separate pictures without combining them into one picture.
- the multi-view encoding method is a method for compressing and encoding the left-view video and the right-view video picture as separate pictures without combining them into one picture.
- FIG. 2 is a diagram showing an encoding method in the MPEG-4 MVC format, which is a multi-view encoding method.
- the MPEG-4 MVC format video stream is processed simultaneously with the base-view video stream and the base-view video stream that can be played back by a playback device that plays back the conventional MPEG-4 AVC-format video stream.
- This is a configuration including a dependent-view video stream that realizes playback.
- the base-view video stream is compression-encoded using inter-picture predictive coding using only the redundancy with the video of the same viewpoint without referring to the video of another viewpoint. Is done.
- the dependent-view video stream is compression-encoded by inter-picture predictive encoding using redundancy with video from another viewpoint in addition to inter-picture predictive encoding using reference of the same viewpoint video. .
- the picture of the dependent-view video stream is compression-coded with reference to a picture showing a view at the same time in the base-view video stream.
- the picture P0 that is the first P picture of the dependent-view video stream refers to the picture I0 that is the I-picture of the base-view video stream.
- a picture B1 that is a B picture of the dependent-view video stream refers to a picture Br1 that is a Br picture of the base-view video stream.
- a picture P3 that is a second P picture of the dependent-view video stream indicates that a picture P3 that is a P-picture of the base-view video stream is referred to.
- the base-view video stream does not refer to the picture of the dependent-view video stream, it can be decoded and reproduced only by this video stream.
- the dependent-view video stream since the dependent-view video stream is decoded with reference to the base-view video stream, it cannot be played back alone.
- the dependent-view video stream performs inter-picture predictive coding using a base-view picture of another viewpoint indicating a view at the same time.
- the right-eye image and the left-eye image at the same time are similar (highly correlated), and the difference is compressed and encoded, so the dependent-view video stream has more data than the base-view video stream. Can be greatly reduced.
- MPEG-2 transport stream is a standard for multiplexing and transmitting various streams such as video and audio. It is standardized in ISO / IEC13818-1 and ITU-T recommendation H222.0.
- FIG. 6 is a diagram showing the structure of a digital stream in the MPEG-2 transport stream format.
- the transport stream 513 is obtained by multiplexing a video TS (Tranport Stream) packet 503, an audio TS packet 506, a subtitle stream TS packet 509, and the like.
- the video TS packet 503 includes the main video of the program
- the audio TS packet 506 includes the main audio portion and sub audio of the program
- the TS packet 509 of the subtitle stream includes subtitle information of the program.
- the video frame sequence 501 is compression encoded using a video encoding method such as MPEG-2, MPEG-4 AVC.
- the audio frame sequence 504 is compression-encoded by an audio encoding method such as Dolby AC-3, MPEG-2 AAC, MPEG-4 AAC, HE-AAC.
- Each stream included in the transport stream is identified by a stream identification ID called PID.
- PID stream identification ID
- the playback apparatus can extract the stream to be processed.
- Information on the correspondence between PID and stream is stored in a descriptor of a PMT packet described below.
- a video stream 501 composed of a plurality of video frames and an audio stream 504 composed of a plurality of audio frames are converted into PES packet sequences 502 and 505, respectively.
- the PES packet sequences 502 and 505 are converted into TS packets 503 and 506, respectively.
- the data of the subtitle stream 507 is converted into a PES packet sequence 508 and further converted into a TS packet 509.
- the MPEG-2 transport stream 513 is configured by multiplexing these TS packets into one stream. The PES packet and TS packet will be described later.
- the video stream is data having a hierarchical structure as shown in FIG.
- a video stream is composed of a plurality of GOPs, and by using this as a basic unit of compression encoding processing, editing of a moving image and random access are possible.
- GOP is composed of one or more video access units.
- the video access unit is a unit for storing compression-encoded data of a picture, and stores data of one frame in the case of a frame structure and one field in the case of a field structure.
- Each video access unit includes an AU identification code, a sequence header, a picture header, supplementary data, compressed picture data, padding data, a sequence end code, and a stream end code.
- Each of these data is stored in units called NAL units in the case of MPEG-4 AVC.
- AU identification code is a start code indicating the head of the access unit.
- the sequence header is a header storing common information in a playback sequence composed of a plurality of video access units, and stores information such as resolution, frame rate, aspect ratio, and bit rate.
- the picture header is a header that stores information such as the encoding method of the entire picture.
- Supplementary data is additional information that is not essential for decoding compressed picture data, and stores, for example, closed caption character information or GOP structure information that is displayed on the TV in synchronization with video.
- Compressed picture data stores compressed and encoded picture data.
- -Padding data stores data for formatting. For example, it is used as stuffing data for maintaining a predetermined bit rate.
- the sequence end code is data indicating the end of the playback sequence.
- the stream end code is data indicating the end of the bit stream.
- the data structure of the AU identification code, sequence header, picture header, supplemental data, compressed picture data, padding data, sequence end code, and stream end code differs depending on the video encoding method.
- the AU identification code is the AU delimiter (Access Unit Delimiter)
- the sequence header is SPS (Sequence Parameter Set)
- the picture header is PPS (Picture Parameter Set)
- the compressed picture data is multiple slices Supplementary data corresponds to SEI (Supplemental Enhancement Information), padding data corresponds to FillerData, sequence end code corresponds to End of Sequence, and stream end code corresponds to End of Stream.
- SEI Supplementary Enhancement Information
- padding data corresponds to FillerData
- sequence end code corresponds to End of Sequence
- stream end code corresponds to End of Stream.
- the sequence header is sequence_Header, sequence_extension, group_of_picture_header, the picture header is picture_header, picture_coding_extension, the compressed code data is a sequence_decode, and the supplement_data is a sequence_decode. To do. There is no AU identification code, but if the start code of each header is used, the break of the access unit can be determined.
- each attribute information is not always necessary.
- the sequence header is necessary only in the video access unit at the head of the GOP, and can be encoded without being added to other video access units.
- the picture header can be referred to that of the previous video access unit in the code order, and the picture header can be omitted in its own video access unit.
- I picture data is stored as compressed picture data, and an AU identification code, a sequence header, a picture header, and compressed picture data are always stored. Supplementary data, padding data, sequence end code, and stream end code are stored as needed.
- AU identification codes and compressed picture data are always stored in video access units other than the GOP head, and supplementary data, padding data, sequence end codes, and stream end codes are appropriately stored as necessary.
- FIG. 10 is a diagram showing how a video stream is stored in a PES packet sequence.
- the 10 shows the video frame sequence of the video stream.
- the second level shows a PES packet sequence.
- a plurality of Video Presentation Units in the video stream are divided into each picture, and stored in the payload of the PES packet. .
- Each PES packet has a PES header, and a PTS that is a picture display time and a DTS that is a picture decoding time are stored in the PES header.
- FIG. 11 is a diagram showing the data structure of TS packets constituting the transport stream.
- the TS packet is a 188-byte fixed-length packet composed of a 4-byte TS header, an adaptation field, and a TS payload.
- the TS header is configured to include transport_priority, PID, adaptation_field_control.
- the PID is an ID for identifying a stream multiplexed in the transport stream as described above.
- Transport_priority is information for identifying the type of packet in TS packets having the same PID.
- Adaptation_field_control is information for controlling the configuration of the adaptation field and the TS payload. There are cases where only one of the adaptation field and the TS payload exists or both, and adaptation_field_control is information indicating the presence / absence thereof.
- adaptation_field_control value When the adaptation_field_control value is “1”, only the TS payload is present. When the adaptation_field_control value is “2”, only the adaptation field is present. When the adaptation_field_control value is “3”, the payload is “3”. Indicates that both adaptation fields are present.
- the adaptation field is a storage area for storing information such as PCR (Program Clock Reference) and stuffing data for making the TS packet a fixed length of 188 bytes.
- PCR Program Clock Reference
- stuffing data for making the TS packet a fixed length of 188 bytes.
- a PES packet is divided and stored in the TS payload.
- TS packets included in the transport stream include packets such as PAT (Program Association Table), PMT, and PCR in addition to packets of each stream such as video, audio, and caption. These packets are called PSI (Program Specific Information).
- PAT indicates what the PID of the PMT used in the transport stream is, and the PID of the PAT itself is “0”.
- FIG. 12 is a diagram showing the data structure of the PMT.
- the PMT includes a PMT header, various descriptors related to the transport stream, and stream information related to each stream such as video / audio / subtitles included in the transport stream.
- copy control information for instructing permission / non-permission of copying of each stream such as video / audio is recorded.
- the stream information regarding each stream includes a stream descriptor in which a stream type, a stream PID, and stream attribute information (for example, a frame rate and an aspect ratio) are described in order to identify the compression encoding method of the stream.
- a stream descriptor in which a stream type, a stream PID, and stream attribute information (for example, a frame rate and an aspect ratio) are described in order to identify the compression encoding method of the stream.
- the PCR In order to synchronize the arrival time of the TS packet at the decoder and the STC (System Time Clock) which is the time axis of the PTS and DTS, the PCR has an STC time corresponding to the time when the PCR packet is transferred to the decoder. Have information.
- STC System Time Clock
- an area to be actually displayed can be designated as a cropping area from among the compression-coded frame areas.
- the frame_cropping information includes differences between the upper line, the lower line, the left line, and the right line of the cropping area and the upper line, the underline, the left line, and the right line of the compression-coded frame area. Is specified as the crop amount of the top, bottom, left and right.
- frame_cropping_flag is set to “1”
- frame_crop_top_offset, frame_crop_left_offset, frame_crop_right, and lower_crop_crop_right crop_crop
- the vertical and horizontal sizes of the cropping area (display_horizontal_size of sequence_display_extension, display_vertical_size), and the difference between the center of the compression-coded frame area and the center of the cropping area
- the cropping area is specified using information (frame_center_horizontal_offset, frame_center_vertical_offset of picture_display_extension).
- an aspect ratio is set as scaling information indicating a scaling method when the cropping area is actually displayed on a television or the like.
- the playback device uses the aspect ratio information to up-convert the cropping area for display.
- aspect ratio information (aspect_ratio_idc) is stored in the SPS as the scaling information.
- the aspect ratio is specified as 4: 3.
- sequence_header there is attribute information for storing aspect ratio information (aspect_ratio_information) called sequence_header, and the same processing as described above can be realized by appropriately setting the value of this attribute information.
- FIG. 15 is a diagram showing an example of the internal configuration of a video stream in the MPEG-4 MVC format.
- the picture of the right-view video stream is compression-coded with reference to the picture at the same display time of the left-view video stream.
- the pictures P1 and P2 of the right-view video stream refer to the pictures I1 and P2 of the left-view video stream, respectively.
- the pictures B3, B4, B6, and B7 of the right-view video stream are respectively pictures Br3 and Br4 of the left-view video stream. , Br6, Br7 with reference to compression coding.
- the second tier in the figure shows the internal structure of the left view video stream.
- This left-view video stream includes pictures I1, P2, Br3, Br4, P5, Br6, Br7, and P9. These pictures are decoded in the order of time set in the DTS.
- the first row shows the left eye image displayed on the display or the like.
- the left-eye image is the second-order decoded picture I1, P2, Br3, Br4, P5, Br6, Br7, P9 in the order of time set in the PTS, that is, I1, Br3, Br4, P2, Br6, Displayed in the order of Br7 and P5.
- the fourth level shows the internal structure of the right-view video stream.
- This right-view video stream includes pictures P1, P2, B3, B4, P5, B6, B7, and P8. These pictures are decoded in the order of time set in the DTS.
- the third row shows a right eye image displayed on a display or the like.
- the right-eye image has the decoded pictures P1, P2, B3, B4, P5, B6, B7, and P8 in the fourth stage in the order set in the PTS, that is, P1, B3, B4, P2, B6, Displayed in the order of B7 and P5.
- display of either the left-eye image or the right-eye image in which the same value is assigned to the PTS is delayed by half the interval until the next PTS time.
- the fifth row shows how the state of the 3D glasses 200 is changed. As shown in the fifth row, the right-eye shutter is closed when viewing the left-eye image, and the left-eye shutter is closed when viewing the right-eye image.
- FIG. 17 shows the configuration of the video access unit for each picture of the base-view video stream and each picture of the dependent-view video stream. As described above, as shown in the upper part of FIG. 17, in the base-view video stream, each picture is configured as one video access unit.
- the dependent-view video stream also has one video access unit in the same manner, but the data structure is different from the video access unit of the base-view video stream.
- the 3D video access unit 1701 is composed of the video access unit of the base-view video stream and the video access unit of the dependent-view video stream having the same PTS value.
- the playback device performs decoding in units of this 3D video access unit.
- FIG. 18 shows an example of the relationship between the PTS and the DTS assigned to each video access unit of the base-view video stream and the dependent-view video stream in the video stream.
- the base-view video stream picture and the dependent-view video stream picture storing the parallax images indicating the same time view are set so that the DTS and the PTS have the same value.
- the playback device that decodes the pictures of the base-view video stream and the dependent-view video stream can perform decoding and display in units of 3D video access units.
- FIG. 19 shows the GOP configuration of the base view video stream and the dependent view video stream.
- the GOP structure of the base view video stream is the same as that of the conventional video stream, and is composed of a plurality of video access units.
- the dependent view video stream is composed of a plurality of dependent GOPs.
- the leading picture of the dependent GOP is a picture displayed as a pair with the I picture at the GOP head of the base-view video stream when playing back 3D video, and has the same value as the PTS of the I picture at the GOP head of the base-view video stream This picture is assigned a PTS.
- FIG. 20 shows the data structure of the video access unit included in the dependent GOP.
- the dependent GOP head video access unit stores, as compressed picture data, picture data displayed at the same time as the GOP head I picture of the base-view video stream.
- the sequence header, picture header, and compressed picture data are always stored. Supplementary data, padding data, sequence end code, and stream end code are stored as appropriate.
- the sub AU identification code is a start code indicating the head of the access unit.
- the sub-sequence header is a header that stores common information in a playback sequence composed of a plurality of video access units, and stores information such as resolution, frame rate, aspect ratio, and bit rate.
- the frame rate, resolution, and aspect ratio values of the sub-sequence header are the same as the frame rate, resolution, and aspect ratio of the sequence header included in the video access unit at the GOP head of the corresponding base-view video stream.
- a video access unit other than the head of the GOP always stores a sub AU identification code and compressed picture data. Supplementary data, padding data, sequence end code, and stream end code are stored as appropriate.
- the video stream data creation device or playback device creates a stream corresponding to 3D video with a reduced amount of data while maintaining playback on an existing playback device that supports MPEG-2 format video streams. Or, since it can be played back, it is useful for an apparatus for generating and playing back 3D video.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
<1-1.概要>
本発明の一実施の形態に係る放送システムは、2D映像として、既存技術であるMPEG-2形式のストリームを生成し、3D映像として、MPEG-4 MVC形式を拡張した新規形式(本明細書において、この形式をMPEG-4 MVCに準拠という。)のベースビュービデオストリームとディペンデントビュービデオストリームとを生成し、送出する。
図21は、本実施の形態における放送システムによって作成されたトランスポートストリームのデータ構造を示している。同図に示されるように、トランスポートストリームは、2D互換ビデオストリームAとマルチビュービデオストリームBとからなる。後者のマルチビュービデオストリームBは、ベースビュービデオストリームB1と、ディペンデントビュービデオストリームB2とからなる。
前記2D互換ビデオストリームAは、左目用画像を圧縮符号化して作成され、ベースビュービデオストリームB1は、黒などの単色映像(以下、「黒画像」という。)を圧縮符号化して作成される。
更に、ディペンデントビュービデオストリームB2は、左目用画像と右目用画像の差分を圧縮符号化して生成される。先に述べたようにベースビュービデオストリームB1は、黒画像を圧縮符号化したものであるので、ディペンデントビュービデオストリームB2を生成する際の参照画像として用いることができない。この点が、既存のMPEG-4 MVC形式と異なっているところであり、参照画像は、2D互換ビデオストリームAの同時刻のフレーム画像に設定されている。
MPEG-4 MVC形式で圧縮符号化されたストリームの復号処理では、ベースビュービデオストリームのフレーム画像を参照してディペンデントビュービデオストリームを復号するが、本実施の形態では、MPEG-2互換ストリーム、すなわち、左目用画像のフレーム画像を参照画像として、ディペンデントビュービデオストリームを復号できるようにしている。
MPEG-4 MVCに準拠した形式では、具体的には、復号の際の参照先を、ベースビュービデオストリームから、MPEG-2互換ビデオストリームに変更することを再生側に指示するディスクリプタ等を規定している。
以下、本発明の実施の形態に係るデータ作成装置及び再生装置について図面を参照しながら説明する。
<1-2.データ作成装置>
<1-2-1.構成>
以下、本発明に係るデータ作成装置の一実施形態について図を参照しながら説明する。
復号ピクチャは、拡張マルチビュービデオエンコーダ2604の2D互換ビデオフレームメモリ2608に格納される。また、2D互換ビデオ符号化情報2606は、ベースビュービデオエンコーダ2605に入力される。
ディペンデントビュービデオエンコーダ2609は、ビュー間参照を実行するための参照ピクチャIDを、ベースビュービデオ符号化情報2607のベースビュービデオフレームメモリ管理情報に基づいて決定する。さらに、ディペンデントビュービデオエンコーダ2609は、ディペンデントビュービデオストリームのビデオの属性情報をベースビュービデオ符号化情報2607に含まれるベースビュービデオストリームの属性情報の値と同じ値を設定する。
次に、データフォーマットについて、図面を参照しながら説明を行う。
図27はデータ作成装置2601のデータ作成フローを示す図である。以下、そのデータ作成フローを説明する。
<1-3.再生装置>
<1-3-1.構成>
次に本実施の形態に係る3D映像を再生する再生装置2823の構成について図を参照しながら説明する。
図30は再生装置2823の3D映像における復号処理と出力処理のフローを示す図である。
再生装置2823のD1(2805)で行う復号処理は、既存のMPEG-2形式のビデオストリームの復号処理と同じなので、既存のMPEG-2形式のビデオの再生装置のLSI(Large Scale Integration)やソフトウェアを利用することができる。また、D2(2817)で行うMPEG-4 MVC形式の復号処理についても、既存のMPEG-4 MVC形式の処理と同じであるので、既存のMPEG-4 MVC形式のビデオの再生装置のLSIやソフトウェアを利用できる。
<再生装置2823の使用態様の一例>
図5を用いて、データ作成装置2823で作成したビデオストリームの3D映像を再生できる3Dデジタルテレビ100と、3D映像の再生をサポートしない従来の2D映像のみを再生できる2Dデジタルテレビ300を例に挙げて説明する。
<2.変形例>
以上、本発明に係るデータ作成装置及び再生装置の実施の形態を説明したが、例示したデータ作成装置及び再生装置を以下のように変形することも可能であり、本発明が上述の実施の形態で示した通りのデータ作成装置及び再生装置に限らないことは勿論である。
図24は、特殊再生について説明するための図である。図24上段は、上記の制約が課されていない場合について示している。この場合、表示順で3番目のピクチャにおいて、2D互換ビデオストリームとベースビュービデオストリームは、Pピクチャ(P3)になっているが、ディペンデントビュービデオストリームのピクチャはBピクチャ(B3)となっている。
一方、前記ディペンデントビューストリームとして、前記ベースビュービデオストリームの対応するフレームと同時刻の前記MPEG-2形式のストリームのフレームを参照画像として前記第2フレーム画像群を構成する各フレームを圧縮符号化したストリームを生成する。
この構成のデータ作成装置は、異なる符号化方式で圧縮符号化されたビデオストリーム間でビュー間参照をすることができる。これにより、3D映像を構成する拡張ビデオストリームにおいて、ディペンデントビュービデオストリームを復号する際に、ベースビューストリームのピクチャを参照せず、2D互換ビデオストリームのビューを参照しながら拡張ビデオストリームを復号可能なビデオストリームを生成することができる。従って、従来の圧縮符号化方式と互換性を保ちながらデータ量を削減したビデオストリームを生成することができる。
これにより、単色(例えば、輝度のない黒色)の画像を圧縮符号化するので、データ量を削減したビデオストリームを生成することができる。
<3.補足説明>
<映像圧縮技術の説明>
<2D映像圧縮技術>
本実施の形態に係るデータ作成装置及び再生装置で使用する2D映像を圧縮符号化する規格であるMPEG-2、及びMPEG-4 AVC(MPEG-4 MVC形式のベースとなる圧縮符号化方式)の符号化方法について簡単に説明する。
次に、視差画像を用いて3D映像としてディスプレイ等で再生する方法、マルチビュー符号化方式であるMPEG-4 MVC形式の圧縮符号化について簡単に説明する。
デジタルテレビの放送波等での伝送では、MPEG-2トランスポートストリーム形式のデジタルストリームが使われている。
次に、映像を上述の符号化方式で圧縮符号化して得られるビデオストリームのデータ構成について説明する。
次に、MPEG-4 MVC形式のビデオストリームについて説明する。
2602 2D互換ビデオエンコーダ
2603 2D互換ビデオデコーダ
2604 拡張マルチビュービデオエンコーダ
2605 ベースビュービデオエンコーダ
2606 2D互換ビデオ符号化情報
2607 ベースビュービデオ符号化情報
2608 2D互換ビデオフレームメモリ
2609 ディペンデントビュービデオエンコーダ
2610 マルチプレクサ
2801 PIDフィルタ
2802 TB(1)
2803 MB(1)
2804 EB(1)
2805 D1(2D互換ビデオ圧縮映像デコーダ)
2806 O(Re-ordering Buffer)
2807 スイッチ
2808 第1プレーン
2809 TB(2)
2810 MB(2)
2811 EB(2)
2812 TB(3)
2813 MB(3)
2814 EB(3)
2815 デコードスイッチ
2816 Inter-view buffer
2817 D2(マルチビュービデオ圧縮映像デコーダ)
2818 DPB
2819 出力プレーンスイッチ
2820 第2プレーン
2821 2D互換ビデオデコーダ
2822 拡張マルチビュービデオデコーダ
2823 再生装置
2901 ビデオデコーダ
2902 マルチビュービデオデコーダ
Claims (19)
- 第1視点における経時的なビューを示す第1フレーム画像群と、第2視点における経時的なビューを示す第2フレーム画像群とを圧縮符号化するデータ作成装置であって、
前記第1フレーム画像群を圧縮符号化することによりMPEG-2形式のストリームを生成する第1符号化手段と、
MPEG-4 MVC形式に準拠し、ベースビュービデオストリームとディペンデントビュービデオストリームとを生成する第2符号化手段と、
前記第1、第2符号化手段により生成されたストリームを送出する送出手段と
を備え、
前記第2符号化手段は、前記ベースビュービデオストリームとして、前記第1フレーム画像群を圧縮符号化して得られるストリームとフレーム数が同一であって、総データ量を少なくした疑似データからなるストリームを生成し、
一方、前記ディペンデントビュービデオストリームとして、前記ベースビュービデオストリームの対応するフレームと同時刻の前記MPEG-2形式のストリームのフレームを参照画像として前記第2フレーム画像群を構成する各フレームを圧縮符号化したストリームを生成する
ことを特徴とするデータ作成装置。 - 前記第2符号化手段は、
前記疑似データからなるストリームを、同一のフレーム画像から構成されるフレーム画像群を圧縮符号化して生成する
ことを特徴とする請求項1記載のデータ作成装置。 - 前記同一のフレーム画像は、単色画像のフレーム画像である
ことを特徴とする請求項2記載のデータ作成装置。 - 前記第2符号化手段は、
前記第1フレーム画像群を圧縮符号化して得られるストリームのビットレートより低いビットレートで、前記ベースビュービデオストリームを生成する
ことを特徴とする請求項1記載のデータ作成装置。 - 前記第2符号化手段は、
ディペンデントビュービデオストリームのフレームを構成するピクチャのPTS(Presentation Time Stamp)を、前記ベースビュービデオストリームの対応するフレームと同時刻のMPEG-2形式のストリームのフレームを構成するピクチャのPTSと同じ値にして圧縮符号化する
ことを特徴とする請求項3記載のデータ作成装置。 - 前記第1符号化手段及び前記第2符号化手段は、
前記第1フレーム画像群及び前記第2フレーム画像群の圧縮符号化を同一のアスペクト比で行い、生成する前記各ストリームに前記アスペクト比を示す属性情報を含める
ことを特徴とする請求項5記載のデータ作成装置。 - 前記第2符号化手段は、
ディペンデントビュービデオストリームのフレームを構成するピクチャのDTSに、前記ベースビュービデオストリームの対応するフレームと同時刻のMPEG-2形式のストリームのフレームを構成するピクチャのDTSの値よりNフレーム分(N≧1の整数値)後の時刻を示す値にして圧縮符号化する
ことを特徴とする請求項5記載のデータ作成装置。 - 前記第2符号化手段は、
ベースビュービデオストリーム及びディペンデントビュービデオストリームにおけるGOP(Group of Pictures)内のフレームの数が、
前記フレームと同時刻のフレームを含む前記MPEG-2形式のストリームにおけるGOP内のフレームの数と同じ数にして圧縮符号化する
ことを特徴とする請求項5記載のデータ作成装置。 - 前記第2符号化手段は、
前記MPEG-2形式のストリームのフレームを構成するピクチャのピクチャタイプがIピクチャである場合には、前記MPEG-2形式のフレームを構成するピクチャと同時刻のベースビュービデオストリームのフレームを構成するピクチャをIピクチャとして圧縮符号化する
ことを特徴とする請求項5記載のデータ作成装置。 - 請求項1記載のデータ作成装置で送出されたストリームを取得する取得手段と、
前記取得手段により取得された前記ストリームに含まれるMPEG-2形式のストリームを復号することにより前記第1フレーム画像群を得る第1復号手段と、
前記取得手段により取得された前記ストリームに含まれるベースビュービデオストリームを復号し、かつ、前記ベースビュービデオストリームの対応するフレームと同時刻の前記第1復号手段により復号されたフレームを参照画像として、前記ストリームに含まれるディペンデントビュービデオストリームを復号することにより、前記第2フレーム画像群を得る第2復号手段と、
前記第1、第2復号手段により得られた前記第1、第2フレーム画像群を再生する再生手段と
を備える
ことを特徴とする再生装置。 - 前記第2復号手段は、
前記第1復号手段により復号されたフレームのうち、前記第2フレーム画像群のフレームを構成するピクチャのPTSと同じ値のPTSのピクチャで構成されるフレームを参照画像として、復号処理を行う
ことを特徴とする請求項10記載の再生装置。 - 第1視点における経時的なビューを示す第1フレーム画像群と、第2視点における経時的なビューを示す第2フレーム画像群とを圧縮符号化するデータ符号化方法であって、
前記第1フレーム画像群を圧縮符号化することによりMPEG-2形式のストリームを生成する第1符号化ステップと、
MPEG-4 MVC形式に準拠し、ベースビュービデオストリームとディペンデントビュービデオストリームとを生成する第2符号化ステップと、
前記第1、第2符号化ステップにより生成されたストリームを送出する送出ステップと
を含み、
前記第2符号化ステップは、前記ベースビュービデオストリームとして、前記第1フレーム画像群を圧縮符号化して得られるストリームとフレーム数が同一であって、総データ量を少なくした疑似データからなるストリームを生成し、
一方、前記ディペンデントビューストリームとして、前記ベースビュービデオストリームの対応するフレームと同時刻の前記MPEG-2形式のストリームのフレームを参照画像として前記第2フレーム画像群を構成する各フレームを圧縮符号化したストリームを生成する
ことを特徴とするデータ符号化方法。 - 第1視点における経時的なビューを示す第1フレーム画像群と、第2視点における経時的なビューを示す第2フレーム画像群とを圧縮符号化するデータ作成装置に圧縮符号化処理を実行させるためのプログラムであって、
前記第1フレーム画像群を圧縮符号化することによりMPEG-2形式のストリームを生成する第1符号化ステップと、
MPEG-4 MVC形式に準拠し、ベースビュービデオストリームとディペンデントビュービデオストリームとを生成する第2符号化ステップと、
前記第1、第2符号化ステップにより生成されたストリームを送出する送出ステップと
を含み、
前記第2符号化ステップは、前記ベースビュービデオストリームとして、前記第1フレーム画像群を圧縮符号化して得られるストリームとフレーム数が同一であって、総データ量を少なくした疑似データからなるストリームを生成し、
一方、前記ディペンデントビューストリームとして、前記ベースビュービデオストリームの対応するフレームと同時刻の前記MPEG-2形式のストリームのフレームを参照画像として前記第2フレーム画像群を構成する各フレームを圧縮符号化したストリームを生成する
ことを特徴とするプログラム。 - 第1視点における経時的なビューを示す第1フレーム画像群と、第2視点における経時的なビューを示す第2フレーム画像群とを圧縮符号化するデータ作成装置に圧縮符号化処理を実行させるためのプログラムを記録した記録媒体であって、
前記データ作成装置に、
前記第1フレーム画像群を圧縮符号化することによりMPEG-2形式のストリームを生成する第1符号化ステップと、
MPEG-4 MVC形式に準拠し、ベースビュービデオストリームとディペンデントビュービデオストリームとを生成する第2符号化ステップと、
前記第1、第2符号化ステップにより生成されたストリームを送出する送出ステップと
を実行させ、
前記第2符号化ステップは、前記ベースビュービデオストリームとして、前記第1フレーム画像群を圧縮符号化して得られるストリームとフレーム数が同一であって、総データ量を少なくした疑似データからなるストリームを生成し、
一方、前記ディペンデントビューストリームとして、前記ベースビュービデオストリームの対応するフレームと同時刻の前記MPEG-2形式のストリームのフレームを参照画像として前記第2フレーム画像群を構成する各フレームを圧縮符号化したストリームを生成する
プログラムを記録した記録媒体。 - 第1視点における経時的なビューを示す第1フレーム画像群と、第2視点における経時的なビューを示す第2フレーム画像群とを圧縮符号化するデータ作成装置を構成する集積回路であって、
前記第1フレーム画像群を圧縮符号化することによりMPEG-2形式のストリームを生成する第1符号化手段と、
MPEG-4 MVC形式に準拠し、ベースビュービデオストリームとディペンデントビュービデオストリームとを生成する第2符号化手段と、
前記第1、第2符号化手段により生成されたストリームを送出する送出手段と
を備え、
前記第2符号化手段は、前記ベースビュービデオストリームとして、前記第1フレーム画像群を圧縮符号化して得られるストリームとフレーム数が同一であって、総データ量を少なくした疑似データからなるストリームを生成し、
一方、前記ディペンデントビューストリームとして、前記ベースビュービデオストリームの対応するフレームと同時刻の前記MPEG-2形式のストリームのフレームを参照画像として前記第2フレーム画像群を構成する各フレームを圧縮符号化したストリームを生成する
ことを特徴とする集積回路。 - 請求項1記載のデータ作成装置で送出されたストリームを取得する取得ステップと、
前記取得ステップにより取得された前記ストリームに含まれるMPEG-2形式のストリームを復号することにより、前記第1フレーム画像群を得る第1復号ステップと、
前記取得ステップにより取得された前記ストリームに含まれるベースビュービデオストリームを復号し、かつ、前記ベースビュービデオストリームの対応するフレームと同時刻の前記第1復号ステップにより復号されたフレームを参照画像として、前記ストリームに含まれるディペンデントビュービデオストリームを復号することにより、前記第2フレーム画像群を得る第2復号ステップと、
前記第1、第2復号ステップにより得られた前記第1、第2フレーム画像群を再生する再生ステップと
を含む
ことを特徴とする再生方法。 - 請求項1記載のデータ作成装置で送出されたストリームの再生処理を再生装置に実行させるためのプログラムであって、
前記再生装置に、
前記送出されたストリームを取得する取得ステップと、
前記取得ステップにより取得された前記ストリームに含まれるMPEG-2形式のストリームを復号することにより、前記第1フレーム画像群を得る第1復号ステップと、
前記取得ステップにより取得された前記ストリームに含まれるベースビュービデオストリームを復号し、かつ、前記ベースビュービデオストリームの対応するフレームと同時刻の前記第1復号ステップにより復号されたフレームを参照画像として、前記ストリームに含まれるディペンデントビュービデオストリームを復号することにより、前記第2フレーム画像群を得る第2復号ステップと、
前記第1、第2復号ステップにより得られた前記第1、第2フレーム画像群を再生する再生ステップと
を実行させる
ことを特徴とするプログラム。 - 請求項1記載のデータ作成装置で送出されたストリームの再生処理を再生装置に実行させるためのプログラムを記録した記録媒体であって、
前記再生装置に、
前記送出されたストリームを取得する取得ステップと、
前記取得ステップにより取得された前記ストリームに含まれるMPEG-2形式のストリームを復号することにより、前記第1フレーム画像群を得る第1復号ステップと、
前記取得ステップにより取得された前記ストリームに含まれるベースビュービデオストリームを復号し、かつ、前記ベースビュービデオストリームの対応するフレームと同時刻の前記第1復号ステップにより復号されたフレームを参照画像として、前記ストリームに含まれるディペンデントビュービデオストリームを復号することにより、前記第2フレーム画像群を得る第2復号ステップと、
前記第1、第2復号ステップにより得られた前記第1、第2フレーム画像群を再生する再生ステップと
を実行させるプログラムを記録した記録媒体。 - 請求項1記載のデータ作成装置で送出されたストリームを再生する再生装置を構成する集積回路であって、
前記送出されたストリームを取得する取得手段と、
前記取得手段により取得された前記ストリームに含まれるMPEG-2形式のストリームを復号することにより前記第1フレーム画像群を得る第1復号手段と、
前記取得手段により取得された前記ストリームに含まれるベースビュービデオストリームを復号し、かつ、前記ベースビュービデオストリームの対応するフレームと同時刻の前記第1復号手段により復号されたフレームを参照画像として、前記ストリームに含まれるディペンデントビュービデオストリームを復号することにより、前記第2フレーム画像群を得る第2復号手段と、
前記第1、第2復号手段により得られた前記第1、第2フレーム画像群を再生する再生手段と
を備える
ことを特徴とする集積回路。
Priority Applications (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
MX2013008310A MX2013008310A (es) | 2011-02-10 | 2012-02-10 | Dispositivo de creacion de datos y dispositivo de reproduccion para imagen de video en flujo de video. |
CA2823289A CA2823289A1 (en) | 2011-02-10 | 2012-02-10 | Data creation device and playback device for video picture in video stream |
BR112013020175A BR112013020175A2 (pt) | 2011-02-10 | 2012-02-10 | dispositivo de criações de dados e dispositivo de reprodução para imagem de vídeo em fluxo de vídeo |
US13/983,657 US9025941B2 (en) | 2011-02-10 | 2012-02-10 | Data creation device and playback device for video picture in video stream |
JP2012556803A JP6008292B2 (ja) | 2011-02-10 | 2012-02-10 | ビデオストリームの映像のデータ作成装置及び再生装置 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201161441383P | 2011-02-10 | 2011-02-10 | |
US61/441,383 | 2011-02-10 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2012108210A1 true WO2012108210A1 (ja) | 2012-08-16 |
Family
ID=46638430
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2012/000899 WO2012108210A1 (ja) | 2011-02-10 | 2012-02-10 | ビデオストリームの映像のデータ作成装置及び再生装置 |
Country Status (7)
Country | Link |
---|---|
US (1) | US9025941B2 (ja) |
JP (1) | JP6008292B2 (ja) |
BR (1) | BR112013020175A2 (ja) |
CA (1) | CA2823289A1 (ja) |
MX (1) | MX2013008310A (ja) |
TW (1) | TW201242373A (ja) |
WO (1) | WO2012108210A1 (ja) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TWI596933B (zh) * | 2013-03-13 | 2017-08-21 | 蘋果公司 | 用於快速切換的編碼解碼器技術 |
Families Citing this family (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101803970B1 (ko) * | 2011-03-16 | 2017-12-28 | 삼성전자주식회사 | 컨텐트를 구성하는 장치 및 방법 |
CN107534790B (zh) * | 2015-05-19 | 2020-06-23 | 索尼公司 | 图像处理装置、图像处理方法、接收装置及发送装置 |
US12212751B1 (en) | 2017-05-09 | 2025-01-28 | Cinova Media | Video quality improvements system and method for virtual reality |
US10944971B1 (en) * | 2017-05-22 | 2021-03-09 | Cinova Media | Method and apparatus for frame accurate field of view switching for virtual reality |
US10594758B2 (en) * | 2017-12-15 | 2020-03-17 | Cisco Technology, Inc. | Latency reduction by sending audio and metadata ahead of time |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004040455A (ja) * | 2002-07-03 | 2004-02-05 | Nippon Television Network Corp | 伝送再生方法、伝送方法、及びそのシステム |
JP2008500790A (ja) * | 2004-05-21 | 2008-01-10 | エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート | 3次元立体ビデオ付加データを用いた3次元立体デジタル放送の送/受信装置及びその方法 |
WO2010067810A1 (ja) * | 2008-12-09 | 2010-06-17 | ソニー株式会社 | 画像処理装置および方法 |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7133449B2 (en) * | 2000-09-18 | 2006-11-07 | Broadcom Corporation | Apparatus and method for conserving memory in a fine granularity scalability coding system |
KR100392384B1 (ko) * | 2001-01-13 | 2003-07-22 | 한국전자통신연구원 | 엠펙-2 데이터에 엠펙-4 데이터를 동기화시켜 전송하는장치 및 그 방법 |
JP2006237656A (ja) * | 2003-02-28 | 2006-09-07 | Secom Co Ltd | 符号化信号分離・合成装置、差分符号化信号生成装置、差分符号化信号抽出装置、符号化信号分離・合成方法、符号化信号分離・合成プログラム |
KR101366092B1 (ko) * | 2006-10-13 | 2014-02-21 | 삼성전자주식회사 | 다시점 영상의 부호화, 복호화 방법 및 장치 |
JP2010245970A (ja) * | 2009-04-08 | 2010-10-28 | Sony Corp | 再生装置、再生方法、およびプログラム |
-
2012
- 2012-02-10 MX MX2013008310A patent/MX2013008310A/es active IP Right Grant
- 2012-02-10 WO PCT/JP2012/000899 patent/WO2012108210A1/ja active Application Filing
- 2012-02-10 US US13/983,657 patent/US9025941B2/en not_active Expired - Fee Related
- 2012-02-10 TW TW101104353A patent/TW201242373A/zh unknown
- 2012-02-10 BR BR112013020175A patent/BR112013020175A2/pt not_active IP Right Cessation
- 2012-02-10 CA CA2823289A patent/CA2823289A1/en not_active Abandoned
- 2012-02-10 JP JP2012556803A patent/JP6008292B2/ja not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2004040455A (ja) * | 2002-07-03 | 2004-02-05 | Nippon Television Network Corp | 伝送再生方法、伝送方法、及びそのシステム |
JP2008500790A (ja) * | 2004-05-21 | 2008-01-10 | エレクトロニクス アンド テレコミュニケーションズ リサーチ インスチチュート | 3次元立体ビデオ付加データを用いた3次元立体デジタル放送の送/受信装置及びその方法 |
WO2010067810A1 (ja) * | 2008-12-09 | 2010-06-17 | ソニー株式会社 | 画像処理装置および方法 |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TWI596933B (zh) * | 2013-03-13 | 2017-08-21 | 蘋果公司 | 用於快速切換的編碼解碼器技術 |
Also Published As
Publication number | Publication date |
---|---|
MX2013008310A (es) | 2013-09-06 |
US20130315573A1 (en) | 2013-11-28 |
JP6008292B2 (ja) | 2016-10-19 |
CA2823289A1 (en) | 2012-08-16 |
BR112013020175A2 (pt) | 2016-11-08 |
JPWO2012108210A1 (ja) | 2014-07-03 |
US9025941B2 (en) | 2015-05-05 |
TW201242373A (en) | 2012-10-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2012111325A1 (ja) | 映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム | |
JP5906462B2 (ja) | 映像符号化装置、映像符号化方法、映像符号化プログラム、映像再生装置、映像再生方法及び映像再生プログラム | |
JP6229962B2 (ja) | 符号化装置及び符号化方法 | |
JP5336666B2 (ja) | 符号化方法、表示装置、復号方法 | |
US9357200B2 (en) | Video processing device and video processing method | |
JP6008292B2 (ja) | ビデオストリームの映像のデータ作成装置及び再生装置 | |
WO2013099290A1 (ja) | 映像再生装置、映像再生方法、映像再生プログラム、映像送信装置、映像送信方法及び映像送信プログラム | |
WO2012169204A1 (ja) | 送信装置、受信装置、送信方法及び受信方法 | |
US20140078256A1 (en) | Playback device, transmission device, playback method and transmission method | |
JP2011211605A (ja) | 画像再生装置、画像再生方法及び画像再生用プログラム |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 12744696 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2823289 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2013/008310 Country of ref document: MX |
|
ENP | Entry into the national phase |
Ref document number: 2012556803 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 13983657 Country of ref document: US |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
REG | Reference to national code |
Ref country code: BR Ref legal event code: B01A Ref document number: 112013020175 Country of ref document: BR |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 12744696 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 112013020175 Country of ref document: BR Kind code of ref document: A2 Effective date: 20130807 |