WO2010146847A1 - 3d映像を再生するための情報記録媒体、及び再生装置 - Google Patents
3d映像を再生するための情報記録媒体、及び再生装置 Download PDFInfo
- Publication number
- WO2010146847A1 WO2010146847A1 PCT/JP2010/003998 JP2010003998W WO2010146847A1 WO 2010146847 A1 WO2010146847 A1 WO 2010146847A1 JP 2010003998 W JP2010003998 W JP 2010003998W WO 2010146847 A1 WO2010146847 A1 WO 2010146847A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- stream
- video
- plane
- file
- information
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N19/00—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
- H04N19/50—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
- H04N19/597—Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/445—Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/10—Indexing; Addressing; Timing or synchronising; Measuring tape travel
- G11B27/19—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier
- G11B27/28—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording
- G11B27/30—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording
- G11B27/3027—Indexing; Addressing; Timing or synchronising; Measuring tape travel by using information detectable on the record carrier by using information signals recorded by the same method as the main recording on the same track as the main recording used signal is digitally coded
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/183—On-screen display [OSD] information, e.g. subtitles or menus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/189—Recording image signals; Reproducing recorded image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
- H04N5/92—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
- H04N9/8227—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being at least another television signal
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2537—Optical discs
- G11B2220/2541—Blu-ray discs; Blue laser DVR discs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/775—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television receiver
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/84—Television signal recording using optical recording
- H04N5/85—Television signal recording using optical recording on discs or drums
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/7921—Processing of colour television signals in connection with recording for more than one processing mode
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/8042—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/806—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal
- H04N9/8063—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal using time division multiplex of the PCM audio and PCM video signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
- H04N9/8211—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being a sound signal
Definitions
- the present invention belongs to the technical field of 3D video and 2D video playback technology.
- 2D video is an image expressed by pixels on the X-Y plane when the display screen of the display device is regarded as the X-Y plane, and is also called a planar view image.
- 3D video is an image obtained by adding the depth in the Z-axis direction to the pixels on the XY plane on the screen of the display device.
- 3D video can be viewed by the user by playing both the left-view video that should be viewed with the left eye and the right-view video that should be viewed with the right eye, and exhibiting the stereoscopic effect in these left-view video and right-view video.
- the user feels that the pixel having the positive Z-axis coordinate is in front of the screen of the display device, and feels that the pixel having the negative Z-axis coordinate exists behind the screen.
- 2D playback device When storing 3D video on an optical disc, playback compatibility with a playback device that can play back only the optical disc storing 2D video (hereinafter referred to as “2D playback device”) is required. If the 2D playback device cannot play back 3D video as 2D video from an optical disc that stores 3D video, it is necessary to produce two types of 3D discs and 2D discs for the same content, resulting in high costs. . Therefore, an optical disc containing 3D video is played back as 2D video on a 2D playback device, and as 2D video or 3D video on a playback device that can play back 2D video and 3D video (hereinafter referred to as “2D / 3D playback device”). It must be reproducible.
- Patent Document 1 As a prior art of a technology for ensuring reproduction compatibility with an optical disc storing 3D video, there is one described in Patent Document 1 below.
- subtitle data is stored on an optical disc in a movie work or the like, and subtitles are usually displayed in combination with moving images during playback.
- subtitle characters occupy a lot of display area on the screen if the dialogue and narration in the play are long.
- the subtitle overlaps with the stereoscopic display of the video and a stereoscopic video that is very difficult to see is played back.
- the subtitle characters are placed close to the edge of the screen so that the stereoscopic display of the video does not overlap with the subtitles.
- the stereoscopic effect differs greatly depending on where it is played in the playback section, and the amount of subtitle characters often varies greatly depending on the language, so one end of the screen is fixed as a subtitle display area. As a result, the screen usage efficiency decreases. Then, even if an expensive large screen 3D television is purchased by spending private expenses, the user may not be able to enjoy the stereoscopic effect as much as possible.
- An object of the present invention is to provide a recording medium capable of avoiding a reduction in stereoscopic effect due to a decrease in screen use efficiency.
- the recording medium of the present invention is a recording medium on which a video stream constituting a stereoscopic image, playlist information, and a plurality of subtitle streams are recorded.
- the playlist information includes a stream selection table, additional information,
- the stream selection table indicates the stream number of the subtitle stream that should be permitted to be played back in the planar playback mode in association with the stream entry and the stream attribute, and the additional information is associated with the stream number in the stream selection table.
- the area reservation flag indicates whether the display area of the caption when the playback device is set to the stereoscopic playback mode is the upper end or the lower end of the video plane;
- the area reservation flag indicates that the subtitle display area is at the top of the video plane. In this case, when the video stream is drawn by shifting downward on the video plane by the playback device and the area reservation flag indicates that the subtitle display area is at the lower end of the video plane, the video stream is The drawing is characterized by being shifted in the direction.
- the additional information including the area reservation flag that defines the display area reservation of the caption is described in the stream selection table for each playback section in association with the stream number, when the playback section changes or when a stream change request occurs
- the stream selection procedure is executed and the stream number corresponding to the language setting on the device side is set in the stream number register, the area reservation flag indicated by the additional information corresponding to the new stream number is provided to the playback device. Will be. By doing so, it is possible to realize control in which a caption display area is secured at the upper end in a certain playback section and a caption display area is secured at the lower end in another playback section.
- the aspect ratio of cinesco size (1: 2.35) is common, and when storing images on an optical disc such as a BD-ROM, the main image without changing the aspect ratio. Is generally arranged at the center of a 16: 9 HD video, and black frames are inserted at the top and bottom of the screen. Therefore, according to the above configuration, it is possible to display subtitles in a wide subtitle display area in which the black frames at the top and bottom of the video are collected above or below the video plane, thereby improving the screen usage efficiency. And the stereoscopic effect can be improved.
- An object of the present invention is to provide a recording medium that can prevent adverse effects caused by viewing a 3D video with a screen size different from that assumed at the time of production.
- a recording medium capable of solving the above problems is A recording medium in which a main-view video stream, a sub-view video stream, and metadata are recorded, wherein the main-view video stream includes picture data constituting a main view of stereoscopic video, and the sub-view video stream is , Including the picture data constituting the sub-view of the stereoscopic video, the metadata including the offset correction value corresponding to the screen size information of a plurality of display devices, and the offset correction value being the picture data constituting the main view
- An offset for moving at least one of a main view video plane on which picture data is drawn and a sub-view video plane on which picture data constituting the sub view is drawn is defined in the right and left directions of the horizontal coordinates.
- FIG. 1 shows a home theater system including a recording medium as a package medium, a playback device as a player device, a display device, and glasses.
- Fig. 2 shows the user's face on the left side, and the right side shows an example when the dinosaur skeleton, the target object, is viewed from the left eye, and an example when the dinosaur skeleton, the target object, is viewed from the right eye.
- An example of an internal structure of the base view video stream and dependent view video stream for stereoscopic vision is shown. It is a figure explaining the concept which puts up the black frame which is not utilized for a main video at the top, and displays subtitle data in the black frame.
- 1 shows an internal configuration of a recording medium according to a first embodiment.
- the internal structure of main TS and sub-TS is shown.
- the internal structure of playlist information is shown.
- An example of a basic stream selection table is shown.
- the internal structure of an extension stream selection table is shown.
- the stream registration sequence in the extended stream selection table is shown. Shows the plane composition when video_shift_mode is set to Keep.
- regenerating apparatus is shown.
- the internal structure of a PG decoder is shown.
- the internal structure of a text subtitle decoder is shown.
- the decoder model of IG decoder is shown.
- the circuit configuration for synthesizing the output of the decoder model and outputting in 3D-LR format is shown.
- a circuit configuration for synthesizing outputs of these decoder models and outputting them in the 1 plane + offset mode system is shown.
- a circuit configuration for synthesizing the outputs of the decoder model and outputting in the upper end side 2D subtitle playback mode method and the lower end side 2D subtitle playback mode method is shown.
- FIG. 3 is a diagram depicting an internal configuration of a register set 203 and a playback control unit. Indicates the bit assignment of PSR24. Indicates the bit assignment of PSR32.
- the playlist playback procedure is shown. It is a flowchart which shows the process sequence about Procedure when playback condition is changed about PGtestST stream. It is a flowchart which shows an upper / lower end reproduction
- a circuit configuration for synthesizing the outputs of the decoder model and outputting in the upper end side 2D subtitle playback mode method and the lower end side 2D subtitle playback mode method is shown. It is a figure explaining the system parameter which shows the shift amount of the vertical direction of each plane. It is a figure explaining the method of shifting & cropping PG plane according to video shift mode. It is a figure explaining the constraints for putting subtitle data in the field which is not cut out by plane composition in the upper end side 2D subtitle reproduction mode and the lower end side 2D subtitle reproduction mode. It is a figure explaining the function of the output offset correction value of the influence between the depths by the screen size of a television.
- 2D playlist and 3D playlist are shown.
- the internal structure of a clip information file is shown.
- the relationship between a clip information file, a playlist, and a stereoscopic interleaved stream file is shown. It is a figure which shows the internal structure of clip base information and clip dependent information. It is a figure which shows a basic entry map and an extended entry map. Indicates an entry that is not allowed in the extended entry map.
- the play item reproduction procedure is shown. It shows how an ATC sequence is restored from data blocks constituting a stereoscopic interleaved stream file. It is a figure which shows how restoration of an ATC sequence is performed.
- An example of extent start point information in base view clip information and an example of extent start point information in dependent view clip information are shown.
- FIG. 1 An ATC sequence restoration procedure is shown. It is a figure explaining the reproduction environment of 2D / 3D reproduction
- An optical disk recording method will be described. It is a flowchart which shows the process sequence of an authoring process. The processing procedure of an AV file writing process is shown.
- the internal structure of a multilayered optical disk is shown.
- An optical disk application format based on a file system is shown.
- the structure of the 2D / 3D playback device is shown.
- FIG. 1 It is a figure which shows in detail how a video stream and an audio stream are stored in a PES packet sequence, and shows a TS packet and a source packet structure in an AV clip. It is a figure which shows the data structure of PMT, and the internal structure of a clip information file. It is a figure which shows the internal structure of stream attribute information, and the internal structure of an entry map. It is a figure which shows the internal structure of a play list, and the internal structure of a play item. It is a figure explaining the internal structure of a 2D reproducing
- the upper part is a diagram explaining how to create a video stream by placing the main video slightly above the center, not in the middle, and the lower part explains how to create a black frame by dynamically changing the transparent color of the PG stream
- FIG. It is a figure explaining the structure in the case of prescribing that each extent has one or more entry points. It is a figure explaining the method to store offset metadata in an AV stream information file. It is a figure explaining the method to store offset metadata for every entry point. It is a figure explaining the method to store offset metadata in a play list. It is a figure explaining the method of not storing offset metadata, when offset metadata is stored in a play list and it is the same as the previous play item.
- FIG. 1 It is a figure which shows an example of the index file in BD-ROM in order to reproduce
- a flow for selecting a 2D playlist and a 3D playlist in the program of the BD program file is shown.
- This is a configuration example in which a 2D / 3D playback device is realized using an integrated circuit.
- the recording medium provided with the above problem solving means can be implemented as a package medium for selling contents over the counter. Further, the playback device corresponding to the recording medium can be implemented as a player device for playing back the package medium, and the integrated circuit corresponding to the recording medium is implemented as a system LSI incorporated in the player device. Can do.
- FIG. 1 shows a home theater system including a recording medium as a package medium, a playback device as a player device, a display device, and glasses.
- the recording medium 100 as the package medium and the playback device 200 as the player device constitute a home theater system together with the display device 300, the 3D glasses 400, and the remote controller 500, and are used by the user. Is done.
- the recording medium 100 supplies, for example, a movie work to the home theater system. Some of these movie works constitute a stereoscopic image.
- the stereoscopic video is composed of at least two viewpoint videos.
- the viewpoint video is a video having some deflection, and the two viewpoint videos are composed of a main view video and a subview video.
- FIG. 1 there are various types of recording media 100 such as a disk type and a memory card type.
- the “recording medium” described below is a disk medium. Suppose there is.
- the playback device 200 is connected to the display device 300 and plays back the recording medium 100.
- the playback device has two output modes, 2D output mode and 3D output mode, to play back the main view video stream representing the main view video and the subview video stream representing the subview video. Is a 2D / 3D playback device (player).
- the display device 300 is a television and provides a user with an interactive operating environment by displaying a playback image of a movie work or displaying a menu or the like.
- the display device 300 according to the present embodiment realizes stereoscopic vision by wearing 3D glasses 400 by a user. However, if the display device 300 is a lenticular type, the 3D glasses 400 are not necessary.
- the 3D glasses 400 are provided with a liquid crystal shutter, and allow the user to view parallax images by the sequential separation method or the polarization glasses method.
- a parallax image is a set of videos composed of a video image that enters the right eye and a video image that enters the left eye, and allows stereoscopic viewing so that only pictures corresponding to each eye enter the user's eyes.
- FIG. 4B shows the display time of the left-eye video. At the moment when the image for the left eye is displayed on the screen, the aforementioned 3D glasses 400 transmit the liquid crystal shutter corresponding to the left eye and shield the liquid crystal shutter corresponding to the right eye.
- FIG. 4C shows the time when the right-eye video is displayed. At the moment when the image for the right eye is displayed on the screen, the liquid crystal shutter corresponding to the right eye is made transparent, and the liquid crystal shutter corresponding to the left eye is shielded from light.
- the remote controller 500 is a device for receiving operation items for AV playback.
- the remote controller 500 is a device that accepts an operation on a hierarchical GUI from a user. To accept such an operation, the remote controller 500 moves a menu key for calling a menu constituting the GUI and a focus of the GUI component constituting the menu. An arrow key to be moved, a determination key for performing a confirmation operation on a GUI component constituting the menu, a return key for returning a hierarchical menu to a higher level, and a numerical key.
- the output mode of the playback device that causes the display device 300 to display an image in the 3D output mode is referred to as “3D output mode”.
- the output mode of the playback device that causes the display device 300 to display an image in the 2D output mode is referred to as a “2D output mode”.
- a feature of the first embodiment is that the capability of whether or not stereoscopic viewing using a right-eye graphics stream and a left-eye graphics stream can be realized is set in a register in the playback device.
- the parallax image method (referred to as 3D-LR mode) is a method of performing stereoscopic viewing by preparing a video entering the right eye and a video entering the left eye so that only a picture corresponding to each eye enters.
- Fig. 2 shows the user's face on the left side, and the right side shows an example when the dinosaur skeleton, the target object, is viewed from the left eye, and an example when the dinosaur skeleton, the target object, is viewed from the right eye. ing.
- the left and right scenes are overlapped by the afterimage reaction of the eyes in the user's brain, and it can be recognized that there is a stereoscopic image on the extension line in the center of the face. .
- the MPEG4-MVC method is used as a video stream encoding method for realizing such a stereoscopic view.
- the “base view video stream” in MPEG4-MVC is a main view video stream
- the “dependent view video stream” in MPEG4-MVC is a subview video stream.
- the MPEG-4 MVC base-view video stream is a sub-bit stream with a view_id of 0 and a sequence of view components with a view_id of 0.
- MPEG-4 MVC base-view video stream obeys the limitations of MPEG-4 AVC video stream.
- MPEG-4 MVC dependent view video stream ⁇ ⁇ is a sub-bit stream with view_id of 1 and a sequence of view components with view_id of 1.
- a view component refers to each of a plurality of picture data that is simultaneously played back for stereoscopic playback in one frame period. Compression encoding using the correlation between viewpoints is achieved by using the view components of the base-view video stream and the dependent-view video stream as picture data and realizing compression encoding using the correlation between pictures. Made.
- a combination of the view component of the base view video stream and the view component of the dependent view video stream allocated in one frame period constitutes one access unit, and this access unit is random. Access is possible.
- Both the base view video stream and the dependent view video stream have a GOP structure in which each view component is a “picture”, and are composed of a closed GOP and an open GOP.
- the closed GOP is composed of an IDR picture, a B picture following the IDR picture, and a P picture.
- the open GOP is composed of a Non-IDR I picture, a B picture following the Non-IDR I picture, and a P picture.
- a B picture refers to a picture composed of Bidirectionally predictive (B) format slice data
- a P picture refers to a picture composed of Predictive (P) format slice data.
- the B picture includes a refrenceB (Br) picture and a nonrefrenceB (B) picture.
- the IDR picture is placed at the top of the closed GOP. Although the IDR picture does not become the head in the display order, other pictures (B picture, P picture) other than the IDR picture cannot have a dependency relationship with pictures existing in the GOP before the closed GOP. Thus, the closed GOP has a role of completing the dependency.
- FIG. 3 shows an example of the internal configuration of a base-view video stream and a dependent-view video stream for stereoscopic viewing.
- the second level of the figure shows the internal structure of the base-view video stream.
- This stream includes view components whose picture types are I1, P2, Br3, Br4, P5, Br6, Br7, and P9. These view components are decoded according to Decode Time Stamp (DTS).
- DTS Decode Time Stamp
- the first row shows a left eye image.
- the decoded view components I1, P2, Br3, Br4, P5, Br6, Br7, and P9 are played in the order of I1, Br3, Br4, P2, Br6, Br7, and P5 according to PTS. Will be played.
- the fourth row shows the internal structure of the dependent-view video stream.
- This dependent view video stream includes picture type view components P1, P2, B3, B4, P5, B6, B7, and P8. These view components are decoded according to DTS.
- the third row shows a right eye image. The right-eye image is reproduced by reproducing the decoded view components P1, P2, B3, B4, P5, B6, B7, and P8 in the order of P1, B3, B4, P2, B6, B7, and P5 according to PTS. Will be played.
- the fifth row shows how the state of the 3D glasses 400 is changed. As shown in the fifth row, the right-eye shutter is closed when the left-eye image is viewed, and the left-eye shutter is closed when the right-eye image is viewed.
- the video frame of the base-view video stream and the video frame of the dependent-view stream are alternately output as “B” ⁇ “D” ⁇ “B” ⁇ “D” in a 1/48 second display period.
- the mode is called “BD presentation mode”.
- 3D-Depth that realizes a stereoscopic effect using 2D images and depth information.
- the same video frame is output twice or more while maintaining the output mode in 3D mode.
- the playback type to be performed is called BB presentation mode.
- BB presentation mode only the video frame of the base-view video stream that can be reproduced independently is repeatedly output as "B"-"B"-"B"-"B".
- the above-mentioned BD presentation mode and BB presentation mode are the basic output modes of the playback device.
- the playback device has a 1 plane + offset mode, an upper end side 2D subtitle playback mode, and a lower end side.
- 1plane + Offset mode (also referred to as 3D-Offset mode) is an output mode that realizes stereoscopic viewing by incorporating a shift unit in the subsequent stage of the plane memory and causing the shift unit to function.
- the plane offset unit shifts the pixel coordinates of the line unit in the plane memory leftward or rightward in each of the left view period and the right view period, and the image points of the right visual line and the left visual line are moved forward or The depth feeling is changed by displacing in the depth direction.
- the image point of the line of sight of both eyes will be in front, and in the right view period in the left view period, the right view period If the pixel coordinates are changed to the left, the image point of the line of sight of both eyes will be in front.
- This plane shift is ideal for easily creating stereoscopic images because only one plane memory is required for stereoscopic viewing.
- This plane shift is only suitable for stereoscopic effects such as menus and subtitles because it only produces a stereoscopic image in which a planar image comes to the front or pulls in the back. This is somewhat unsatisfactory for realizing the stereoscopic effect. This is because the indentation and unevenness of the character's face cannot be reproduced.
- the playback device configuration When supporting 1 plane + Offset mode, the playback device configuration is as follows.
- the playback apparatus includes a plane memory, a CLUT unit, and a combining unit, and a shift unit is incorporated between the CLUT unit and the combining unit.
- achieves the coordinate change of a pixel as mentioned above using the offset in the offset sequence integrated in the access unit structure of dependent view video stream. By doing so, the degree of popping out of the pixels in the 1 plane + offset mode is precisely synchronized with the MVC video stream.
- the 1 plane + offset modes there is a 1 plane + zero offset mode.
- the 1 plane + zero-offset mode is a display mode in which, when the pop-up menu is on, the offset value is set to zero and a stereoscopic effect is given only to the pop-up menu.
- the plane memory subject to shift control by the offset sequence is a plurality of plane memories constituting a predetermined layer model.
- the plane memory stores pixel data for one screen obtained by decoding the elementary stream in units of lines, and outputs these pixel data along the horizontal synchronization signal and the vertical synchronization signal. It is memory.
- Each plane memory stores pixel data for one screen obtained by decoding of a video decoder, a PG decoder, and an IG decoder.
- the predetermined layer model is composed of a left-eye video plane layer, a right-eye video plane layer, a PG plane layer, and an IG / BD-J plane layer. This is intended for layer composition in the order of the / BD-J plane.
- the process of superimposing the pixel values of the pixel data stored in the plane memory of the two layers is executed for all combinations of the two layers in the layer model. That is done.
- the plain memory of each layer will be described.
- the left-eye video plane is a plane memory that can store the pixel data for one screen obtained by decoding the view component and that constitutes picture data for the left eye.
- the right-eye video plane is a plane memory that can store the pixel data for one screen obtained by decoding the view component, that constitutes the picture data for the right eye.
- the presentation graphics (PG) plane is a plane memory used for storing graphics obtained by a decoding process performed by a graphics decoder operating in a pipeline manner.
- the IG / BD-J plane is a plane memory that functions as an IG plane in one operation mode and functions as a BD-J plane in another operation mode.
- An interactive graphics (IG) plane is a plane memory used for storing graphics obtained by decoding by a graphics decoder that operates on the premise of interactive processing.
- the BD-J plane is a plane memory that can be used to store drawing image graphics obtained by an object-oriented programming language application performing drawing processing.
- the IG plane and the BD-J plane are exclusive, and when one of them is used, the other is not used. Therefore, the IG plane and the BD-J plane share one plane memory. .
- the IG / BD-J plane and the PG plane are for the base view and the dependent view. There is no plane memory for each of the dent views. Therefore, the IG / BD-J plane and PG plane are subject to shift control.
- the upper-end 2D subtitle playback mode is an output mode that secures a 2D subtitle display area at the upper end of the video frame by incorporating a shift section after the video plane memory and causing the shift section to function.
- the subtitle playback mode is an output mode that secures a 2D subtitle display area at the lower end of the video stream by causing the shift unit to function.
- the plane offset unit shifts the pixel coordinates of the picture data stored in the video plane memory downward in each of the left-view period and the right-view period, so that the lower-end 2D subtitle playback mode Then, in each of the left view period and the right view period, the coordinates of the pixels of the picture data stored in the video plane memory are shifted upward.
- the configuration of the playback device is as follows.
- the playback apparatus includes a plane memory and a synthesis unit, and a shift unit is incorporated between the video plane memory and the synthesis unit.
- a shift part implement achieves the coordinate change of the above pixels using the offset built in the stream registration sequence of a graphics stream.
- a video frame suitable for use in the upper-end 2D subtitle playback mode and the lower-end 2D subtitle playback mode has an aspect ratio of 16: 9 in the center of the screen with an aspect ratio of 16: 9 and a resolution of 1920 ⁇ 1080 pixels as shown in FIG.
- a 2.35: 1 Sinesco-sized image having a resolution of 1920 ⁇ 818 pixels is arranged, and a 1920 ⁇ 131 pixel black frame is arranged above and below the Sinesco-sized image.
- FIG. As shown in (c), the black frames existing at the upper and lower ends of the original video frame are combined at the upper end or the lower end. By doing so, it is possible to prepare a black frame large enough to display subtitles.
- FIG. 5 shows an internal configuration of the recording medium according to the first embodiment.
- the recording medium according to the first embodiment includes “index table”, “program file of operation mode object”, “playlist information file”, “stream information file”, “stream”. File "is recorded.
- the index table is management information relating to the entire recording medium, and the recording medium is uniquely recognized by the reproducing apparatus by first reading the index table after the recording medium is inserted into the reproducing apparatus.
- the program file of the operation mode object stores a control program for operating the playback device. Some programs are described by commands, and some programs are described by an object-oriented compiler language.
- the former program supplies a plurality of navigation commands as a batch job to the playback device, and operates the playback device based on these navigation commands. This command-based operation mode is called “HDMV mode”.
- the latter program supplies a bytecode application, which is an instance of a class structure, to the playback device in the object-oriented compiler language-based operation mode, and operates the playback device based on this instance.
- a Java (TM) application which is one of bytecode applications, can be used as an instance of the class structure.
- the operation mode based on the object-oriented compiler language is called “BD-J mode”.
- the stream file stores a transport stream obtained by multiplexing a video stream, one or more audio streams, and a graphics stream.
- the 2D dedicated stream file has a normal transport stream format
- the 2D-3D combined stream file has a stereoscopic interleaved stream file format.
- the stereoscopic interleaved stream file format is an interleaved format of the extent of the main transport stream (main TS) including the base view stream and the extent of the sub transport stream (sub TS) including the dependent view video stream. Are arranged alternately.
- the main TS stored in the stream file includes packet management information (PCR, PMT, PAT) defined in the European digital broadcasting standard as information for managing and controlling a plurality of types of PES streams.
- packet management information PCR, PMT, PAT
- PCR Program_Clock_Reference
- ATS Arriv Time Clock
- STC System Time Clock
- PMT Program_map_table
- PID of each stream such as video / audio / graphics included in the stream file
- stream attribute information corresponding to each PID and various descriptors related to TS.
- the descriptor includes copy control information for instructing whether or not to permit copying of the stream file.
- PAT Program Association Table
- PCR, PMT, and PAT have a role to define a partial TS that constitutes one broadcast program (Program) in the European digital broadcasting standard, and the playback device constitutes one broadcast program in the European digital broadcasting standard.
- the TS can be used for processing by the decoder as if it were a partial TS. This is intended to be compatible with a European digital broadcasting standard terminal device and a recording medium playback device.
- the extent set in the main TS and the extent in the sub-TS are set to a data size that does not cause the double buffer to underflow during playback, and the playback device can read these extent sets seamlessly.
- the stream information file is a stream information file that guarantees random access to an arbitrary source packet in the transport stream in the stream file and continuous reproduction with other transport streams. Through this stream information file, the stream file is managed as an “AV clip”.
- the stream information file is a basic entry map that indicates information such as the encoding format, frame rate, bit rate, resolution, etc. of the stream in an AV clip, and the source packet number at the head position of the GOP in association with the presentation time stamp of the frame period. Therefore, if this stream information file is loaded into the memory before accessing the stream file, it can be understood what the transport stream in the stream file to be accessed is. , Random access execution can be guaranteed.
- the stream information file includes a 2D stream information file and a 3D stream information file.
- the 3D stream information file includes clip information for base view (clip base information) and clip information for dependent view ( Clip dependent information) and an extended entry map for stereoscopic viewing.
- the clip base information includes extent start point information for the base view, and the clip dependent information includes extent start point information for the dependent view.
- the extent start point information for the base view is composed of a plurality of source packet numbers. Each source packet number indicates in what packet the extent division position in the main TS exists.
- the extent start point information for the dependent view is also composed of a plurality of source packet numbers, and indicates how many packets have the division position in the sub-TS.
- the “ATC sequence” is an array of source packets, which means that there is no discontinuity (no arrival time-base discontinutiy) in the Arrival_Time_Clock referenced by the Arrival_Time_Stamp. Since the ATC sequence is a source packet sequence in which ATC time stamps are continuous, each source packet constituting the ATC sequence is continuous while the clock counter that measures the arrival time clock of the playback device is measuring time. Source packet depacketizing processing and continuous packet filtering processing.
- the ATC sequence is an array of source packets, an array of TS packets with consecutive time stamps on the STC time axis is called an STC sequence.
- the “STC sequence” is an array of TS packets, and does not have an STC (System Time Clock) discontinuity point (system time-base discontinuity) that is a TS system base time.
- the discontinuity point of STC is a point where discontinuity information (discontinuity_indicator) of a PCR packet carrying PCR (Program Clock Reference) that is referred to by the decoder to obtain STC is ON.
- the STC sequence is a TS packet sequence in which STC time stamps are continuous, each TS packet constituting the STC sequence is played back while the clock counter that counts the system time clock of the playback device is counting. It will be used for the continuous decoding process of the decoder existing in the apparatus.
- the extended entry map indicates the source packet number of the access unit delimiter that is the head position of the GOP head view component in the dependent-view video stream in association with the presentation time stamp representing the GOP head frame period.
- the basic entry map in the 3D stream information file is compatible with the 2D stream information file, but is associated with the presentation time stamp indicating the GOP head frame period, and the GOP head view component of the base view video stream. Indicates the source packet number of the access unit delimiter at the head position.
- the playlist information file is a file that stores information for causing the playback device to play back the playlist.
- a “playlist” is a playback path that is defined by specifying playback sections on the time axis of the transport stream (TS) and logically specifying the playback order between the playback sections. Of these, it plays the role of defining which part is played back and in what order the scene is developed.
- the playlist information defines the “type” of the playlist.
- the playback path defined by the playlist information is a so-called “multipath”. Multipath is a bundle of a playback path (main path) defined for a main TS and a playback path (subpath) defined for a subordinate stream. If the playback path of the base-view video stream is defined in this multi-pass and the playback path of the dependent-view video stream is defined in the sub-path, the combination of video streams for reproducing stereoscopic vision can be preferably defined. it can.
- An application based on an object-oriented programming language orders the generation of a framework player instance that reproduces this playlist information, so that AV reproduction by multipass can be started.
- the framework player instance is actual data generated on the heap memory of the virtual machine based on the media framework player class.
- the command-based program can also start playback by multipass by issuing a playback command specifying this playlist information as an argument.
- Playlist information includes one or more play item information.
- the play item information is playback section information that defines one or more sets of a time point that becomes In_Time and a time point that becomes Out_Time on the playback time axis of the video stream.
- the playlist information has a hierarchical structure of play item information-stream information-transport stream, and the ratio of the combination of the transport stream and stream information and the play item information has a one-to-many relationship.
- One transport stream can be referred to from a plurality of pieces of play item information. Therefore, it is possible to efficiently create a variation of a movie work by adopting a transport stream created for a playlist as a bank film and referring to it from play item information in a plurality of playlist information files.
- Yes bank film is a term in the movie industry and refers to video content that is reused in multiple scenes).
- a unit of a playlist is not recognized by the user, a plurality of variations (for example, a theater release version or a TV broadcast version) derived from a stream file are recognized as a playlist.
- the playlist information includes 2D playlist information and 3D playlist information, and the difference between them is that a base view indicator and a stereoscopic stream number table exist in the 3D playlist information.
- the “stereoscopic stream number table” is a table showing stream entries and stream attributes of elementary streams to be reproduced only in the 3D output mode in association with stream numbers.
- Base view indicator is information indicating whether a base-view video stream serving as a basis for compression coding using correlation between viewpoints should be displayed as a left eye or a right eye.
- the stereoscopic effect can be reversed.
- a playlist intended for a stereoscopic effect in which an object appears in front of the screen has already been created
- a stereoscopic effect in which the object appears in the back of the screen is separate from the playlist.
- An intended playlist can be created. Therefore, in the 3D playlist, there is an effect that it is possible to easily create variations of the playlist with different stereoscopic effects.
- FIG. 5B shows the internal configuration of the main TS
- FIG. 5C shows the internal configuration of the sub-TS.
- the main TS includes one base-view video stream, 32 left-eye PG streams, 32 left-eye IG streams, and 32 audio streams.
- the sub-TS includes one dependent-view video stream, 32 right-eye PG streams, and 32 right-eye IG streams.
- Elementary streams (ES) multiplexed in TS include video streams, audio streams, presentation graphics streams, and interactive graphics streams.
- the base view video stream constitutes the primary video stream in the picture-in-picture application.
- the picture-in-picture application includes a secondary video stream in addition to the primary video stream.
- the primary video stream is a video stream composed of picture data that becomes a parent screen in a picture-in-picture application.
- the secondary video stream is a video stream composed of picture data inserted into a part of the parent screen as a child screen in picture-in-picture.
- the picture data constituting the primary video stream and the picture data constituting the secondary video stream are decoded and stored in separate plane memories. There are components (Scalling & Positioning) for changing the scaling of the picture data constituting the secondary video stream and positioning the display coordinates in the previous stage of the plane memory for storing the picture data constituting the secondary video stream.
- Audio stream There are two types of audio streams, a primary audio stream and a secondary audio stream.
- the primary audio stream is an audio stream that should be a main sound when mixing reproduction is performed
- the secondary audio stream is an audio stream that should be a sub sound when mixing reproduction is performed.
- the secondary audio stream includes information for downsampling for mixing and information for gain control.
- the PG stream is a graphics stream suitable for subtitle display that can be precisely synchronized with video by adopting a pipeline in the decoder.
- Up to 32 2DPG streams, up to 32 left-eye PG streams, and up to 32 right-eye PG streams can be defined. Each of these PG streams is provided with a different packet identifier. By instructing the demultiplexer to specify a packet identifier to be reproduced, a desired one of these PG streams is provided for reproduction. Will be.
- PG stream is not limited to character reproduction like subtitles. Any graphics playback that requires precise synchronization, such as displaying a mascot character of a movie work and synchronizing it with a moving image, should be adopted as a playback target by the PG stream. Can do.
- -Streams that are not multiplexed into a stream file but show subtitles include text subtitle (textST) streams in addition to PG streams.
- the textST stream is a stream that represents the content of subtitles in character code.
- the PG stream and the text subtitle stream are registered in the same stream registration column on the assumption that these types are not distinguished and are the same stream type. Then, when executing the stream selection procedure, the PG stream or text subtitle stream to be reproduced is determined according to the stream registration order in the stream registration sequence. Since the PG stream and the text subtitle stream are provided to the stream selection procedure without distinguishing the stream type, the PG stream and the text subtitle stream are classified into one stream type, that is, “PG_text subtitle stream (abbreviated, It may be called a subtitle stream).
- the PG_text subtitle stream for 2D is played back in 1 plane + Offset mode, upper end 2D subtitle playback mode, and lower end 2D subtitle playback mode.
- Interactive graphics (IG) stream IG stream is a graphics stream that can display a menu as the video stream is played and can display a pop-up menu according to the user's operation by providing information on interactive operations. is there.
- the IG stream has two types, 2DIG stream and stereoscopic IG stream, like PG stream.
- the control information of the IG stream (referred to as the dialog control segment) has information (User_interface_model) that defines the user interface model, and the author can set the user interface model information to set the menu as the video stream playback progresses. Can be specified (always on), or a pop-up menu can be displayed according to user operation (pop-up menu on).
- the significance of the IG stream having dialogue operation information is as follows.
- the Java virtual machine instructs the playback control engine, which is the main player of playback control, to start playback of a playlist in response to a request from the application
- the Java virtual machine instructs the playback control engine to play, and then plays the playlist. Returns a response indicating that it has started to the application.
- the Java virtual machine does not wait for execution completion. This is because the Java virtual machine is a so-called event-driven operating entity and can operate while the playback control engine is playing back a playlist.
- the command interpreter when the command interpreter commands the playback control engine to play a playlist, it waits for completion of the playlist playback until the playlist playback ends. While playback by the playback control engine continues, the command execution unit cannot execute interactive processing. Instead of this command interpreter, the graphics decoder performs an interactive operation. In order to cause the graphics decoder to perform an interactive operation, control information defining an interactive operation using a button member is incorporated in the IG stream. *
- 3D display mode which 3D display mode is allowed depends on the stream type.
- two output modes such as a BD presentation mode and a BB presentation mode are allowed.
- the BB presentation mode is allowed only when the pop-up menu is turned on.
- the type of primary video stream in the case of playback in the BD presentation mode is called “stereoscopic BD playback type”.
- the type of primary video stream in the case of playback in the BB presentation mode is called stereoscopic BB playback type.
- the PG stream 3D display mode allows five output modes: BD presentation mode, 1 plane + offset mode, 1 plane + zero offset mode, upper end 2D subtitle playback mode, and lower end 2D subtitle playback mode.
- the 1 plane + zero offset mode is allowed only when the pop-up menu is on.
- the type of PG stream in the case of playback in the BD presentation mode is called “stereoscopic playback type”.
- the type of PG stream and PG_text subtitle stream when played back in the 1 plane + Offset mode is referred to as a 1 plane + Offset type.
- the types of PG stream and PG_text subtitle stream when played back in the 1 plane + Zero Offset mode are referred to as 1plane + Zero Offset type.
- the types of PG stream and PG_text subtitle stream when played back in the upper end 2D subtitle playback mode are referred to as “upper end 2D subtitle playback type”.
- the types of PG stream and PG_text subtitle stream when played back in the lower-end 2D subtitle playback mode are referred to as “lower-end 2D subtitle playback type”.
- 3D display mode of IG stream allows 3 output modes such as BD presentation mode, 1 plane + offset mode, 1 plane + zero offset mode.
- the 1 plane + zero offset mode is allowed only when the pop-up menu is turned on.
- picture-in-picture cannot be used when the 3D output mode is executed unless otherwise specified. This is because both the picture-in-picture and 3D output modes require two video planes for storing uncompressed picture data. Unless otherwise specified, sound mixing cannot be used in 3D output mode.
- FIG. 6 shows the internal structure of the main TS and sub-TS.
- Figure (a) shows the internal structure of the main TS.
- the main TS is composed of the following source packets.
- a source packet having a packet ID of 0x0100 constitutes Program_Map_Table
- a TS packet having a packet ID of 0x1001 constitutes a PCR.
- a source packet sequence having a packet ID of 0x1011 constitutes a primary video stream.
- Thirty-two 2DIG streams are configured from the source packet sequence having packet IDs of 0x1400 to 0x141F.
- a primary audio stream is configured from a source packet sequence having a packet identifier of 0x1100 to a source packet sequence having a packet identifier of 0x111F.
- Figure (b) shows the internal structure of the sub-TS.
- the sub-TS is composed of the following source packets.
- the source packet sequence having a packet identifier of Ox1012 constitutes a dependent view video stream.
- Thirty-two left-eye IG streams are configured from a source packet sequence having a packet identifier of 0x1420 to a source packet sequence having a packet ID of 0x143F.
- a source packet sequence having a packet identifier of 0x145F from a source packet sequence having a packet identifier of Ox1440 constitutes 32 right-eye IG streams.
- FIG. 7 shows the internal structure of the playlist information.
- the playlist information includes “main path information”, “sub path information”, “play list mark information”, and “extension data”.
- main path information “main path information”
- sub path information “sub path information”
- play list mark information “extension data”.
- Main path information is composed of one or more main playback section information.
- FIG. 7B is a diagram showing an internal configuration of main path information and sub path information. As shown in this figure, the main path information is composed of one or more main playback section information.
- the sub-path information is composed of one or more subordinate playback section information.
- the main playback section information is called play item information, and is information that defines a logical playback section by defining one or more sets of time points that become In_Time and time points that become Out_Time on the TS playback time axis. It is.
- the playback device includes a play item number register that stores the number of the current play item. Among a plurality of pieces of playlist information, the information stored in the play item number register is the current playback target.
- FIG. 7C shows the internal structure of the play item information. As shown in the figure, it includes “stream reference information”, “in timeout time information”, “connection state information”, and “basic stream selection table”.
- the stream reference information includes “stream information file name information (clip_information_file_name)” indicating a stream information file that manages a transport stream constituting a play item as an “AV clip”, and “clip code indicating an encoding method in the TS. And “STC identifier reference (STC_ID_referrence)” indicating which STC sequence has in-time and out-time set in the STC sequence of the TS.
- Subordinate playback section information is called sub-path information and is composed of a plurality of sub-playitem information.
- FIG.7 (d) shows the internal structure of a sub play item.
- the sub play item information is information that defines a playback section constituting a sub path by defining a combination of in time and out time on the time axis of the STC sequence.
- Information “ in timeout time information ”,“ synchro play item reference ”, and“ sync start time information ”.
- Stream reference information includes “stream information file name information”, “clip encoding scheme identifier”, and “STC identifier reference”, as with play item information.
- “In timeout time information (SubPlayItem_In_Time, SubPlayItem_Out_Time)” indicates the start point of the sub play item on the STC sequence time axis and the end point of the sub play item on the STC sequence time axis.
- Synchronized play item reference (Sync_PlayItem_Id) is information for uniquely specifying a play item to be synchronized with the sub play item.
- the sub play item in time exists on the playback time axis of the play item specified by the synchronous play item reference.
- Playlist mark information is information that defines mark points specific to the playback section, and indicates where the mark point is located on the reference line indicating the playback section and the time axis of the digital stream. Including a time stamp to indicate and attribute information indicating an attribute of the mark point, The attribute information indicates whether the mark point defined by the playlist mark information is a link point or an entry mark.
- the link point is a mark point that can be linked by a link command but is not a selection target when the chapter skip operation is performed by the user.
- the entry mark is a mark point that can be linked by a link command and is a selection target when a chapter skip operation is performed by the user.
- the link command embedded in the button information of the IG stream specifies the cue position in the form of indirect reference via playlist mark information.
- the basic stream selection table is referred to by a multi-pass main path when a current play item becomes a current play item among a plurality of play items constituting a playlist.
- the basic stream selection table is referred to by a multi-pass main path when a current play item becomes a current play item among a plurality of play items constituting a playlist.
- the stream type here refers to a type such as a primary video stream in picture-in-picture, a secondary video stream in picture-in-picture, a primary audio stream in sound mixing, a secondary audio stream in sound mixing, a PG_text subtitle stream, and an IG stream
- the basic stream selection table can register a stream that should be permitted to be reproduced for each of these stream types.
- the basic stream selection table is composed of a stream registration array.
- stream registration refers to the stream number, which ES is to be permitted to be played when the play item to which the basic stream selection table belongs becomes the current play item.
- the stream registration has a data structure in which a combination of a stream entry and a stream attribute is associated with a logical stream number.
- the stream number in the stream registration is represented by an integer value such as 1, 2, 3, and the maximum number of stream numbers is the number of streams of the corresponding stream type.
- the playback apparatus has a stream number register for each stream type, and the ES indicated by the stream number stored here becomes the ES that is the current playback target, that is, the current stream.
- the ES packet identifier to be reproduced is described. Since the ES packet identifier to be reproduced can be described in the stream entry, the stream number in the stream registration is stored in the stream number register of the reproduction apparatus, and the reproduction apparatus is based on the packet identifier in the stream entry in the stream registration.
- the PID filter causes the playback device to perform packet filtering. By doing so, the ES TS packet whose reproduction is permitted in the basic stream selection table is output to the decoder, and the ES is reproduced.
- stream registrations in the basic stream selection table are arranged in accordance with the order of stream numbers, and the order of stream registration based on the order of stream numbers is “can be played back by playback device” and “language attribute of stream”.
- the stream to be selected is determined according to the order of the stream numbers in the stream registration sequence.
- the selection procedure for selecting which of the streams satisfies the condition that “the language attribute of“ matches the language setting of the playback device ” is called a stream selection procedure.
- the stream selection procedure is executed when the current play item is switched to a new one, or when stream switching is requested by the user.
- the stream registration column in the basic stream selection table uniformly assigns a priority order to the stream specified by the sub play item information and the stream specified by the play item information, it is multiplexed with the video stream. Even if the stream is not specified, it is a target for selection when selecting a stream to be reproduced in synchronization with the video stream if it is specified by the sub play item information.
- the playback device can play back the stream specified by the sub play item information, and the priority order of the stream specified by the sub play item information is the priority of the graphics stream multiplexed with the video stream.
- the stream specified by the sub play item information can be used for reproduction instead of the stream multiplexed with the video stream.
- the stream number described in the basic stream selection table can be used as an operand of the set stream command.
- a set stream command is a command that sets the stream number specified in the operand as the current stream number in the stream number register and instructs the playback device to switch the current stream.
- the command-based program plays back the stream switching. Used when the device is executed.
- the stream change UO is a user operation event that sets the stream number specified in the argument as the current stream number in the stream number register and instructs the playback apparatus to switch the current stream.
- the set stream API is an API that sets the stream number specified in the argument as the current stream number in the stream number register and instructs the playback device to switch the current stream. This is used when the playback apparatus executes switching.
- FIG. 8 shows an example of the basic stream selection table.
- FIG. 6A shows a plurality of streams provided in the basic stream selection table when there are types such as primary video stream, secondary video stream, PG stream, IG stream, secondary video stream, and secondary audio stream. Indicates the registration column.
- FIG. 5B shows what ES is separated from the main TS and sub-TS by the basic stream selection table. The left side of the figure shows the main TS and sub-TS, and the middle shows the basic stream selection table and the demultiplexing unit. The right side shows a primary video stream, a primary audio stream, a PG stream, an IG stream, a secondary video stream, and a secondary audio stream that are separated based on the basic stream selection table.
- the extended stream selection table must be stored in the data block in the extension data of the playlist information file.
- the extension information (subpath block extension) of the subpath information must be stored in the data block of the extension data in the playlist information file.
- the extended stream selection table is a stream selection table that is used together with the stream selection table only in the stereoscopic output mode, and can be selected when a play item is reproduced or a sub-path associated therewith is being reproduced.
- the extended stream selection table indicates an ES that should be allowed to be reproduced only in the stereoscopic output mode, and includes a stream registration column.
- Each stream registration information in the stream registration sequence includes a stream number, a stream entry corresponding to the stream number, and a stream attribute. Since the extended stream selection table means an extension specific to the stereoscopic output mode, a playlist in which the extended stream selection table (STN_table_SS) is associated with each play item information is referred to as a “3D playlist”.
- the stream entry in the extended stream selection table is a packet that the playback device should use for demultiplexing when the corresponding stream number is set in the stream number register in the playback device when the playback device is set to the stereoscopic output mode. Indicates an identifier.
- the difference from the basic stream selection table is that the stream registration sequence in the extended stream selection table is not subject to the stream selection procedure. That is, the stream registration information in the stream registration sequence in the basic stream selection table is interpreted as the priority order of each ES, and the stream number in any of the stream registration information is written into the stream number register.
- the stream registration sequence in the extension stream selection table is not subject to the stream selection procedure, and the stream registration information in the extension stream selection table is stored in the stream number when any stream number is stored in the stream number register. Used only for the purpose of retrieving the corresponding stream entry and stream attributes.
- the target stream selection table is switched from the basic stream selection table to the extended stream selection table.
- the identity of the numbers cannot be maintained, and the identity of the language attributes may be lost.
- the extended stream selection table is also used for the above.
- the stream number described in the extended stream selection table can be used as an operand of a set stream command or a set stereoscopic stream command.
- the set stereoscopic stream command is a command that sets a stream number for stereoscopic viewing specified in an operand as a current stream number in the stream number register and instructs the playback device to switch the current stream. Is used when causing the playback device to switch to the stereoscopic stream.
- the extended stream selection table is composed of a stream registration column for a dependent-view video stream, a stream registration column for a PG stream, and a stream registration column for an IG stream.
- the stream registration sequence in the extended stream selection table is combined with the stream registration sequence of the same stream type in the stream selection table.
- This combination combines the stream registration column of the dependent-view video stream in the extension stream selection table with the stream registration column of the primary video stream in the stream selection table, and the extension stream into the stream registration column of the PG stream in the stream selection table. This is done by combining the stream registration sequence of the PG stream in the selection table and combining the stream registration sequence of the IG stream in the extension stream selection table with the stream registration sequence of the IG stream.
- the above procedure is executed for the stream registration column in the basic stream selection table among the combined stream selection tables.
- FIG. 9 shows the internal structure of the extension stream selection table.
- the extended stream selection table includes an overall length (length) of the extended stream selection table, a pop-up period fixed offset (Fixed_offset_during_Popup), and a stream registration sequence corresponding to each stream type in each play item.
- a stream registration sequence corresponding to each of the play items # 1 to #N is provided in the extended stream selection table.
- the stream registration sequence corresponding to each play item is a dependent view stream registration sequence, a PG stream registration sequence, and an IG stream registration sequence.
- “Fixed_offset_during_Popup” is a pop-up period fixed offset, and controls the playback type of video or PG_text subtitle stream when the pop-up menu by IG stream is set to ON.
- the “Fixed_offset_during_Popup” field is set to ON when the user_interface_model field in the IG stream is ON, that is, the user interface of the pop-up menu is set to ON. If the user_interface_model field in the IG stream is off, that is, if it is set to an AlwaysON user interface, it is set to off.
- the stereoscopic PG stream is a stereoscopic reproduction type.
- the PG_text subtitle stream is in the 1 plane + Offset mode.
- the video stream is in the BB presentation mode.
- the stereoscopic PG stream is in 1 plane + Offset mode, and the PG stream for 1 plane + Offset is reproduced as a 1 plane + zero offset reproduction type.
- PG_text subtitle stream is 1plane + zero offset in 1plane + offset mode.
- FIG. 10 shows a stream registration sequence in the extended stream selection table.
- FIG. 10A shows the internal structure of the stream registration sequence of the dependent-view video stream.
- the stream registration sequence of the dependent view stream is composed of v (x) SS_dependet_view_blocks.
- v (x) is the number of primary video streams permitted to be reproduced in the basic stream selection table of play item information #x.
- the lead line in the figure shows the internal structure of the dependent view stream stream registration sequence in close-up.
- SS_dependet_view_block is composed of a stream number, a stream entry, a stream attribute, and the number of offset sequences (number_of_offset_sequence).
- the stream entry includes a sub-path identifier reference (ref_to_Subpath_id) that specifies a sub-path to which a dependent-view video stream playback path belongs, and a stream file reference that specifies a stream file in which the dependent-view video stream is stored ( ref_to_subClip_entry_id) and a packet identifier (ref_to_stream_PID_subclip) of the dependent-view video stream in the stream file.
- a sub-path identifier reference ref_to_Subpath_id
- ref_to_subClip_entry_id a stream file in which the dependent-view video stream is stored
- ref_to_stream_PID_subclip packet identifier of the dependent-view video stream in the stream file.
- Stream attribute includes the language attribute of the dependent-view video stream.
- Number of offset sequences (number_of_offset_sequence)” indicates the number of offsets present in the dependent-view video stream.
- Offset sequence number information (number_of_offset_sequence in the figure) indicates the number of offset sequences in the dependent view stream.
- extension stream selection table This value in the extension stream selection table is the same as the number of offset sequences included in the dependent view stream.
- the stream registration column of the dependent-view video stream is provided with stream registration information for a plurality of dependent-view video streams in terms of the data structure. Normally, since the number of base-view video streams is one, the number of pieces of stream registration information in the dependent-view video stream is also unique.
- FIG. 10B shows the internal configuration of the stream registration sequence of the PG stream.
- the stream registration sequence of the PG stream is composed of P (x) pieces of stream registration information.
- P (x) is the number of PG streams permitted to be reproduced in the basic stream selection table of play item information #x.
- the leader line in the figure shows a close-up of the common internal structure of the stream registration sequence.
- PG_text subtitle offset sequence ID reference information (PGtextST_offset_sequence_id_ref) is PG_text subtitle stream offset sequence reference information, and indicates an offset sequence for a PG_text subtitle stream in 1 plane + Offset mode.
- the offset metadata is supplied by the dependent view video stream access unit.
- the playback device must apply the offset provided by this field to the 1 plane + Offset mode type presentation graphics (PG) plane.
- PG Offset mode type presentation graphics
- this field is an undefined value (FF)
- the playback device does not apply this offset to the PG stream plane memory.
- Steposcopic PG presence / absence flag indicates the validity and existence of the left-eye IG stream entry, right-eye IG stream entry, and stream attribute in the PG stream. If there is no structure in the stereoscopic PG stream, this field must be set to 0. If there is a structure in the stereoscopic PG stream, this field must be set to 1.
- the “left-eye stream entry” includes a sub-path identifier reference (ref_to_Subpath_id) that specifies the sub-path to which the playback path of the left-eye PG stream belongs, and a stream file reference (ref_to_subClip_entry_id) that specifies the stream file in which the left-eye PG stream is stored. And the packet identifier (ref_to_stream_PID_subclip) of the left-eye PG stream in the stream file.
- the “right-eye stream entry” includes a sub-path identifier reference (ref_to_Subpath_id) that specifies the sub-path to which the playback path of the right-eye PG stream belongs, and a stream file reference (ref_to_subClip_entry_id) that specifies the stream file in which the right-eye PG stream is stored. And the packet identifier (ref_to_stream_PID_subclip) of the right-eye PG stream in the stream file.
- stream_entry_for_depentdent_view in the stream registration information of the extension stream selection table is different from the stream file referenced by the stream entry of the basic stream selection table, the stream file storing the right-eye PG stream is revisited. Must be read.
- Communication stream attribute includes the language attribute of the left-eye PG stream and the dependent PG stream.
- the “stereoscopic PG_text subtitle offset sequence reference information (SS_PG_textST_offset_sequence_id_ref)” is reference information for referring to the offset sequence for the PG_text subtitle stream, and indicates an offset sequence for the PG_text subtitle stream.
- the playback device must apply the offset provided by this field to the PG plane.
- this field is an undefined value (FF)
- the playback device does not apply this offset to the PG stream plane memory.
- Video shift mode (video_shift_mode) is an area securing flag that defines the process for securing the caption display area. Whether the caption display area is secured at the upper end or the lower end of the video plane, and the caption display area is the video. Indicate whether the upper or lower end of the plane. When the subtitle display area is not secured at the upper or lower end of the video plane, the video shift mode is set to “Keep”. When video_shift_mode is set to Keep, as shown in FIG. 11, the picture data stored in the video plane memory is not shifted in the vertical direction, but is directly combined with the subtitles stored in the PG stream plane memory. .
- the video shift mode is set to "Up"
- the video shift mode is " Set to "Down”.
- FIG. 10C shows the internal configuration of the stream registration sequence of the IG stream.
- the stream registration sequence of the IG stream includes I (x) pieces of stream registration information.
- I (x) is the number of IG streams permitted to be reproduced in the basic stream selection table of play item information #x.
- the lead lines in the figure show a close-up of the common internal configuration of the stream registration sequence.
- IG offset sequence ID reference information (IG_offset_sequence_id_ref) is an interactive graphics offset sequence reference, which is a reference of the sequence ID of the IG stream in the 1 plane + Offset mode. This value indicates the offset sequence ID defined in the offset sequence.
- offset metadata is provided by the dependent-view video stream. The playback device must apply the offset provided by this field to the 1 plane + Offset mode type IG stream.
- the playback device does not apply this offset to the interactive graphics plane.
- Offset direction information in BB mode indicates the offset direction in the interactive graphics (IG) plane in the 1 plane + offset mode while the IG stream is played on the user interface of the pop-up menu in the BB presentation mode.
- the front setting that is, the plane memory exists between the TV and the viewer, and the plane is shifted to the right in the left view period, and the plane is shifted to the left in the right view period.
- Offset value information in BB mode indicates the offset value of the IG plane in the 1 plane + Offset mode in units of pixels while the IG stream is played back by the user interface of the pop-up menu in the BB presentation mode.
- Steposcopic IG presence / absence flag indicates the validity and existence of the left-eye IG stream entry, the right-eye IG stream entry, and the stream attribute in the IG stream. If the data structure of the stereoscopic IG stream does not exist, this field must be set to the value 0. If the IG stream that is permitted to be played is a stereoscopic IG stream, this field must be set to a value of 1.
- the “left-eye stream entry” includes a sub-path identifier reference (ref_to_Subpath_id) that specifies the sub-path to which the playback path of the left-eye IG stream belongs, and a stream file reference (ref_to_subClip_entry_id) that specifies the stream file in which the left-eye IG stream is stored. ) And the packet identifier (ref_to_stream_PID_subclip) of the left-eye IG stream in the stream file.
- Light-eye stream entry includes a sub-path identifier reference (ref_to_Subpath_id) that specifies the sub-path to which the playback path of the right-eye IG stream belongs, and a stream file reference (ref_to_subClip_entry_id) that specifies the stream file in which the right-eye IG stream is stored. ) And the packet identifier (ref_to_stream_PID_subclip) of the right-eye IG stream in the stream file.
- the stream file referenced by stream_entry_for_depentdent_view in the stream registration information of the extended stream selection table is different from the stream file referenced by the stream entry of the basic stream selection table, the stream file storing the right-eye IG stream is It must be read again.
- Communication stream attribute includes language attribute of left-eye IG stream and right-eye IG stream.
- Steposcopic IG offset sequence reference information is a reference of an offset sequence ID for a stereoscopic type IG stream, and indicates an offset sequence in offset metadata of a dependent-view video stream.
- the playback device must apply the offset provided by this field to the stereoscopic type IG plane.
- the reference information of the offset sequence for the PG_text subtitle stream and the reference information of the offset sequence for the IG stream are described in the stream registration information in association with the stream number.
- Stream entries in the stereoscopic dependent view block must not change in the playlist.
- the sub path ID reference and the sub clip entry ID reference do not change in the playlist. .
- the stream attribute stream encoding method in the stereoscopic dependent view block is set to “0x20”.
- FIG. 14 shows what ESs are demultiplexed from the main TS and sub-TS by the basic stream selection table and the extended stream selection table.
- the demultiplexing unit is shown, and on the upper side, a combination of a basic stream selection table and an extended stream selection table is shown.
- FIG. 15 shows changes in stream number assignment in 2D output mode and 3D output mode.
- the column indicates the stream number of primary video stream # 1, the stream number of primary audio streams # 1 and # 2, the stream number of PG_text subtitle streams # 1, # 2 and # 3, and the stream of IG streams # 1 and # 2. Numbers are shown.
- the ES surrounded only by the broken line on the left side is an ES that is subject to demultiplexing only in the 2D output mode, and indicates that playback is permitted by the stream selection table (STN_table).
- the ES surrounded only by the broken line on the right side is an ES that is subject to demultiplexing in the 3D output mode, and indicates that playback is permitted by the extension stream selection table (STN_table_SS).
- ES ES surrounded by both the left and right broken line frames indicates the ES to be demultiplexed in 3D output mode.
- the MPEG4-MVC base-view video stream is surrounded by both the left and right dashed frames, so it can be seen that it is the playback target in both 2D output mode and 3D output mode .
- the MPEG4-MVC dependent-view video stream is only played back in the 3D output mode because it is surrounded only by the broken line on the right side.
- PG stream # 1 and PG stream # 2 are 2DPG streams, and are surrounded by left and right broken line frames. This indicates that PG stream # 1 and PG stream # 2 are playback targets in both the 2D output mode and the 3D output mode. On the other hand, since the left-eye PG stream and the right-eye PG stream are surrounded only by the right broken line frame, it can be seen that they are reproduced only in the 3D output mode.
- IG stream # 1 and IG stream # 2 are 2DIG streams, and are surrounded by left and right broken line frames. This indicates that IG stream # 1 and IG stream # 2 are to be played back only in the 2D output mode.
- IG stream # 1 and IG stream # 2 are to be played back only in the 2D output mode.
- the left-eye IG stream and the right-eye IG stream are surrounded only by the right dashed line frame, it can be seen that they are reproduced only in the 3D output mode.
- the dependent view video stream is added as a playback target in the stream type of the video stream in the 3D output mode.
- the PG stream stream type is added with the left-eye PG stream and the right-eye PG stream.
- the IG stream type is the left-eye IG stream and the right-eye IG stream. I understand that.
- the reason why the left-eye PG stream and the right-eye PG stream are added as playback targets is to realize stereoscopic playback using the left-eye PG stream and right-eye PG stream in the 3D output mode.
- the reason why the left-eye IG stream and the right-eye IG stream are added as playback targets is to realize stereoscopic playback using the left-eye IG stream and right-eye IG stream in the 3D output mode.
- FIG. 16 shows the internal structure of the playback device.
- the playback apparatus includes a reading unit 201, a memory 202, a register set 203, a decoder 204, a demultiplexing unit 205, a plane memory set 206, a shift unit 207, a layer synthesis unit 208, a transmission / reception unit 209, and a playback control unit. 210.
- the internal configuration of this figure only describes the minimum necessary components for implementing the playback apparatus including the problem solving means. A more detailed internal configuration will be left to the description of the subsequent embodiment.
- the reading unit 201 reads an index table, a program file, a playlist information file, a stream information file, and a stream file from the recording medium.
- the reading unit 201 uses the extent start point information in the clip base information in the 3D stream information file and the extent start point information in the clip dependent information to perform stereoscopic interleaving.
- the stream file is divided into an ATC sequence 1 corresponding to the main TS and an ATC sequence 2 corresponding to the sub-TS, and the ATC sequence 1 and the ATC sequence 2 are stored in separate read buffers. .
- This division includes processing for extracting source packets from the stereoscopic interleaved stream file and adding them to the ATC sequence 1 by the number of packets of the source packet number indicated in the extent start point information in the clip dependent information; This is done by repeating the process of extracting the source packet from the stereoscopic interleaved stream file and adding it to the ATC sequence 2 by the number of packets of the source packet number indicated in the extent start point information in the clip base information.
- the memory 202 stores a combined stream registration sequence obtained by combining the basic stream selection table and the extended stream selection table included in the playlist information.
- the register set 203 is composed of various registers for the playback device to operate.
- the decoder 204 for each stream type includes a video decoder 211, a PG decoder 212, an IG decoder 213, and an audio decoder 214.
- the demultiplexing unit 205 includes a source depacketizer that converts a source packet into a TS packet, a PID filter that performs packet filtering, and has a packet identifier described in the stream entry of the basic stream selection table of 3D playlist information
- the source packet is converted to a TS packet and then output to the decoder.
- the source packet having the packet identifier described in the stream entry of the stereoscopic stream selection table of the 3D playlist information is converted to a TS packet and then the decoder Output to.
- Which packet identifier is used among the plurality of entries in the basic stream selection table and the plurality of entries in the extended stream selection table depends on the setting of the stream number register in the player status register.
- the stream number register is a register that stores the current stream number.
- the plane memory set 206 is composed of a plurality of plane memories.
- plane memories constitute a layer model, and the contents stored in each plane memory are used for layer synthesis.
- the plane memory set includes plane memory for the left eye and plane memory for the right eye, and decodes uncompressed picture data and dependent view view components obtained by decoding the base view view components of individual access units.
- the uncompressed picture data obtained is written into the left-eye plane memory and the right-eye plane memory. This writing is performed when the playback start time indicated in the presentation time stamp of each access unit comes.
- the decoded picture data is written to either the left-eye plane memory or the right-eye plane memory depends on the base view indicator in the playlist information. If the base view indicator designates the base view video stream as "for left eye", the picture data that is the view component of the base view video stream is written to the plane memory for the left eye, and the view component of the dependent view video stream Is written in the plane memory for the right eye.
- the base view indicator specifies the base view video stream as “for right eye”
- the picture data that is the view component of the base view video stream is written to the plane memory for the right eye, and the view of the dependent view video stream is displayed.
- the picture data as a component is written into the plane memory for the left eye.
- the shift unit 207 shifts the pixel coordinates.
- the layer synthesis unit 208 performs layer synthesis in a plurality of plane memories.
- the transmission / reception unit 209 When the transmission / reception unit 209 is connected to another device in the home theater system via an interface, the transmission / reception unit 209 goes to the data transmission phase through the negotiation phase and performs data transmission.
- this negotiation phase the capabilities (including decoding capability, playback capability, and display frequency) of the counterpart device are ascertained and set in the player setting register to determine the transmission method for subsequent transmissions.
- Including a mutual authentication phase in which each device is validated.
- one line of uncompressed and plain text pixel data in the combined picture data is transferred to the display device at a high transfer rate in accordance with the horizontal synchronization period of the display device.
- audio data in uncompressed / plaintext format is sent to other devices (including not only the display device but also an amplifier and a speaker) connected to the playback device. Forward.
- devices such as a display device, an amplifier, and a speaker can receive uncompressed / plaintext picture data and uncompressed / plaintext audio data, and can realize reproduction output.
- the counterpart device has a decoding capability, it is possible to pass-through transmission of a video stream and an audio stream.
- pass-through transmission a video stream and an audio stream can be transmitted in a compressed / encrypted format.
- the playback control unit 210 performs random access from an arbitrary time point on the time axis of the video stream. Specifically, when playback from an arbitrary time point on the time axis of the video stream is ordered, an access unit corresponding to the arbitrary point of time using the basic entry map and the extended entry map in the 3D stream information file Search the source packet number of.
- the access unit includes a set of a view component of a base view video stream and a view component of a dependent view video stream, and a source packet storing an access unit delimiter for the access unit by the search. Source packet numbers are identified. Random access is performed by executing reading from the source packet number and decoding.
- the main TS, sub-time is used by using the in-time and out-time specified in the main path information of the 3D playlist information and the in-time and out-time specified in the sub-path information.
- Play list reproduction is started by executing random access to each TS.
- the video decoder 211 is a representative decoder among the decoders in the decoder set 204, and is preloaded with view components constituting the dependent-view video stream, and is located at the head of the closed GOP in the base-view video stream. Decodes the view component of the picture type (IDR type) intended for refresh. In this decoding, all of the coded data buffer and the decoded data buffer are cleared. After decoding the IDR type view component in this way, the subsequent view component of the base view video stream that is compression-encoded based on the correlation with the view component and the view component of the dependent view video stream are decoded. . If uncompressed picture data for the view component is obtained by decoding, the picture data is stored in the decoded data buffer, and the picture data is used as a reference picture.
- IDR type picture type
- motion compensation is performed for the subsequent view component of the base view video stream and the view component of the dependent view video stream. If uncompressed picture data is obtained for the subsequent view component of the base-view video stream and the view component of the dependent-view video stream by motion compensation, these are stored in the decoded data buffer and used as reference pictures.
- the above decoding is performed when the decoding start time indicated in the decoding time stamp of each access unit arrives.
- the configuration of the decoder in the PG stream is 1 decoder + 1 plane when the 1 plane + offset method is adopted.
- the configuration is 2 decoders + 2 planes.
- the decoder configuration in the IG stream is a 2 decoder + 2 plane configuration when the 3D-LR system is adopted.
- the configuration is 1 decoder + 1 plane.
- the structure is 1 decoder + 1 plane.
- the left-eye PG stream and the right-eye PG stream each include a plurality of display sets.
- a display set is a collection of functional segments that constitute one screen display.
- the functional segment is a processing unit that is stored in the payload of a PES packet of about 2 Kbytes and supplied to the decoder, and reproduction control is performed using DTS and PTS.
- Epoch start display set is a collection of functional segments that start the memory management by resetting the composition buffer, code data buffer, and graphics plane in the graphics decoder, and are necessary for screen configuration. Includes all functional segments.
- Normal case display set is a display set that performs screen configuration while maintaining memory management of the composition buffer, code data buffer, and graphics plane in the graphics decoder. It includes functional segments that are differences.
- An acquisition point display set is a display set that includes all functional segments necessary for screen composition, but is a composition buffer, code data buffer, and graphics plane memory in a graphics decoder. It is a display set that does not reset management. This acquisition point display set may have functional segments with different contents from the previous display set.
- the start and end points of these display sets are assigned to the same point on the playback time axis in the STC sequence for the left eye and the right eye.
- the types of display sets existing at the same time point on the time axis are the same. That is, if the left-eye display set is an epoch start display set, the right-eye display set at the same time point in the STC sequence time axis becomes an epoch start display set.
- the right-eye acquisition point display set at the same time on the time axis of the STC sequence is also an epoch start display set.
- Each display set includes multiple functional segments.
- the plurality of functional segments include the following.
- the object definition segment is a functional segment that defines a graphics object.
- the graphics definition segment defines a graphics object by using a code value and a run length of the code value.
- the pallet definition segment includes pallet data indicating the correspondence between each code value and luminance, red color difference / blue color difference.
- the correspondence between the code value, the luminance, and the color difference is set to the same content.
- Window definition segment is a functional segment that defines a rectangular frame called a window in a plane memory for expanding an uncompressed graphics object on the screen. Drawing of graphics objects is restricted inside the plane memory, and drawing of graphics objects is not possible outside this window.
- the playback device Since a part of the plane memory is designated as a window for displaying graphics, the playback device does not need to draw graphics for the entire plane. Graphics only need to be drawn for a window of a limited size. Since it is possible to omit drawing of the display plane other than the window, the burden on the software on the playback device side is much lighter.
- the screen configuration segment is a functional segment that defines the screen configuration using the graphics object, and includes a plurality of control items for the composition controller in the graphics decoder.
- the screen configuration segment is a functional segment that defines the details of the display set in the graphics stream and also defines the screen configuration using the graphics object.
- Such screen configurations include Cut-In / Out, Fade-In / Out, Color Change, Scroll, and Wipe-In / Out.
- End segment A functional segment located at the end of a plurality of functional segments belonging to one display set.
- the playback device interprets the screen segment to the end segment as a functional segment constituting one display set.
- the start point of the display set is specified by the DTS of the PES packet storing the screen composition segment
- the end point of the display set is specified by the PTS of the PES packet storing the screen composition segment.
- the graphics stream for the left eye and the graphics stream for the right eye are packetized elementary streams (PES), the screen configuration segment is stored in the PES packet, and the PTS of the PES packet storing the screen configuration segment is the screen configuration segment. Indicates when to perform display by the display set to which it belongs.
- PES packetized elementary streams
- the PTS value of the PES packet storing the screen composition segment is the same for the left-eye video stream and the right-eye video stream.
- the PG decoder is obtained by decoding a "coded data buffer” that stores functional segments read from the PG stream, a "stream graphics processor” that decodes screen composition segments to obtain graphics objects.
- the "object buffer” that stores the graphics objects that are stored, the “composition buffer” that stores the screen composition segments, and the screen composition segments that are stored in the composition buffer are decoded, and the control items in these screen composition segments And a “composition controller” that configures a screen on the graphics plane using the graphics object obtained in the object buffer.
- a transport buffer for adjusting the input speed of TS packets that make up the functional segment exists in the previous stage of the graphics plane.
- a CLUT unit that converts the pixel code that configures the graphics object stored in the graphics plane into luminance and color differences based on the graphics plane and the palette definition segment, and a shift unit for plane shift And exist.
- the pipeline in the PG stream is a graphics object obtained by decoding the object definition segment belonging to the display set that is a graphics decoder and writing the graphics object to the object buffer, and decoding the object definition segment belonging to the preceding display set. Is performed simultaneously with the process of writing the data from the object buffer to the plain memory.
- FIG. 17 shows the internal structure of the PG decoder.
- FIG. 5A shows a decoder model for displaying in the 1 plane + offset mode method.
- FIG. 5B shows a decoder model for displaying LR data.
- the part corresponding to the main body of the PG decoder is surrounded by a black frame, and the part corresponding to the latter stage of the graphics decoder is surrounded by an alternate long and short dash line.
- the PG decoder has a single decoder configuration
- the graphics plane also has a single plane configuration.
- the graphics plane output is divided into left-eye output and right-eye output, and a shift unit is added to each left-eye output and right-eye output.
- FIG. 5B there are two sets of transport buffer-graphics decoder-graphics plane-CLUT part, and the left-eye stream and the right-eye stream can be processed independently.
- the graphics decoder Since the offset sequence is included in the video stream for the right eye, in the plain offset format, the graphics decoder has one decoder configuration, and the output of this one graphics decoder is switched between the left eye and the right eye.
- a text subtitle stream is composed of a plurality of subtitle description data.
- the text subtitle decoder uses a “subtitle processor” that separates text codes and control information from subtitle description data, a “management information buffer” that stores text codes separated from subtitle description data, and font data. , "Text render” that expands the text code in the management information buffer into a bitmap, "object buffer” that stores the bitmap obtained by the expansion, and control information separated from the caption description data, A “drawing control unit” that executes control of text subtitle reproduction along the axis.
- the text subtitle decoder there is a “font preload buffer” that preloads font data, a “TS buffer” that adjusts the input speed of TS packets that make up the text subtitle stream, and a text subtitle stream that is prior to play item playback.
- a “caption preload buffer” for preloading.
- FIG. 18 shows the internal structure of the text subtitle decoder.
- FIG. 4A shows a decoder model of the text subtitle decoder in the 1 plane + Offset mode
- FIG. 4B shows a decoder model of the text subtitle decoder in the 3D-LR system.
- the portion corresponding to the text subtitle decoder main body is surrounded by a black frame
- the portion corresponding to the subsequent stage of the text subtitle decoder is surrounded by an alternate long and short dash line.
- a portion corresponding to the preceding stage of the text subtitle decoder is surrounded by a broken line frame.
- the graphics plane output is divided into left eye and right eye outputs, and a shift unit is added to each left eye output and right eye output.
- FIG. 5B there are a left-eye graphics plane and a right-eye graphics plane, and a bitmap developed by the text subtitle decoder is written to each of these graphics planes.
- the text subtitle stream Unlike the PG stream, the text subtitle stream generates subtitles by rendering engine by sending font data and character codes instead of sending graphics data as a bitmap. Therefore, stereoscopic viewing of subtitles is realized by 1plane + Offset mode. To do.
- Each of the left-eye IG stream and the right-eye IG stream includes a plurality of display sets, and each display set includes a plurality of functional segments.
- epoch start display sets normal case display sets
- acquisition point display sets acquisition point display sets
- epoch continuation display sets There are the following types.
- This object definition segment is the same as that of the PG stream, except that the graphics object of the IG stream defines the in-effect, out-effect of the page, normal state of the button member, selected state, and active state. Is.
- the object definition segments are grouped together that define the same state of the button member and those that constitute the same effect video.
- a group of object definition segments that define the same state is called a graphics data set.
- the dialog control segment includes a plurality of page information, and the plurality of page information is information defining the screen configuration of the multi-page menu.
- Each page information includes an effect sequence and a plurality of button information. And a reference value of the palette identifier.
- the button information is information that realizes an interactive screen configuration on each page constituting the multi-page menu by displaying the graphics object as one state of the button member.
- the effect sequence is an in-effect that is reproduced prior to display of a page corresponding to page information or an out-effect that is reproduced after display of the page by using a graphics object, and includes effect information.
- the effect information is information that defines an individual screen configuration when playing an in-effect or an out-effect, and what screen configuration should be executed in the window (partial area) defined by the window definition segment on the graphics plane.
- a screen composition object that defines the kika and effect period information indicating a time interval between the next screen composition in the same area.
- the screen configuration object in the effect sequence specifies the same control content as the screen configuration segment of the PG stream.
- the one that defines the graphics object used for the in-effect is arranged before the object definition segment that defines the graphics object used for the button member in the graphics data string.
- Each button information in the page information is information that realizes an interactive screen configuration on each page constituting the multi-page menu by displaying the graphics object as one state of the button member.
- the button information includes a set button page command.
- the set button page command causes the playback device to perform processing for setting a page other than the first page as the current page when the corresponding button member is activated. It is a command.
- End segment A functional segment located at the end of a plurality of functional segments belonging to one display set.
- the dialogue control segment to this end segment are interpreted as functional segments constituting one display set.
- control items of the same dialog control segment in the left-eye graphics stream and the right-eye graphics stream include button proximity information, selection timeout time stamp, user timeout duration, and composition timeout information.
- Button proximity information is information that designates which button should be selected when a certain button is in the selected state and there is a key operation instructing one of the up, down, left, and right directions. .
- Selection time-out time stamp indicates a time-out time for automatically activating the button member in the current page and causing the playback device to execute the set button page command.
- the user timeout duration indicates a timeout period for returning the current page to the first page so that only the first page is displayed.
- composition time-out information indicates a time for ending the interactive screen display by the dialog control segment.
- the start time of the display set is specified by the DTS of the PES packet storing the dialog control segment
- the end time of the display set is specified by the composition timeout time of the dialog control segment.
- these DTS and composition timeout time are set to the same time.
- the IG decoder is obtained by decoding a "coded data buffer” that stores functional segments read from the IG stream, a "stream graphics processor” that decodes screen composition segments to obtain graphics objects, and
- the "object buffer” that stores the graphics objects that are stored, the “composition buffer” that stores the screen composition segments, and the screen composition segments that are stored in the composition buffer are decoded, and the control items in these screen composition segments
- a “composition controller” that configures a screen on the graphics plane using the graphics object obtained in the object buffer.
- FIG. 19 shows a decoder model of the IG decoder.
- the portion corresponding to the IG decoder body is surrounded by a black frame, and the portion corresponding to the latter stage of the graphics decoder is surrounded by a one-dot chain line.
- the portion corresponding to the previous stage of the IG decoder is surrounded by a broken line frame.
- FIG. 19A shows a decoder model for displaying a 2D format IG stream in the LR format by the 1 plane + Offset mode method.
- FIG. 4B shows an IG stream decoder model, which is a decoder model for displaying LR data. *
- These decoders include a circuit for reflecting the value of the system parameter in the offset in order to control the depth information of the menu graphics from the program.
- Figure (b) is a two-decoder model, and the offset value can be changed by a command. Therefore, the depth information of the menu can be changed with a command. Offset values are also given different values on the left and right. On the other hand, in the case of the Depth method, Offset becomes invalid.
- the composition controller in the graphics decoder displays, using the graphics data of the graphics data set corresponding to the selected state, the button member that is the current button among the button members that exist on the dialog screen, and the other button members are displayed as normal. By displaying the graphics data set corresponding to the state, the initial display of the interactive screen is realized.
- the button number of the button member in the normal state located around the current button and existing in the direction designated by the user operation Writing to the register, the button member that has newly become the current button is changed from the normal state to the selected state by the writing.
- the dialog screen is updated by taking out the graphics data constituting the active state from the graphics data set and displaying it. Is realized. *
- dialog screen updates need to be executed in common for the left eye and right eye, so in the 2-decoder model, the composition controller is shared between the left-eye graphics decoder and the right-eye graphics decoder. It is desirable to do.
- the left-eye navigation command and the right-eye navigation command in the stereoscopic IG stream are made the same, and the button configuration of the graphics object for 3D and 2D is made the same, thereby realizing mutual switching.
- the 2DIG stream and the stereoscopic IG stream have the same attributes and number of navigation commands and button information, only the display of the graphics object can be switched. Switching from 3D-LR mode to L image only can be switched without reloading, but the display position may be biased. It is desirable that the title creator's intention of which one is adopted is indicated by a flag, and the playback apparatus performs switching based on this flag.
- ⁇ Reload does not occur when switching between 1plane + Offset mode and 2D mode. This is because it is not necessary to load the IG stream and only invalidate the Offset.
- ⁇ Reloading occurs when switching between 3D-LR mode and 2D mode because the stream is different.
- the process of superimposing the pixel values of the pixel data stored in the plane memory between layers in the layer model of the plane memory is executed for all combinations between the layers in the layer model. That is done.
- the process of superimposing the pixel values of the pixel data stored in the plane memory of the two layers in the layer model of the plane memory is performed for all combinations of the two layers in the layer model. It is done by executing.
- the pixel value of the line unit of the plane memory located in a certain layer is multiplied by the transmittance ⁇ as a weight
- the pixel value of the line unit of the plane memory located in the lower layer is multiplied by (1-transmission ⁇ ) Is added to the pixel values that have been weighted for luminance, and the addition result is used as the pixel value of the pixel in line units in the hierarchy.
- the layer composition is realized by repeatedly performing the superimposition between the hierarchies between the line-unit pixels located in two adjacent hierarchies of the layer model.
- a multiplication unit for multiplying individual pixel values by an equivalence rate and an addition unit for performing addition between pixels are provided. Including.
- FIG. 20 shows a circuit configuration for synthesizing the output of the decoder model and outputting it in the 3D-LR system.
- the layer model of the video plane, PG plane, and IG plane is surrounded by a black frame, and the portion corresponding to the subsequent stage of the plane memory is surrounded by a one-dot chain line.
- the plane memory configuration in the 3D-LR system is divided into a left plane and a right plane for each of the video plane, PG plane, and IG plane.
- the output of the plane memory is subjected to layer synthesis for each of the left eye and the right eye.
- FIG. 21 shows a circuit configuration for synthesizing the outputs of these decoder models and outputting them in the 1 plane + offset mode method.
- the layer model of the left-eye video plane, right-eye video plane, PG plane, and IG plane is enclosed by a black frame, and the portion corresponding to the subsequent stage of the plane memory is enclosed by a one-dot chain line.
- a video plane for the left eye and one for the right eye are prepared.
- the PG plane and IG plane are not separated for the left eye and the right eye, and there is only one common plane memory for the left eye and the right eye.
- the layer synthesis as described above is performed for each of the left-eye output and the right-eye output.
- FIG. 22 shows a circuit configuration for synthesizing the outputs of the decoder model and outputting in the upper end 2D subtitle playback mode method and the lower end 2D subtitle playback mode method.
- video planes for the left eye and for the right eye are prepared.
- the PG plane and IG plane are not separated for the left eye and the right eye, and there is only one common plane memory for the left eye and the right eye.
- pixels are shifted by 131 pixels upward or downward in accordance with the setting of the register (PSR32) indicating the video shift mode of the playback device described later.
- Each of the left-eye output and the right-eye output of the PG plane is composed of layers.
- the playback device Since the playback device needs to support all of the BD presentation mode, 1 plane + Offset mode, upper end 2D subtitle playback mode, and lower end 2D subtitle playback mode, the playback device basically has 2 hardware configurations.
- 1 plane + offset mode, 2D output mode, upper end 2D subtitle playback mode, and lower end 2D subtitle playback mode one of the set of 1 decoder + 1 plane is selected. It is invalidated to have a 1 decoder + 1 plane configuration.
- the manufacturer that manufactures the playback device can freely determine whether the above decoder model is a 1-decoder configuration, a 2-decoder configuration, a plane model is a 1-plane configuration, or a 2-plane configuration.
- a high-end model with a 2-decoder configuration and 2-plane configuration by design set it as having stereoscopic PG playback capability and stereoscopic IG playback capability, and selling as a low-priced model
- the lineup can be enhanced by setting no stereoscopic PG playback capability and no stereoscopic IG playback capability.
- Such a configuration for presence / absence of the reproduction capability of the stereoscopic PG and a configuration for the presence / absence of the reproduction capability of the stereoscopic IG exist in the register set.
- the register set includes a plurality of player status registers and a plurality of player setting registers.
- Each of the player status register and the player setting register is a register having a word length of 32 bits.
- Each register having a 32-bit length is given a register number, and the register to be accessed is specified by using this register number.
- each bit data that constitutes one word (32 bits) of each register is called b0 to b31.
- the most significant bit is called b31, and the least significant bit is called b0.
- the bit range from the bit position of the bx bit to the bit position of the by bit is expressed by the notation [bx: by].
- a 32-bit length bit string stored in the player setting register / player status register of a predetermined register number and having a value in an arbitrary bit range [bx: by] is the value of the operating system when the program operates. Treated as environment variables (system parameters or player variables).
- a program that performs playback control can acquire system parameters through system properties and an application programming interface (API). Unless otherwise prohibited, the program can rewrite the values of the player status register and player setting register. For an object-oriented programming language-based program, it is a requirement that legitimate authority is given to obtain and rewrite system parameters.
- the player status register is a hardware resource for storing the numerical value that is the operand when the MPU of the playback device performs an arithmetic operation or a bit operation, and the initial value is set when the optical disc is loaded.
- This is a register for determining the validity of the stored value when the state of the playback device changes, such as a change of the current play item.
- the stored value includes a current title number, a current playlist number, a current play item number, a current stream number, a current chapter number, and the like. Since the initial value is stored when the optical disk is loaded, the stored value is temporary. If the optical disk is ejected or the playback apparatus is turned off, the stored value loses its validity.
- the player setting register is different from the player status register in that power supply measures are taken. Since power supply measures are taken, the stored value is saved in a non-volatile memory when the playback device is turned off, and the stored value is restored when the playback device is powered on.
- FIG. 23 is a diagram depicting the internal configuration of the register set 203 and the playback control unit.
- the left side of the figure shows the internal configuration of the register set 203.
- the right side shows the internal structure of the playback control unit.
- the player status register and the player setting register to which the respective register numbers are assigned indicate what they are.
- PSR0 is a stream number register for the IG stream, and stores the current IG stream number.
- PSR2 is a stream number register for the PG stream and stores the current PG stream number.
- PSR24 is a setting of “Player Capability for 3D”. This indicates whether or not the playback device has the ability to perform stereoscopic playback.
- the playback control unit refers to the PSR 24 in the register set 203 and the stream selection table of the current playlist information in the memory, and a stream selection procedure that uniquely determines the current PG stream number and the current IG stream number in the current playlist. Exists.
- This stream selection procedure includes “Initialization”, “Procedure when playback condition changed”.
- FIG. 25 shows the bit assignment of PSR32.
- PSR32 represents the video shift mode of the playback device. This value is set through the API or command of the BD program, and video_shift_mode for the current PG stream selected by stream switching is acquired and set from the stream registration information in the extension stream selection table.
- FIG. 24 shows the bit assignment of PSR24.
- PSR24 represents the 3D capability of the playback device.
- the program in the recording medium cannot change the value of PSR24.
- B0 indicates stereoscopic display 1280 x 720 50p video display capability.
- the stereoscopic 1280 ⁇ 720 50p video display capability when set to “0”, indicates that the playback device does not have the processing capability to display 1280 ⁇ 720/50 Hz progressive video.
- B2 indicates stereoscopic PG capability.
- the stereoscopic PG capability is set to “0” to indicate that the playback device does not have the capability of playing back the stereoscopic PG. Setting it to “1” indicates that the playback device has the ability to play back the stereoscopic PG.
- B3 indicates stereoscopic IG capability.
- the stereoscopic IG capability is set to “0” to indicate that the playback device does not have the ability to reproduce the stereoscopic IG. Setting it to “1” indicates that the playback device has the ability to play back the stereoscopic IG.
- B5 indicates the capability of BD-J mode in 3D output mode. When this b5 is set to “1”, it indicates that the playback apparatus can process the BD-J mode in the 3D output mode. When b5 is set to “0”, it indicates that the playback device can process the BD-J mode in the 3D output mode. Since the usage of b5 of PSR24 is out of the main point of the present embodiment, the explanation will be given to the subsequent embodiment.
- each of the IG decoder and PG decoder is configured with two decoders and each of IG and PG is stereoscopically supported.
- a configuration is possible in which only PG is supported for stereoscopic viewing and IG is only supported for 1 plane + offset.
- each IG and PG in order to sell the playback device as a low-priced model, it is possible to configure each IG and PG to be compatible with 1 plane + Offset while each IG decoder and PG decoder is configured with two decoders. Since it is configured with a common configuration of two decoder models, the availability of IG stereoscopic view and PG stereoscopic view can be changed according to the grade of the model, so when the manufacturer commercializes the playback device It will be possible to enhance the lineup.
- both PG and IG are configured with one decoder, the possibility of stereoscopic playback is clearly indicated, so even if the playlist to be played is compatible with stereoscopic viewing, the playback type is incorrect. Setting to stereoscopic PG and stereoscopic IG can be prevented.
- reproduction control described above can be realized by describing the processing procedure shown in the flowcharts of FIGS. 26 to 32 in an object-oriented compiler language and causing the computer to execute it.
- FIG. 26 shows a playlist playback procedure.
- This flowchart constitutes a loop in which the current play item number is set to 1 in step S1, and then the processing in steps S2 to S6 is repeated.
- the stream number is determined by Procedure when playback condition is changed (step S2), the stream file storing the ES corresponding to the stream number is opened, the source packet sequence is read (step S3), and read.
- the source packet sequence corresponding to the stream number is instructed to be demultiplexed (step S4), and the read source packet is output from the in-time to the out-time of the play item and out from the in-time of the sub-play item.
- the process of instructing the decoder to play until time is repeated until the current play item number reaches the final number. If it is not the final number (No in step S6), the current play item number is incremented and the process proceeds to step S2.
- step S7 it is determined whether or not there is a stream selection request in step S7, and if there is a request, the requested stream number is set to x and Procedure when stream change is requested is executed (step S8). If it is the final number, the process ends (Yes in step S6).
- the PG_text subtitle stream referred to by the PG_text subtitle stream number of PSR2 is selected using the output mode (PSR22), the stereoscopic PG capability in PSR24, and is_SS_PG.
- FIG. 27 is a flowchart showing the processing procedure for the Procedure when playback condition is changed for the PGtestST stream.
- the processes in steps S11 to S22 are common to the 3D output mode and the 2D output mode.
- Steps S23 to S28 are processing specific to the 3D output mode.
- step S11 the current PG_text subtitle stream number is acquired from PSR2, and in step S12, it is determined whether the current PG_text subtitle stream number is for the PG stream or the text subtitle stream.
- step S13 checks whether or not the PG stream corresponding to the current PG_text subtitle stream number satisfies the condition (A) and the condition (B).
- condition (A) and condition (B) are defined as follows.
- the playback device has the ability to decode the PG stream specified from the current PG_text subtitle stream number.
- the playback device has the ability to play the specified language.
- step S14 it is checked whether or not the text subtitle stream corresponding to the current PG_text subtitle stream number satisfies the condition (A) and the condition (B).
- the playback device has the ability to expand the character code of the text subtitle stream of the current PG_text subtitle stream number into a bitmap. Such playback capability is indicated by PSR 30 in the PSR set 23.
- the playback device must have the ability to support the language characteristics of the text subtitle stream of the current PG_text subtitle stream number.
- the text subtitle stream in that language can be said to be “decodable” by expanding the character code in that language into a bitmap. Capabilities and support capabilities that support the characteristics of the language need to exist on the playback device side.
- Japanese subtitles can be judged to support language characteristics only when the functions “horizontal writing”, “vertical writing”, “line ending prohibition”, and “ruby” are supported.
- Arabic subtitle display can be judged to support language characteristics only after the functions of “Right-to-Left Drawing” and “Language / Hugging” are supported.
- a text subtitle stream of a certain language can be developed into a bitmap and the characteristics of the language can be supported, it can be said that the above conditions (A) and (B) are satisfied. If the text subtitle stream of the language can be expanded into a bitmap, but the characteristics of the language cannot be supported, the condition (B) is not satisfied and only the condition (A) is satisfied be able to.
- the support capability for each language is set for each language in each bit of PSR48 to 61 in the register set. Specifically, in PSR48 to 61, there are flags corresponding to each of the 3-byte language codes defined in IS0639-2 / T. In these flags, text subtitles corresponding to each language code are present. The presence or absence of display capability is set.
- the 3-byte code “ita” indicates Italian
- the 3-byte code “jpn” indicates Japanese
- the 3-byte code “jav” indicates Javanese. Since the language codes defined in IS0639-2 / T cover 430 languages in this way by 3-byte codes, the text subtitle stream described in the stream number table can be decoded when determining the current PG_text subtitle stream. In determining whether or not, the flags in the PSRs 48 to 61 are referred to. In this way, it is possible to properly determine whether or not even a minority language can be decoded.
- step S15 the playback device determines whether or not the condition (Z) is satisfied.
- the condition (Z) means that the user plays back subtitles in an unsupported language that does not support language characteristics. This intention is shown in PSR30 in the register set.
- step S16 is executed. This is to determine whether or not the number of PG_text subtitle streams in the stream selection table of the current play item is zero. If there is no PG_text subtitle stream that is permitted to be played back in the stream selection table, the PG_text subtitle stream number in PSR2 is maintained (step S17).
- step S18 is executed. This is a step for asking the validity of the current PG_text subtitle stream number, is the current PG_text subtitle stream number less than or equal to the total number of stream_entry in the stream selection table, and satisfies the conditions (A) and (B)? Determine if.
- step S20 determines whether or not the current PG_text subtitle stream number is equal to or less than the total number of stream_entry in the stream selection table and satisfies the conditions (A) and (Z). If it is satisfied, the PG_text subtitle stream number of the text subtitle in the unsupported language is set in PSR2, but since the user intends to play back the unsupported language, PSR2 is not changed (step S21). If it is determined in step S20 that the condition is not satisfied, an optimum stream for the current play item is selected (step S22).
- the subsequent steps S23 to S28 are processing specific to the 3D output mode. Specifically, in the 3D output mode, first, upper and lower end reproduction type determination processing is executed (step S23). In the upper / lower end playback type determination process, when the playback type is not set to any one of the upper end side 2D subtitle playback and the lower end side 2D subtitle playback (step S24: No), the PG identified by the PG stream number of PSR2 The stream is_SS_PG is acquired from the stream registration information in the extension stream selection table (step S25), and the acquired is_SS_PG flag is “1” and the b2 stereoscopic PG capability in the PSR24 is “1”. Is determined (step S26). If step S26 is Yes, the playback type is set to stereoscopic PG using the left-eye PG stream and the right-eye PG stream (step S27).
- the demultiplexing unit performs demultiplexing of the TS packet of the packet identifier corresponding to the stream number of the current stream stored in PSR2 but indicated by the packet identifier reference in the right eye stream entry and the left eye stream entry. Make it.
- Step S28 the playback type is set to 1 plane + Offset (Step S28).
- the playback type is 1 plane + Offset PG
- the offset sequence will be described. There are a plurality of offset sequences used in the 1 plane + Offset mode in the video access unit of the dependent-view video stream.
- the video access unit of the dependent-view video stream is configured by arranging the video access unit delimiter, sequence parameter set, picture parameter set, MVC scalable nesting SEI message, first view component, sequence end code, stream end code Is done.
- a user data container exists in the MVC scalable nesting SEI message.
- the user data container is unregistered user data (unregistered user data), and there are three types: closed caption information, GOP structure map, and offset metadata. Which of these types is specified in the type_indeicator in the container.
- the offset metadata is a sequence list for the PG plane, the IG plane, and the BD-J plane.
- the presentation graphics_text subtitle and the IG / BD-J plane are played back in the 1 plane + Offset mode in the stereoscopic playback mode. Used for offset setting. Specifically, offset control for the PG plane, the IG plane, and the BD-J plane when the graphics to be combined with the picture data is reproduced in the 1 plane + Offset mode is shown.
- the offset metadata must be stored in the MVC scalable nesting SEI message of the first video component of each GOP in the dependent view access unit encoding order.
- the offset sequence is a parameter sequence that indicates control parameters in the case of synthesizing each piece of picture data belonging to the group of pictures and graphics, and is composed of the number of control parameters indicated by number_of_displayed_frames_in_GOP.
- the This control parameter includes plane offset direction information and a plane offset value.
- the plane offset direction information (Plane_offset_direction) indicates the offset direction in the plane.
- the value “0” is set to the front, that is, the plane memory exists between the TV and the viewer, and the plane is shifted to the right in the left view period, and the plane is shifted to the left in the right view period.
- Behind setting with value 1, that is, the plane memory exists behind the TV or screen, and the plane is shifted to the left in the left view period, and the plane is shifted to the right in the right view period.
- the plane offset direction information indicates the front setting
- the Z-axis coordinate of the control parameter in the three-dimensional coordinate system is a positive coordinate.
- the plane offset direction information indicates behind setting, the Z-axis coordinate of the control parameter in the three-dimensional coordinate system is a negative coordinate.
- the plane offset value (Plane_offset_value) is a designation of the horizontal displacement amount of the pixels constituting the graphics, and indicates the plane offset value in units of pixels.
- FIG. 28 is a flowchart showing the upper / lower end reproduction type determination processing procedure.
- the video_shift_mode of the PG stream specified by the PG stream number of PSR2 is acquired from the stream registration information in the extension stream selection table (step S101), the acquired video_shift_mode is “Down”, and It is determined whether or not the video shift mode of the playback device in PSR32 is “Down” (step S102).
- Step S102 the playback type is set to 2D subtitle playback at the upper end side (Step S103).
- the playback type is 2D subtitle playback at the upper end side
- PG playback in the upper end 2D subtitle playback mode is executed.
- the demultiplexing unit performs demultiplexing of the TS packet of the packet identifier indicated by the packet identifier reference in the stream entry corresponding to the stream number of the current stream stored in PSR2, and the shift unit The picture data stored in the right-eye and left-eye video planes are shifted downward by 131 pixels.
- Step S102 it is determined whether or not the video_shift_mode acquired in Step S101 is “Up” and the video shift mode of the playback device in PSR32 is “Up” (Step S104). If step S104 is Yes, the playback type is set to 2D subtitle playback at the lower end (step S105).
- the playback type is the lower-end 2D subtitle playback
- PG playback in the lower-end 2D subtitle playback mode is executed.
- the shift unit shifts the picture data stored in the right-eye and left-eye video planes upward by 131 pixels.
- FIG. 29 is a flowchart showing a processing procedure for selecting an optimal PG_text subtitle stream for the current playitem.
- step S30 it is checked whether or not all the PG_text subtitle streams satisfy the conditions (a) to (c).
- conditions (a) to (c) are defined as follows.
- the playback device has the ability to decode PG stream i.
- PG_language_code of PG stream i matches the language setting in the playback device.
- the language setting in the playback apparatus is indicated by PSR17 in the register set.
- conditions (a) to (c) are defined as follows.
- the playback device must have the ability to expand the character code of the text subtitle stream i into a bitmap.
- the playback device has the ability to support the language attribute of the text subtitle stream i.
- step S31 it is determined whether or not the playback device satisfies the condition (Z) (: playback of unsupported language) described in the previous flowchart. If not, it is determined in step S32 whether there is a PG_text subtitle stream that satisfies the conditions (a) to (c). If there is a PG_text subtitle stream satisfying the conditions (a) to (c), select the stream entry whose corresponding stream entry is at the top in the stream selection table, and select the selected PG_text subtitle stream number. , PSR2 is set (step S33).
- step S34 If there is no PG_text subtitle stream that satisfies the conditions (a) to (c), it is determined in step S34 whether or not there is a PG_text subtitle stream that satisfies a more relaxed condition.
- the relaxed condition means that the condition (a) and the condition (b) are satisfied, and it is determined in step S34 whether there is a PG_text subtitle stream that satisfies this condition. If there is a PG_text subtitle stream satisfying the conditions (a) and (b), select the stream entry whose corresponding stream entry is at the top in the stream selection table and select the selected PG_text subtitle stream number. , PSR2 is set (step S36).
- step S35 If not, the PG_text subtitle stream number of 0xFFF is set in PSR2 (step S35). If it is determined in step S31 that the condition (Z) is satisfied, it is determined in step S37 whether a PG_text subtitle stream that satisfies another relaxed condition exists. The other relaxed conditions satisfy the conditions (a) and (c). In step S37, it is determined whether or not there is a PG_text subtitle stream that satisfies these conditions.
- the PG_text subtitle stream number of the selected PG_text subtitle stream satisfying the condition (a) and the condition (c) is selected by selecting the stream stream corresponding to the first entry in the stream selection table. Is set to PSR2 (step S38).
- step S39 it is determined in step S39 whether there is a PG_text subtitle stream that satisfies the condition (a). If present, select the PG_text subtitle stream that satisfies the condition (a) whose corresponding stream entry is at the top in the stream selection table, and set the selected PG_text subtitle stream number to PSR2 (Step S40). If not, 0xFFF is set to PSR2 (step S35).
- FIG. 30 is a flowchart showing a procedure of “Procedure when stream change is requested” to be executed when a stream change is requested by the set stereoscopic stream command (SetstreamSSCommand).
- step S41 it is determined whether the number X designated by the operand in the set stereoscopic stream command means the stream number for the PG stream or the stream number for the text subtitle stream.
- step S42 checks whether or not the PG stream (PGx) corresponding to the number X satisfies the following conditions (A) and (B).
- the playback device must be capable of decoding the PG stream specified by number X.
- Step S43 checks whether or not the text subtitle stream (textSTx) corresponding to the number X satisfies the following condition (A) and condition (B).
- the playback device has the ability to expand the character code of the text subtitle stream X corresponding to the number X into a bitmap.
- the playback device must have the ability to support the language attribute of the text subtitle stream X corresponding to the number X.
- step S44 the playback device checks whether or not the condition (Z) is satisfied, and performs the determination in step S45. In this determination, it is determined whether or not the number is equal to or less than the total number of stream_entry in the stream selection table and the conditions (A) and (B) are satisfied. If it satisfies, the PG_text subtitle stream of the PG_text subtitle stream number corresponding to the number X is selected and set to PSR2 (step S46).
- step S45 If it is determined in step S45 that the condition is not satisfied, the determination in step S47 is executed. This determination is to determine whether the number is equal to or less than the total number of stream_entry in the stream selection table and satisfies the conditions (A) and (Z). If satisfied, the PG_text subtitle stream corresponding to the number X is selected, and the PG_text subtitle stream number is set to PSR2 (step S48).
- step S49 the determination in step S49 is performed. This determination is to determine whether the number X is 0xFFF. Otherwise, the PSR2 value is maintained assuming that there is no PG_text subtitle stream that is permitted to be played back in the stream selection table (step S50).
- an optimal PG_text subtitle stream is selected for the current play item (step S51).
- the selection of the optimum PG_text subtitle stream is the same as that shown in FIG.
- step S52 upper / lower end reproduction type determination processing is executed (step S52).
- step S53 when the playback type is not set to any one of the upper end side 2D subtitle playback and the lower end side 2D subtitle playback (step S53: No), the PG stream specified by the PG stream number X Is_SS_PG of X is acquired from the stream registration information in the extension stream selection table (step S54), and whether or not the acquired is_SS_PG flag is “1” and the stereoscopic PG capability in PSR24 is “1” is determined. Determination is made (step S55). If step S55 is Yes, the playback type is set to stereoscopic PG (step S56). If step S55 is No, the playback type is set to 1 plane + Offset (step S57).
- FIG. 31 is a flowchart showing the processing procedure of Procedure when stream is requested to be executed when a stream change is requested by a set stream command or a user operation requesting a stream number change.
- step S58 the stream number designated by the operand of the set stream command or the stream number designated by the user operation requesting the stream number change is set to the number X, and then steps S41 to S57 are performed. Execute the process. Since the processing contents of steps S41 to S57 are the same as those in FIG. 30, the same reference numerals are given and the description thereof is omitted.
- the IG stream referred to by the IG stream number of PSR0 is selected using the output mode in PSR22, the stereoscopic PG capability in PSR24, and is_SS_IG.
- FIG. 32 is a flowchart showing a processing procedure for determining the current IG stream and determining the reproduction type of the current IG stream.
- FIG. 32 (a) shows Procedure when playback condition is changed for determining the current IG stream when the play item changes and the playback state of the playback device changes.
- steps S61 to S65 are steps common to the 2D output mode and the 3D output mode
- steps S64 to S67 are steps unique to the 3D output mode.
- Step S61 is a determination of whether or not the number of entries in the stream selection table is 0. If it is 0, the value of PSR0 is maintained (step S64).
- Step S62 determines whether the number of entries in the stream selection table is larger than that in PSR0 when the number of entries in the stream selection table is not zero. If Step S62 is Yes, PSR0 is maintained (Step S65). If the value of PSR1 is larger than the entry number, 1 is set in PSR0 (step S63). The subsequent steps S64 to S67 are unique to the 3D output mode. In the 3D output mode, is_SS_IG of the IG stream specified by the IG stream number of PSR0 is acquired from the stream registration information in the extension stream selection table (step S64), the acquired is_SS_IG flag is “1”, and in PSR24 It is determined whether or not the stereoscopic IG capability indicated by b3 is “1” (step S65).
- step S65 the playback type is set to stereoscopic IG (step S66).
- the playback type is stereoscopic IG
- the packet identifier reference in the right eye stream entry and the left eye stream entry corresponding to the stream number of the current stream stored in PSR0 among the plurality of stream registration information in the extended stream selection table is displayed. Used to perform stereoscopic reproduction.
- the demultiplexing unit performs demultiplexing of the TS packet of the packet identifier corresponding to the stream number of the current stream stored in PSR0 but designated by the packet identifier reference in the right eye stream entry and the left eye stream entry. Make it.
- Step S65 If Step S65 is No, the playback type is set to 1 plane + Offset (Step S67). *
- the offset sequence indicated by the stereoscopic IG offset sequence ID reference information corresponding to the stream number of the current stream stored in PSR0 among a plurality of stream registration information in the extended stream selection table Is used to perform IG playback in 1plane + Offset mode.
- the offset sequence indicated by the stereoscopic IG offset sequence ID reference information is extracted from the video decoder and supplied to the shift unit To do.
- FIG. 32 (b) is a flowchart showing the setting procedure of PSR0 when a stream change is requested by a set stereoscopic stream command (SetstreamSSCommand), a set stream command, and a user operation requesting a stream number change.
- SetstreamSSCommand set stereoscopic stream command
- FIG. 32 (b) is a flowchart showing the setting procedure of PSR0 when a stream change is requested by a set stereoscopic stream command (SetstreamSSCommand), a set stream command, and a user operation requesting a stream number change.
- the stream number specified by the command operand or the user operation is number X.
- Step S71 in this flowchart determines whether the number of entries in the stream selection table is greater than the number X. If X satisfies this condition, X is set in PSR0 (step S74). If X is larger than the entry number, 1 is set to PSR0 (step S72). In the 3D output mode, is_SS_IG of the IG stream specified by the IG stream number of PSR0 is acquired from the stream registration information in the extension stream selection table (step S73), the acquired is_SS_IG flag is “1”, and in PSR24 It is determined whether or not the stereoscopic IG capability is “1” (step S74). If step S74 is Yes, the playback type is set to stereoscopic IG (step S75). If step S74 is No, the playback type is set to 1 plane + Offset (step S76).
- FIG. 33 shows what packet identifiers are output to the demultiplexing unit according to the combined stream registration sequence.
- the combined stream registration sequence includes three stream registration information in the basic stream selection table and three stream registration information in the extension stream selection table.
- the three stream registration information in the basic stream selection table has stream numbers “1”, “2”, and “3”, and the stream attributes in the three stream registration information are English, Japanese, and Chinese. Language attribute.
- the three stream registration information in the extended stream selection table has stream numbers “1”, “2”, and “3”, respectively.
- the stream attributes in the three stream registration information are English, Japanese, and Chinese. Language attribute.
- the stream identifier in the stream entry is different between the stream registration information in the basic stream selection table and the stream registration information in the extension stream selection table.
- the stream registration information in the extension stream selection table is the left-eye PG stream for the BD presentation mode. Packet identifier for the right eye PG stream.
- FIG. 5B shows the setting of the stream number and the packet identifier when such a combined stream registration sequence is supplied to a playback device whose language setting is Chinese and whose output mode is set to 2D output mode. Output.
- the arrows with a1, a2, and a3 in the figure schematically show language setting match determination, stream number setting in the stream number register, and packet identifier output to the demultiplexing unit.
- FIG. 6C shows the setting of the stream number and the packet identifier when such a combined stream registration sequence is supplied to a playback device whose language setting is Chinese and whose playback type is set to stereoscopic PG. Output.
- the arrows a4, a5, and a6 in the figure schematically indicate language setting match determination, stream number setting for the stream number register, and packet identifier output to the demultiplexing unit.
- FIG. 34 shows what packet identifiers are output to the demultiplexing unit according to the combined stream registration sequence.
- the combined stream registration sequence includes three stream registration information in the basic stream selection table and three stream registration information in the extension stream selection table.
- the three stream registration information in the basic stream selection table has stream numbers “1”, “2”, and “3”, respectively.
- the stream attributes in the three stream registration information are all Chinese language attributes. have.
- Each of the three stream registration information in the extended stream selection table has stream numbers “1”, “2”, and “3”, and the stream attributes in the three stream registration information also have Chinese language attributes. is doing.
- the stream identifier in the stream entry is different between the stream registration information in the basic stream selection table and the stream registration information in the extension stream selection table.
- the stream registration information in the extension stream selection table is the left-eye PG stream for the BD presentation mode. Packet identifier for the right eye PG stream.
- FIG. 5B shows the setting of the stream number and the packet identifier when such a combined stream registration sequence is supplied to a playback device whose language setting is Chinese and whose output mode is set to 2D output mode. Output.
- the arrows with a1, a2, and a3 in the figure schematically indicate language setting match determination, stream number setting, and packet identifier output.
- the packet identifier in the stream entry in the basic stream selection table is output to the demultiplexing unit.
- the TS packet specified by the packet identifier of the stream entry in the stream registration information of the stream number “1” in the basic stream selection table is output to the decoder.
- (C) shows the setting of the stream number and the output of the packet identifier when such a combined stream registration sequence is supplied to a playback device whose language setting is Chinese and whose playback type is set to 1 plane + Offset type. Show.
- the arrows a4, a5, and a6 in the figure schematically indicate language setting match determination, stream number setting, and packet identifier output.
- the stream registration information of the stream number “1” it is determined whether the language setting on the playback device side matches the stream attribute, and the stream number included in the stream registration information of the stream number “1” is the stream number register. Is written to.
- the packet identifier in the stream entry in the basic stream selection table is output to the demultiplexing unit.
- two types of TS packets specified by the set of packet identifiers stored in the stream entry in the stream registration information of the stream number “1” in the extension stream selection table are output to the decoder.
- the video shift mode that defines the securing of the subtitle display area is described in the extended stream selection table in association with the stream number.
- a stream selection procedure is executed at the time of occurrence of a video stream and a new stream number is set in the stream number register, a video shift mode corresponding to the new stream number is provided to the playback apparatus.
- the aspect ratio of cinesco size (1: 2.35) is common, and when storing images on an optical disc such as a BD-ROM, the main image without changing the aspect ratio. Is generally arranged at the center of a 16: 9 HD video, and black frames are inserted at the top and bottom of the screen. Therefore, according to the above configuration, it is possible to display subtitles in a wide subtitle display area in which the black frames at the top and bottom of the main video are gathered at the top or bottom of the video plane, thereby improving screen utilization efficiency. And the stereoscopic effect can be improved.
- Modification As a modification of the present embodiment, when picture data and subtitles are combined, not only the picture data stored in the video plane memory but also the subtitles stored in the PG plane memory are shifted upward or downward. The method of making it explain.
- FIG. 35 shows a stream registration sequence of the extension stream selection table according to this modification.
- FIG. 5B shows the internal structure of the stream registration sequence of the PG stream.
- “PG shift value when moving up video (PG_v_shift_value_for_Up)” and “PG shift value when moving down video (PG_v_shift_value_for_Down)” are added to the stream registration information of the PG stream.
- the “PG shift value when moving up video (PG_v_shift_value_for_Up)” is stored in the PG plane memory when the video shift mode is set to “Up” and the subtitle display area of the PG_text subtitle stream is secured at the lower end of the video plane. This is a shift amount for shifting the stored caption data downward.
- the “PG shift value when moving down video (PG_v_shift_value_for_Down)” is stored in the PG plane memory when the video shift mode is set to “Down” and the subtitle display area of the PG_text subtitle stream is secured at the upper end of the video plane. This is a shift amount for shifting the stored caption data upward.
- the PSR 33 includes a plane shift amount when the video shifts upward and a plane shift amount when the video shifts downward.
- PG_shift_value_for_UP and PG_shift_value_for_Down are prepared for the PG plane.
- PG_v_shift_value_for_Up and PG_v_shift_value_for_Down for the current PG stream selected by stream switching are acquired from the stream registration information in the extension stream selection table and set.
- FIG. 36 shows a circuit configuration for synthesizing the outputs of the decoder model and outputting them in the upper end 2D subtitle playback mode method and the lower end 2D subtitle playback mode method.
- the upper-end 2D subtitle playback mode method and the lower-end 2D subtitle playback mode method only 131 pixels upward or downward depending on the PSR32 setting for each video plane for the left eye and right eye As the pixels are shifted, the left eye output and the right eye output of the PG plane are also shifted downward or upward according to the value of PG_ shift_value_for_Up or PG_ shift_value_for_Down set in PSR33, respectively. It is designed to be synthesized.
- the coordinates of the area are represented by (0, PG_v_shfit_value_for_Down), (0, height + PG_v_sfhit_value_for_Up), (width, PG_v_shfit_value_for_Down), (width, height + PG_v_sfhit_value_for_Up). For example, if PG_v_sfhit_value_for_Up is ⁇ a, for example, PG_v_sfhit_value_for_Down is + b, the regions are (0, b), (0, height-a), (width, b), (width, height-a).
- PG restrictions include that the display position does not exceed the above area, that the size of the object to be displayed at the display position does not exceed the above area, that the window display position does not exceed the above area, The window size plus the display position does not exceed the above area.
- the screen size affects the sense of depth of the 3D video, as shown on the left side of FIG. This is because the difference value between the left-eye video and the right-eye video changes depending on the size of the television screen. For example, as shown on the left side of FIG. 40, when the left-eye video and the right-eye video are created so that the optimum depth can be obtained for 50 inches, optimal viewing is possible with 50 inches, but from 50 inches. In a small television, the difference value between the left-eye video and the right-eye video is small, so the video does not have a powerful depth.
- an offset value for correcting the screen size is applied in the direction to increase the sense of depth in the case of 32 inches, and output to the television in the direction of increasing the sense of depth in the direction of 100 inches. To output to the TV with the offset value.
- applying an offset value to a plane means that the final plane output from the player is shifted in the horizontal direction by the offset value and cropped in the same manner as in the 1 plane + offset method with respect to the plane.
- the offset value applied to the final player output plane is set as an output offset correction value.
- a table as shown in FIG. 41A is stored in files such as an index file, a playlist file, and an AV stream information file.
- files such as an index file, a playlist file, and an AV stream information file.
- a plurality of pieces of screen size information in which the TV inch number and the output offset correction value are paired are registered.
- the number of inches is divided every 10 inches, but how many inches are divided may be determined according to a predetermined standard or the like, or may be set by the user. It should be noted that several tables as shown in FIG. 41A may be prepared according to the standard, and only the reference IDs of the tables may be registered in the index file, playlist file, AV stream information file, and the like.
- a function for determining the output offset value according to the number of inches may be prepared.
- an optimum TV size (assumed_TV_size_when authoring) value indicating how many inches the corresponding content is created as a target may be entered.
- Various correction processes are possible using this value. For example, when displaying on a TV that is larger than the optimal size, As shown in FIG. 42, it is possible to perform processing such as displaying an image with the optimum size in the center of the screen and displaying a black frame in the surrounding area indicated by diagonal lines.
- the playback device includes a system parameter PSR35 in which an output offset correction value is stored, and an output offset correction value application unit.
- PSR 35 the screen size (number of inches) of the television connected via the HDMI cable or the like is acquired by the playback control unit, and the output offset correction value corresponding to this screen size is specified and stored based on the table described in FIG.
- the output offset correction value application unit refers to the value of PSR35 and applies an offset with the value of PSR35 to the plane of the left eye video and the right eye video synthesized by the plane addition unit.
- the output offset correction value is stored in PSR35
- the screen size is stored in PSR35
- the output offset correction value application unit specifies the output offset correction value with reference to the table described in FIG. May be.
- the output offset correction value may be adjusted according to the viewer. For example, in the case of a young child, the difference between the left eye image and the right eye image is preferably small because the distance between the left and right eyes is narrow. Therefore, an “output offset correction value alpha” for correcting the output offset correction value is prepared.
- the output offset correction value application unit executes the offset correction process with a value obtained by multiplying the output offset correction value by the “output offset correction value alpha”. Specifically, it is realized by a configuration as shown in FIG. In FIG. 44, a PSR 36 for storing the value of the output offset correction value alpha is prepared.
- the playback control unit or program execution means sets a value in the PSR 36 through a menu screen, a player OSD screen, or the like.
- the depth when it is desired to weaken the depth for viewing by a child, the depth can be weakened by setting a value larger than 1.
- the output offset correction value application unit applies an offset to the plane by referring to PSR35 and PSR36 and by multiplying the output offset correction value and the output offset correction value alpha. By doing in this way, a feeling of depth can be adjusted according to a user's liking.
- the “output offset correction value alpha” value is configured on the menu screen of the BD program by selecting from the “weak”, “normal”, and “strong” modes for the 3D depth. May be.
- the “output offset correction value alpha” may be stored for each SEI message of the video stream, descriptor of the PMT packet, each play item, etc., and the value may be changed according to the scene. As a result, it is possible to increase the “output offset correction value alpha” in a scene having a large depth.
- the output offset amount correction value is changed according to the screen size of the television.
- the output offset correction value and the output offset correction value alpha may be changed according to the viewing distance.
- the glasses used for 3D viewing may measure the distance from the screen, and the television may acquire the distance and notify the playback device from the television via the HDMI cable.
- the output offset amount correction value is changed in accordance with the television screen size.
- the projector may measure the television size as follows.
- the first method is to output a laser such as infrared rays from the projector to the screen, measure the distance by rebounding from the screen, and calculate the screen size from the optical system parameters of the lens.
- Another method is to display a “line segment” of the size that is a projector, the user measures the length of the line segment projected on the screen, and the length of the measured line segment through the OSD of the projector. You may make it input.
- the projector can calculate the screen size.
- the optimal stereoscopic view corresponding to the screen size is performed by performing the offset process for changing the difference value between the left-eye video and the right-eye video according to the screen size for displaying the video.
- the internal structure of the stereoscopic interleaved stream file is improved.
- a file in the UDF is composed of a plurality of extents managed by file entries.
- the “file entry” includes a “descriptor tag”, an “ICB tag”, and an “allocation descriptor”.
- Descriptor tag is a tag indicating that it is a file entry.
- the tag includes types such as a file entry descriptor and a space bitmap descriptor. In the case of a file entry, “261” indicating a file entry is described as the descriptor tag.
- the “allocation descriptor” includes a logical block number (LBN) that indicates the recording position of the extent that constitutes a lower file under a certain directory.
- the allocation descriptor includes data indicating the extent length and a logical block number indicating the recording position of the extent. However, the upper 2 bits of the data indicating the extent length are set to “0” to indicate that they are allocated and recorded extents, and are set to “1” to indicate allocated and unrecorded extents. It shows that it is. When set to “0”, it indicates that the extent is a continuation of the allocation identifier.
- the file entry has a plurality of allocation descriptors for each extent.
- Stereoscopic interleaved stream file is a stream file (2TS interleaved file) in which 2TS is converted into an interleaved format. It is a 5-digit integer value and an extension indicating that the file is an interleaved format file for stereoscopic playback. identified by (ssif).
- the stereoscopic interleaved stream file is composed of extent SS [n], and extent SS [n] (EXTSS [n]) is specified by index number n.
- the index number n is a number that is incremented one by one from the beginning of the stereoscopic interleaved stream file.
- Extent SS [n] is configured as a combination of a dependent view data block and a base view data block.
- the base view data block and dependent view data block that make up extent SS [n] are subject to cross-reference from file 2D, file base, and file dependent.
- Cross reference refers to registering one data object recorded on a recording medium in a file entry as an extent of a plurality of files.
- the start address and the continuous length of the base view data block and the dependent view data block are registered in the file entry of the file 2D, the file base file entry, and the file dependent file entry. .
- the file base is a virtual stream file that “stores” the main TS indicated by the extent start point information in the clip information corresponding to the file 2D, and includes at least one extent 1 [i ] (Referred to as EXT1 [i]).
- Extent 1 [i] is the i-th extent in the file base, i is the index number of the extent, and is incremented starting from 0 at the beginning of the file base.
- the file base is a virtual stream file for handling a stereoscopic interleaved stream file, which is a 2TS file, as a 1TS file, and is created virtually by constructing the file entry in the memory of the playback device.
- the file base is specified by opening the file using the file name of the stereoscopic interleaved stream file.
- the playback device middleware generates a file entry that identifies the file-based extent in the memory when the file open using the file name of the stereoscopic interleaved stream file is called, and Is virtually opened.
- a stereoscopic interleaved stream file can be regarded as “includes only 1TS”, and a 2TS stereoscopic interleaved stream file can be read from a recording medium as a 1TS file base.
- the BB presentation mode when only the base view data block is to be read, only the extents constituting this file base are read. Even if the mode is changed from the BB presentation mode to the BD presentation mode, the reading range is changed from the extent recording range constituting the file base to the extent recording area constituting the stereoscopic interleaved stream file. If it is enlarged, both the base view data block and the dependent view data block can be read, so that the efficiency of reading the file is not lowered.
- a file dependent is a stream file that “stores” a sub-TS, and is composed of extents 2 [i] (EXT2 [i]).
- EXT2 [i] is the i-th extent in the file dependent, i is the index number of the extent, and is incremented starting from 0 at the beginning of the file dependent.
- a file dependent is a virtual stream file for handling a stereoscopic interleaved stream file, which is a 2TS file, as a 1TS file containing a sub-TS, and the file entry is constructed in the memory of the playback device. It is generated virtually.
- Dependent video stream is given a file name that is a 5-digit number that is the file name of the stereoscopic interleaved stream file plus one. Accessed using this file name.
- a dummy file is recorded on the recording medium, and the identification number of the dependent-view video stream, “number added by 1”, is assigned to the dummy file.
- a dummy file is a file that has only a file name and no actual extent, and the dependent-view video stream is treated as being stored in this dummy file.
- the file 2D is a 1TS stream file storing the main TS that is played back in the 2D output mode, and is composed of extents 2D.
- the file 2D is identified by a 5-digit integer value and an extension (ssif) indicating that the file is an interleaved format file for stereoscopic playback.
- FIG. 45 shows the correspondence between extents, file 2D / file base, and file dependent.
- the first row shows file 2D / file base, file dependent 00001.m2ts, 00002.m2ts, and the second row shows the extent and dependent view data block storing the base view data block. Indicates the stored extent.
- the third row shows 00001.ssif, which is a stereoscopic interleaved stream file.
- extents EXT1 [i], EXT2 [i], EXT1 [i + 1], and EXT2 [i + 1] are registered as extents of 00001.ssif. You can see that As described above, it is understood that the extents EXT1 [i] and EXT1 [i + 1] have the duality that they belong to 00001.ssif and at the same time belong to 00001.m2ts.
- the extension “ssif” is an acronym for StereoScopic Interleave File, and indicates that it is in an interleaved format for stereoscopic playback.
- an interleaved extent unit a combination of extents constituting a file base and extents constituting a file dependent, which are specified by the same extent identifier, is referred to as an “interleaved extent unit”.
- EXT1 [ A pair of i + 1] and EXT2 [i + 1] is an interleaved extent unit [i + 1].
- it In random access to the stereoscopic interleaved stream file, it must be ensured that the interleaved extent unit specified by the extent identifier is read from the recording medium at a time.
- FIG. 46 shows the relationship between the stereoscopic interleaved stream file and the file 2D / file base.
- the third tier in Fig. 3 (a) shows the internal structure of the interleaved stream file.
- Stereoscopic interleaved stream files are the extents EXT1 [1] and EXT1 [2] that store the base view data block, and the extents EXT2 [1] and EXT2 [2] that store the dependent view data block. Are arranged alternately.
- the first row shows the internal structure of file 2D and file base.
- the file 2D / file base is composed of only the extents EXT1 [1] and EXT1 [2] storing the base view data block among the extents constituting the interleaved stream file in the third stage.
- the file name of the file 2D is the same as that of the interleaved stream file, but the extension is different.
- the second tier shows the internal structure of the Dependent.
- Fi dependent consists only of the extents EXT2 [1], EXT2 [2], and EXT2 [2] that store the dependent view data block among the extents constituting the interleaved stream file in the third stage. ing.
- the file dependent file name is obtained by adding 1 to the file name of the interleaved stream file, and the extension is different.
- the 2D playback device is a discriminating method that is the same as the conventional 2D playback method, and it is necessary to be able to access only the 2D playlist and 2D stream, so the file format that the 2D playback device can recognize for the base view video stream Is stored in.
- the first method is to refer to the playlist information as described above, that is, the main TS uses the same file name as the 2D playback method so that it can also be used for 2D playback, and the interleaved stream file changes the extension. It is.
- 00001.m2ts and 00001.ssif are coupled by the same file name “00001” while one is a 2D system and the other is a 3D system.
- the existing 2D playback device plays only the file 2D.
- the 3D-compatible playback device only refers to the file 2D containing the main TS in the playlist, but if there is a file with the same identification number but a different extension, the file is found and used for 3D video.
- the main TS and the sub TS are output.
- the second method is to divide folders.
- the main TS is stored in a folder with an existing folder name (eg STREAM), but the sub TS is stored in the folder with a 3D-specific name (eg SSIF) with the same file name “00001”. deep.
- a playlist refers to a file
- the 2D playback device refers only to the file in the “STREAM” folder
- the 3D playback device refers to a file with the same name from the “STREAM” and “SSIF” folders at the same time.
- the third method is based on an identification number.
- the file dependent identification number is obtained by adding “1” to the file 2D identification number, as shown in FIG.
- association is performed according to a certain rule, such as assigning an identification number “0002”.
- the file dependent named according to the above rule is treated as a dummy file having no substance. This is because the substance of the file dependent is only a stereoscopic interleaved stream file.
- the file name thus associated is described in the stream registration information in the basic stream selection table and the sub clip entry ID reference (ref_to_subclip_entry_id) of the stream entry in the stream registration information in the extension stream selection table.
- the playback device authenticates that the file name of the identification number obtained by adding “1” to the identification number described in the sub clip entry ID reference is the file name of the dummy file, and virtualizes the file dependent. The process to open automatically. This ensures that the file dependent associated as described above is read from the recording medium in the stream selection procedure.
- the clip information file is also identified by the same rule.
- the base view data block (B [i]) is the i-th data block of the main TS.
- the main TS is a TS specified as the main path base axis through the clip information file name information (clip information file name information) of the current play item information.
- “I” of B [i] is an index number that is incremented by setting the first data block of the file base to 0.
- the base view data block shared by the file 2D and file base, and the base view data block specific to the file 2D are extents of the file 2D, and set to a length that does not cause buffer underflow in the playback device. Has been.
- the head sector address is described in the allocation descriptor in the file entry of the file 2D.
- the base sector data block specific to the file base has its head sector address not described in the allocation descriptor in the file entry. Instead, the source packet of the head source packet in the base view data block is pointed by the extent start point information in the clip information of the clip information file corresponding to the main TS. Therefore, it is necessary to derive the head sector address of the base view data block specific to the file base using the allocation descriptor in the file entry of the stereoscopic interleaved stream file and the extent start point information in the clip information.
- the dependent view data block (D [i]) is the i-th data block of the sub-TS.
- the sub-TS is a TS that is designated as a sub-path base in the stream entry in the stream registration sequence of the extended stream selection table corresponding to the current playitem information.
- “I” in D [i] is an index number that is incremented with the first data block of the file dependent as 0.
- Dependent view data block is a file dependent extent and is set to a length that does not cause double buffer underflow in the playback device.
- the dependent view data block is arranged before the base view data block to be reproduced at the same reproduction time. Therefore, when the stereoscopic interleaved stream file is read, the dependent view data block is always read before the base view data block.
- Dependent view data block is not shared with file 2D, so the start sector address is not described in the allocation descriptor in the file entry of file 2D. Instead, the source packet of the head source packet in the dependent view data block is pointed by the extent start point information in the clip information. Therefore, it is necessary to derive the start sector address of the dependent view data block using the allocation descriptor in the file entry of the file 2D and the extent start point information in the clip information.
- the extent of the file 2D may be common with the file-based extent or may not be common with the file base.
- File 2D extent consists of B [0], B [1], B [2], B [3] 2D, B [4] 2D, and file-based extents are B [0], B [1] , B [2], B [3] ss, B [4] ss.
- B [0], B [1], and B [2] are base view data blocks that are shared with the file base.
- B [3] 2D and B [4] 2D are file 2D-specific base view data blocks that are not shared with the file base.
- B [3] ss and B [4] ss are base view data blocks specific to the file base that are not shared with the file 2D.
- the data in B [3] 2D and the data in B [3] ss have bit-for-bit identity.
- the data in B [4] 2D and the data in B [4] ss have bit-for-bit identity.
- the data blocks B [2], B [3] 2D, and B [4] 2D in these files 2D constitute an extent (big extent) having a large continuous length immediately before the place where the long jump occurs.
- File 2D can form a big extent immediately before a long jump, so there is no need to worry about underflow of the read buffer even when playing a stereoscopic interleaved stream file in 2D output mode. .
- File 2D and file base are identical, although their extents are partially different. These files 2D and file base are collectively referred to as “file 2D / file base”.
- FIG. 47 shows the interrelationship between the stereoscopic interleaved stream file, file 2D, and file base.
- the first row shows a file 2D
- the second row is a data block on the recording medium
- the third row is a stereoscopic interleaved stream file
- the fourth row is a file base
- the fifth row is a file dependent. Indicates a dent.
- the data blocks in the second stage are the above-described D [1], B [1], D [2], B [2], D [3], B [3] ss, D [4], B [4 ] ss, B [3] 2D, B [4] 2D.
- the arrows ex1, ex2, ex3, and ex4 represent the belonging relationship that B [1], B [2], B [3] 2D, and B [4] 2D of the data blocks constitute the extent of the file 2D. Indicates.
- arrows ex5 and ex6 indicate D [1], B [1], D [2], B [2], D [3], B [3] ss, D [4], B [4 among the data blocks.
- ] ss indicates the belonging relationship that forms an extent of a stereoscopic interleaved stream file.
- B [1], B [2], B [3] ss, and B [4] ss among the data blocks constituting this stereoscopic interleaved stream file are file-based extents.
- the fifth row indicates that D [1], D [2], D [3], and D [4] among the data blocks constituting the stereoscopic interleaved stream file are file dependent extents. .
- FIG. 48 shows a 2D playlist and a 3D playlist.
- the first level is 2D playlist information
- the second level is a bases data block
- the third level is a 3D playlist
- the fourth level is a dependent view data block.
- Arrows rf1, rf2, and rf3 indicate playback paths by combining the file name 00001 described in clip_information_file_name in the playitem information of the 2D playlist information and the extension m2ts.
- a playback path on the base view side is configured by the data blocks B [1], B [2], and B [3] 2D.
- Arrows rf4, rf5, rf6, and rf7 indicate playback paths specified by the play item information of the 3D playlist information.
- a playback path on the base view side is configured using B [1], B [2], B [3] ss, and B [4] ss.
- Arrows rf8, rf9, rf10, and rf11 indicate playback paths specified by the sub play item information of the 3D playlist information.
- the dependent view side playback path is configured using D [1], D [2], D [3], and D [4].
- the data block constituting the playback path specified by these play item information and sub play item information is obtained by combining the file name described in clip_information_file_name in the play item information and the extension ssif to open the file. Can be read.
- the playlist information that defines these 3D playlist and 2D playlist The common description is sufficient to describe (see symbols df1 and df2). Therefore, if playlist information that realizes this 3D playlist is described, it functions as a 3D playlist when the output mode of the playback device is the stereoscopic output mode, and the output mode of the playback device is the 2D output mode. When it is, it will function as a 2D playlist.
- the 2D playlist and 3D playlist in this figure are interpreted as 2D playlists and 3D playlists according to the output mode of the playback device that interprets one playlist information. It is possible to reduce the time and effort of the person who performs the authoring.
- clip_information_file_name in the play item information of the 2D playlist describes the file name of the file 2D.
- Clip_information_file_name in the play item information of the 3D playlist describes a file base file name. Since the file base is a virtual file and the file name is the same as that of the stereoscopic interleaved stream file, the file name of the stereoscopic interleaved stream file may be described in clip_information_file_name in the play item information.
- Ref_to_subclip_entry_id in the stream registration information of the extension stream selection table describes the file name of the file dependent. The file name of the file dependent is obtained by adding 1 to the identification number of the stereoscopic interleaved stream file.
- the stereoscopic interleaved stream file is converted into the file 2D, the file base, and the file. Since the file can be opened as any of the dependent files, the stereoscopic interleaved stream file can be handled in the same manner as a normal stream file on the decoder side. Therefore, it is possible to positively incorporate a stereoscopic interleaved stream file into the storage system of the base-view video stream and the dependent-view video stream.
- FIG. 49 shows the internal structure of the clip information file.
- (A) shows a 2D clip information file
- (b) shows a 3D clip information file.
- These clip information files include “clip information”, “sequence information”, “program information”, and “feature point information”.
- “Clip information” indicates for each ATC sequence what kind of AV clip each source packet sequence stored in the stream file is.
- Sequence information is configured to indicate, for each ATC sequence, information (ATC sequence information) indicating what ATC sequence is one or more source packet sequences stored in the stream file.
- the ATC sequence information includes information indicating the source packet number where the source packet that is the starting point of the ATC exists, an offset between the STC sequence identifier and the ATC sequence identifier, and STC sequence information for each of a plurality of STC sequences. Including.
- the STC sequence information includes the packet number of the source packet storing the PCR in the STC sequence, information indicating where in the ATC sequence the source packet that is the start point of the STC sequence exists, and the playback start time in the STC sequence , Including playback end time.
- Program information is information indicating the program configuration of the main TS and sub-TS managed as “AV clips” by the clip information file, and what ES is multiplexed with the AV clip. Indicates. Specifically, it indicates what packet identifier the ES multiplexed in the AV clip has and what encoding method it uses. It is clearly shown in this program information which encoding method is used to compress and encode the video stream among MPEG2-video, MPEG4-AVC and the like.
- Feature point information is information indicating, for each ES, where a plurality of ES feature points multiplexed in an AV clip exist. Information indicating feature points for each ES is called a basic entry map.
- an access unit delimiter indicating the head of an I picture type view component located at the head of an open GOP or a closed GOP is a feature point.
- an access unit delimiter that indicates the start position of an audio frame that exists every other period, such as every second, is a feature point.
- An access unit delimiter that indicates the head position of a display that has all the functional segments necessary for display is a feature point.
- a feature point in an ATC sequence is represented by a source packet number.
- the same feature point is expressed using PTS indicating the time point on the STC time axis.
- each entry point that makes up the entry map has a source packet number that indicates the location of the feature point in the ATC sequence and the PTS that indicates the location of the feature point in the STC sequence. It includes a flag (is_angle_change flag) indicating whether or not the angle can be switched to a point, and information (I_size) indicating the size of the intra picture existing at the head of the GOP. Since the source packet located at the head of the interleave unit constituting the multi-angle section can be switched in angle, the is_angle_change flag of the entry point that points to the head source packet of the interleave unit is always set to ON. In addition, the entry point to which the head source packet of the interleave unit is inserted is associated with In_Time in the play item information by the entry point.
- the entry map for each ES shows the source packet number of the feature point for each stream type in association with the PTS, so by referring to this entry map, from any point in the STC sequence to that point A source packet number indicating the location of the feature point for each closest ES can be derived.
- the clip information file for 3D has the internal configuration shown in FIG. 49B.
- clip information for file 2D which is normal clip information (management information)
- a clip for file dependent Information “clip dependent information (base view management information)”
- file base clip information “clip base information (base view management information)”
- clip dependent information and clip base information are stored in a clip information file corresponding to the file 2D.
- the difference between 2D clip information, clip base information, and clip dependent information is that there is metadata including an extent start point sequence inside the clip base information and clip dependent information.
- clip dependent information includes an extent start point sequence
- clip base information also includes an extent start point sequence
- the feature point information includes an extended entry map
- the extension data includes an extended entry map
- the clip information file is divided into a clip base information file and a clip dependent information file.
- (C) in the figure shows a clip base information file generated based on the clip base information file. It consists of clip base information and basic entry map.
- the clip base information includes extent start point information.
- FIG. 4D shows that the clip dependent information file includes clip dependent information and an extended entry map.
- the clip dependent information includes extent start point information.
- the clip information file in the 2D output mode exists in the clip information file directory (CLPI directory).
- CLPI directory The clip base information file is generated from the clip information file in the 3D output mode and is handled as being stored in the clip information file in the 2D output mode.
- This dummy clip information file has a number corresponding to the file dependent, that is, a file 2D / file base. A file name having a number obtained by adding “1” to the identification number is given.
- the clip dependent information file is generated from the clip information file corresponding to the file 2D in the 3D output mode, and is handled as being stored in the dummy clip information file.
- the clip information file in 2D output mode is 00001.clpi
- the clip base information file is treated as being stored in 00001.clpi in 3D output mode, and clip dependent information in 3D output mode.
- the file is handled as being stored in 00002clpi.
- the stereoscopic interleaved stream file is composed of two clip AV streams (BDAV MPEG2 transport stream).
- the pair of extent start point information provides a way to divide a stereoscopic interleaved stream file into two AV streams.
- Extent start point information is supplied as follows.
- the extent start point information table pair is The clip information referenced by the angle ID value and the clip information referenced by the sub clip entry ID value are stored and supplied to the playback device.
- the ID1 value and ID2 value in the extension data in ext_data_entry () must be set to 0x0002 and 0x0004, respectively.
- the clip information file with extent start point information must satisfy the following two conditions.
- FIG. 50 shows the relationship among a clip information file, a playlist, and a stereoscopic interleaved stream file.
- the right side shows a stereoscopic interleaved stream file
- the left side shows a clip information file.
- the middle shows the file base in the first row, the clip information file in the second row, the 3D playlist in the third row, the clip dependent information file in the fourth row, and the file dependency in the fifth row. Indicates a dent.
- Arrows bk1 and bk2 indicate that the file base and file dependent can be obtained by dividing the stream file on the right side.
- the clip information file on the left side includes feature point information, extension data, clip base information, and clip dependent information.
- Arrows bk3 and bk4 schematically indicate that the extent start point information in the clip base information and the extent start point information in the clip dependent information provide a way to divide the stereoscopic interleaved stream file.
- the ID1 value and ID2 value in the extension data in ext_data_entry () must be set to 0x0002 and 0x0004, respectively.
- the clip information file with extent start point information must satisfy the following two conditions.
- the stereoscopic interleaved stream file is composed of two clip AV streams (BDAV MPEG2 transport stream).
- the pair of extent start point information provides a way to divide a stereoscopic interleaved stream file into two AV streams.
- Extent start point information is supplied as follows.
- the extent start point information table pair includes the clip information referenced by the angle ID value and the clip information referenced by the sub clip entry ID value, respectively. It is stored and supplied to the playback device.
- FIG. 51 shows the internal structure of clip base information and clip dependent information.
- the clip base information and the clip dependent information are configured by “clip stream type information” indicating what type of stream the corresponding AV clip belongs to and the corresponding AV clip.
- “Application type” indicating what type of application is a movie application, time-based slide show application, browsable slide show application, etc., and after the source packet passes through the source packet depacketizer in the playback device , “TS recording rate” indicating how much TS packets are transferred in the AV clip, “number of source packets” indicating the number of source packets constituting the corresponding AV clip, and the previous AV clip Structure Composed from the "ATC difference value”, which is the ATC of the difference between the ATC sequence that.
- (B) shows the internal structure of the extent start point information table.
- the extent start point information table is composed of number_pf_extent_start_points and number_pf_extent_start_points SPN_start_points.
- “Number_pf_extent_start_points” indicates the number of extents belonging to the related AV stream file.
- a pair of extent start point information in the clip base information and extent start point information in the clip dependent information has the same value of number_pf_extent_start_points.
- SPN_extent_start (SPN_extent_start [0] to SPN_extent_start [number_of_extent_start_point]) is composed of number_of_extent_start_point + 1 SPN_extent_start.
- SPN_extent_start is indicated by an extent identifier [extent_id], and is a 32-bit value indicating the source packet number of the source packet from which the extent_id-th extent in the AV stream file starts.
- An extension entry map exists in the extension data. Similar to the basic entry map, the extended entry map is composed of a plurality of entry points. More specifically, each entry point constituting the extended entry map has a source packet number indicating the location of the feature point in the ATC sequence associated with the PTS indicating the location of the feature point in the STC sequence. It includes a flag (is_angle_change flag) indicating whether or not the angle can be switched to the feature point, and information (I_size) indicating the size of the intra picture existing at the head of the GOP.
- the extended entry map is different in that the following restrictions are added to the extended entry map.
- FIG. 52 is a diagram showing a basic entry map and an extended entry map.
- the fifth row shows a plurality of sets of dependent view data blocks and base view data blocks.
- the fourth level shows a source packet sequence constituting the dependent view data block and the base view data block.
- the first level shows view components specified by the PTS.
- the second row shows a basic entry map, and the third row shows an extended entry map.
- the application type 8
- the source packet located at the head of GOP (i) of the base-view video stream and the source packet located at the head of GOP (i) of the dependent-view video stream are in the same interleave extent unit. If it belongs, the entry pointing to the GOP (i) head source packet of the base-view video stream and the GOP (i) head source packet of the dependent-view video stream is in both the basic entry map and the extended entry map. Added. Therefore, in this case, if both the basic entry map and the extended entry map are used, continuous reading of the GOP (i) of the base-view video stream and the GOP (i) of the dependent-view video stream can be ensured.
- FIG. 53 shows entries that are not allowed in the extended entry map.
- GOP ( i) Do not add an entry pointing to the first source packet to either the basic entry map or the extended entry map. In this case, since the GOP (i) of the base-view video stream and the GOP (i) of the dependent-view video stream are excluded from the random access destinations, it is possible to prevent a decrease in access performance.
- FIG. 54 shows a play item playback procedure
- Step S201 is a determination as to whether or not the current output mode is the 3D output mode. If the current output mode is the 2D output mode, steps S203 to S206 are executed.
- step S203 the stream file specified by “XXXXX” described in Clip_Information_file_name of the current play item and the extension “m2ts” is opened.
- step S204 the entry point corresponding to the packet ID of the video stream is opened. Is used to convert the current PlayItem.In_Time and the current PlayItem.Out_Time into Start_SPN [i] and End_SPN [i].
- step S205 the extent belonging to the read range [i] for reading the TS packet [i] of the packet ID [i] from Start_SPN [i] to End_SPN [i] is specified, and in step S206, the read range [i] The drive of the recording medium is instructed to continuously read the extents belonging to.
- step S300 If the current output mode is the stereoscopic output mode, the process proceeds to a loop of step S300 to step S60.
- step S300 the stream file specified by “XXXXX” described in Clip_Information_file_name of the current play item and the extension “ssif” is opened, and in step S301, according to the base view indicator of the current play item information,
- the base view video stream is assigned to one of the left view video plane and the right view video plane
- the dependent view video stream is assigned to the other of the left view video plane and the right view video plane.
- step S302 the current PlayItem.In_Time and the current PlayItem.Out_Time are converted into Start_SPN [i] and End_SPN [i] using the basic entry map corresponding to the base view video stream.
- step S303 SubPlayItemIn_Time and SubPlayItemOut_Time are converted to Start_SPN [j] and End_SPN [j] using the extended entry map corresponding to the dependent-view video stream (step S304).
- step S305 The extent belonging to the read range [i] for reading the TS packet [i] constituting the base-view video stream from Start_SPN [i] to End_SPN [i] is specified (step S305), and the TS packet with the packet ID [j]
- step S306 The extent belonging to the read range for reading [j] from Start_SPN [j] to End_SPN [j] is specified (step S306).
- step S307 the extents belonging to the reading ranges [i] and [j] are sorted in ascending order of the addresses, and the extents belonging to the reading ranges [i] and [j] are continuously used by using the addresses sorted in step S308. Instruct the drive to read automatically. After that, when the source packet sequence is read out, in step S309, the base view ATC sequence and the dependent view ATC sequence are restored and sent to the base view PID filter and the dependent view PID filter.
- the entry in the extended entry map is dependent-view picture data, Only the base view picture data to be reproduced at the same reproduction time is pointed to that pointed by the entry of the basic entry map.
- the extent is determined using the basic entry map and extended entry map as a clue. If you access, you can play the base view GOP and the dependent view GOP together. Thereby, the delay of the reproduction start can be eliminated.
- FIG. 88A it may be specified that one or more entry points exist in each extent. By defining in this way, it is possible to prevent the interval from the entry point to the next entry point from becoming long as shown in FIG. 88B, and to suppress the amount of delay in processing such as jumping reproduction.
- the present embodiment relates to an improvement for restoring an ATC sequence from data blocks constituting a stereoscopic interleaved stream file.
- FIG. 55 shows how the ATC sequence is restored from the data blocks constituting the stereoscopic interleaved stream file.
- the fourth row shows a plurality of data blocks constituting a stereoscopic interleaved stream file
- the third row shows a source packet sequence multiplexed into the main TS and sub-TS.
- the second row shows the STC sequence 2 that constitutes the dependent view, the entry map, and the ATC sequence 2 that constitutes the dependent view.
- the first row shows the STC sequence 1 that constitutes the base view.
- a set of ATC sequence 1 composing the entry map and base view is shown.
- the arrows from the 3rd level to the 2nd level and the 1st level indicate the ATC sequence 1 and ATC from the data blocks of the two TSs (main TS and sub-TS) interleaved in the stereoscopic interleaved stream file. It shows schematically that sequence 2 is restored.
- These ATC sequences correspond to the STC sequences by the entry map in the clip information.
- the reading unit of the playback device in the present embodiment is configured to accept source packet input from two recording media, and is input from two drives for accessing each of the two recording media. 2 read buffers for temporarily storing source packets and outputting them to the decoder.
- An ATC sequence restoration unit exists between the two drives and the two read buffers. This ATC sequence restoration unit separates the ATC sequence that constitutes the base view and the ATC sequence that constitutes the dependent view stream from the source packet in the interleaved stream file read from one recording medium, It writes to each of the two read buffers. By doing so, the playback apparatus can process the ATC sequence constituting the base-view video stream and the ATC sequence constituting the dependent-view stream as if they were read from different recording media.
- FIG. 56 is a diagram showing how the ATC sequence is restored.
- FIG. 4A shows the internal configuration of the reading unit provided with the ATC sequence restoration unit. As described above, the ATC sequence restoration unit is interposed between the two drives and the two read buffers.
- the arrow B0 in the figure symbolically shows the source packet input from one drive, the arrow B1 writes the ATC sequence 1 constituting the base view video stream, and the arrow D1 shows the dependent view The writing of the ATC sequence 2 constituting the stream is schematically shown.
- FIG. 56 (b) shows how two ATC sequences obtained by the ATC sequence restoration unit are handled.
- a PID filter existing in the demultiplexing unit is shown in the middle of the figure.
- the left side shows two ATC sequences obtained by the ATC sequence restoration unit.
- the right side shows the base-view video stream, dependent-view video stream, left-eye PG stream, right-eye PG stream, base-view IG stream, and dependent-view IG stream obtained by demultiplexing these two ATC sequences. .
- FIG. 57 shows an example of extent start point information in the base view clip information and an example of extent start point information in the dependent view clip information.
- (A) shows extent start point information of base-view clip information and extent start point information of dependent-view clip information. *
- (B) is a base view data block B [0], B [1], B [2]... B [n] constituting the ATC sequence 1, and a dependent view data block constituting the ATC sequence 2.
- (C) shows the number of source packets of the dependent view data block and the number of source packets of the base view data block.
- (D) indicates a plurality of data blocks included in the stereoscopic interleaved stream file.
- the ATC sequence 2 is composed of dependent view data blocks D [0], D [1], D [2]... D [n]
- 0, b1, b2, b3, b4 which are the relative source packet numbers of the dependent view data blocks D [0], D [1], D [2], ... D [n] .. bn is described in SPN_extent_start of the extent start point information of the file dependent.
- ATC sequence 1 is composed of base view data blocks B [0], B [1], B [2]... B [n], 0 is the number of relative source packets of the base view data block.
- a1, a2, a3, a4... an are described in SPN_extent_start of the file-based extent start point information.
- FIG. 4C shows the number of source packets of an arbitrary dependent view data block D [x] and an arbitrary base view data block b [x] in the stereoscopic interleaved stream file.
- the head source packet number of the base view data block B [x] is ax and the head source packet number of the base view data block B [x + 1] is ax + 1, B [n] is formed.
- the number of source packets is ax + 1 ⁇ ax.
- the source packets that constitute B [n] The number is number_of_source_packet1-an.
- leading source packet number of the dependent view data block and the leading source packet number of the base view data block are as shown in (d).
- the first SPN of D [0] is “0”, and the first SPN of B [0] is “b1”.
- the leading SPN of D [1] is “b1 + a1” because it is the sum of the source packet number b1 of the preceding D [0] and the source packet number a1 of B [0].
- the source packet number b1 of the preceding D [0] the source packet number a1 of the B [0] the source packet number b2-b1 of the preceding D [1]
- FIG. 58 is a diagram for explaining the source packet number of an arbitrary data block in the ATC sequences 1 and 2.
- the source packet number in the stereoscopic interleaved stream file of D [x] existing in the source packet number of bx is obtained in the ATC sequence 2 of FIG.
- the head source packet number of D [x] is D [0], B [0], D [1], B [1], D [2], B [2], ... D [x -1] and B [x-1]
- the total number of source packets is “bx + ax” as shown in FIG.
- the head source packet number of B [x] is D [0], B [0], D [1], B [1], D [2], B [2]... Since it is the sum of the source packet numbers of the relative source packet numbers of D [x ⁇ 1], B [x ⁇ 1], and D [x], “bx + 1 + ax”.
- FIG. 5C shows a file base having the base view data block as an extent and a file dependent having the dependent view data block as an extent.
- the first LBN and continuous length of EXT1 [x], which is the file-based extent corresponding to B [x], and the first LBN and continuous length of EXT2 [x], which is the extent of the file dependent corresponding to D [x], are as follows: Asking.
- the LBN of the extent of the stereoscopic interleaved stream file closest to these LBNs is calculated by using the LBN obtained by the above conversion for file_offset which is an argument of the function SSIF_LBN (file_offset).
- the function SSIF_LBN is a function that returns the LBN corresponding to file_offset by following the SSIF allocation descriptor from file_offset.
- the first LBN of EXT2 [x] is SSIF_LBN ((bx + ax) * 192/2048), and the first LBN of EXT1 [x] is SSIF_LBN ((bx + 1 + ax) * 192/2048). .
- EXT2 [x] is (SSIF_LBN ((bx + 1 + ax) * 192/2048) ⁇ SSIF_LBN ((bx + ax) * 192/2048)).
- the continuous length of EXT1 [x] is (SSIF_LBN ((bx + 1 + ax + 1) * 192/2048) ⁇ SSIF_LBN ((bx + 1 + ax) * 192/2048)). If a file entry indicating the head LBN and the continuous length is generated on the memory, a file base and a file dependent can be virtually obtained.
- This ATC sequence restoration unit is realized by creating a program that causes a hardware resource to execute the processing of FIG.
- FIG. 59 shows an ATC sequence restoration procedure.
- step S91 the ATC sequence for base view is ATC sequence 1
- the ATC sequence for dependent view is ATC sequence 2.
- the variable x is initialized to 1. This variable x indicates the dependent view data block and the base view data block. Thereafter, the loop from step S94 to step S96 is repeated.
- step S93 It is determined whether or not the source packet number bx indicated by the variable x is the source packet number bn indicated by the last numerical value n of the base view data block (step S93). From the source packet (bx + ax) indicated by the number bx + ax to the packet immediately before the source packet (bx + 1 + ax) indicated by bx + 1 + ax is added to the ATC sequence 2 (step S94), and the source packet (bx + 1 + ax) ) To the packet immediately before the source packet (bx + 1 + ax + 1) is added to the ATC sequence 1 (step S95), and the variable x is incremented (step S96). Repeat until.
- step S93 If the determination in step S93 is Yes, (number_of_source_packet2-bn) source packets from the source packet number bn are added to the ATC sequence 2 (step S97), and (number_of_source_packet1-an) source packets from the source packet number an Is added to the ATC sequence 1 (step S98).
- Step S99 a file entry indicating the first LBN and the continuous length of the base view data block in the number of sectors is generated on the memory, and the file base is virtually opened.
- Step S100 a file entry indicating the first LBN and the continuous length of the dependent view data block by the number of sectors is generated on the memory, and the file dependent is virtually opened.
- Sector search is a process in which when random access is performed from an arbitrary time point, the source packet number of the source packet corresponding to that point is specified, and the file is read from the sector including the source packet of that source packet number. It is.
- Stereoscopic interleaved stream files have a large extent, so the search range for sector search is wide, and when random access is commanded from an arbitrary point in time, a considerable amount of processing time is required to specify the sector to be read. It may take.
- interleave stream file consists of a data block that makes up the base-view video stream and a data block that makes up the dependent-view stream. The reason is that this allocation descriptor only indicates the start address of the long extent.
- the file base is composed of a plurality of extents with a short length, and the start address of each extent is indicated in the allocation descriptor.
- the identification of the sector to be read is completed in a short time.
- the data blocks that make up the base-view video stream are managed as file-based extents, and the start address of the data block is specified in the allocation descriptor in the file entry corresponding to the file base. If the sector search is started from the head address of the extent including the position, it is possible to reach the sector including the source packet that becomes the random access position at an early stage.
- the data blocks constituting the base-view video stream are managed as file-based extents, and the start address and continuous length of each extent are indicated in the allocation descriptor of the file entry for the file base. Random access from any point in time in the view video stream becomes faster.
- the specific sector search procedure is as follows. By using an entry map corresponding to the base-view video stream, a source packet number that is a random access position corresponding to an arbitrary time point is derived.
- extent start pointing information in the clip information corresponding to the base-view video stream it is specified which extent includes the source packet number serving as a random access position.
- the allocation descriptor of the file entry corresponding to the file base it is possible to specify the head sector address of the extent that includes the source packet number serving as the random access position.
- a file pointer is set at the head sector address, file read is performed, and packet analysis is performed on the read source packet, thereby specifying a source packet having a source packet number serving as a random access position. Then, the source packet of the specified source packet number is read out. As a result, random access to the main TS is efficiently performed. The same applies to the sub-TS.
- the base-view video stream extent and the dependent-view video stream extent in the interleaved stream file are aligned based on the extent start point information and multiplexed. Since the data is supplied to the separation unit and the decoder, the decoder and the program are assumed to have two files, that is, a file dependent that stores a file-based dependent-view video stream that stores a base-view video stream, on a recording medium. Can be handled.
- the base-view video stream and the dependent-view video stream for stereoscopic viewing are recorded as an interleaved stream file, the base-view video stream and the dependent-view video stream can be accessed independently. The processing efficiency can be improved.
- 3D display / glasses system identification will be described with reference to FIG. If multiple TVs are connected to a player capable of 3D display, it is necessary to specify whether each TV requires glasses for 3D viewing, and if so, what type of glasses is required. It is desirable to notify the player via the I / F. For example, in a BD player, if the connected TV can know that glasses with active shutter are required for 3D viewing, a message indicating that glasses with active shutter are required is notified to the viewer before 3D playback starts. Programming is also possible.
- I / F such as HDMI requires whether or not glasses are required when the TV displays in 3D, and if so, what type of glasses is required (anaglyph, circular deflection, active). It is desirable that the information can be shared between devices connected by E-EDID, InfoFrame, etc. so that the player can be notified of information for identifying the shutter. If there is a communication function between the TV and the glasses, the position information of the glasses (specifically, the vertical line extending from the center of the TV screen and the horizontal / vertical angle and distance of the glasses) is returned to the player, so that each 3D image can be displayed. It is also possible to change appropriately for each user.
- TV-1 is a 2D-only TV and TV-2 is a 3D-compatible TV
- the right-eye video or left-eye video output to TV-2 is extracted and output as the output for TV-1.
- FIG. 61 when switching to 2D video during playback of 3D video, it is preferable to output one video in duplicate while maintaining the frame rate of 3D video. This is because when the frame rate changes, a delay such as the need for HDMI authentication occurs.
- the TV is a video with increased brightness in consideration of the fact that the user's field of vision is darkened by glasses, and there is a possibility that appropriate processing for displaying 2D video has not been performed.
- the stream is recorded so that the same subtitle (the subtitle pattern and the display position differ between C, L, and R) is displayed at the same display time between C, L, and R. It is desirable.
- the same menu operation (C, L, R, R, R, etc.) with the same menu structure (C, L, R) It is desirable that the stream is recorded so that only the symbol of the button and the display position are different between R). This is necessary so that 2D viewers on TV-1 and 3D viewers on TV-2 can see the same subtitles and menus and control the menu from either TV.
- simultaneous 2D / 3D output the player may synthesize L, R, and C separately as shown in FIG.
- the video of L2 may contain video with higher relevance in the video R1, R2, R3 on the opposite side than L1, L3, etc. Therefore, even in the case of 2D display, it is possible to improve the image quality better by referring to the video that is not displayed.
- the method of improving the image quality is not questioned here, but it is highly accurate to use the one-eye video (after all binocular image) that is not displayed in the image quality improvement processing while displaying 2D images. It is extremely important for conversion.
- 2D display is provided to the viewer, but the player and the TV need to be connected as 3D. Even an I / F such as HDMI requires connection authentication as 3D, and it is necessary to perform switching control so that the player outputs L / R binocular video.
- the parallax image for the right eye and left eye is larger than the binocular interval of the viewer, the stereoscopic image is formed.
- the right eye / left eye overall image is shifted left and right so that the right eye / left eye parallax image does not shift more than the smallest eyeglasses used, that is, the smallest eye space between viewers.
- Display is desirable as a TV display process.
- the player can input or select the viewer's age, stereoscopic strength, etc. so that the appropriate binocular parallax can be displayed.
- the output processing may be performed after shifting.
- an AV file that is a stream file and a non-AV file that is a file other than a stream file are created in real time, and are directly written in the AV data recording area and the non-AV data recording area on the recording medium.
- the recording medium according to the present embodiment is also specified by a recording method using real-time recording and a recording method using preformat recording.
- a recording apparatus that executes the recording method creates an AV clip in real time and records it on a BD-RE or BD-R, a hard disk, or a semiconductor memory card.
- the AV clip may be a TS obtained by real-time encoding of an analog input signal by the recording device, or may be a TS obtained by partializing the TS digitally input by the recording device.
- a recording apparatus that performs real-time recording includes a video encoder that encodes a video signal to obtain a video stream, an audio encoder that encodes an audio signal to obtain an audio stream, a video stream, an audio stream, and the like.
- a multiplexer that obtains TS and a source packetizer that converts TS packets that make up a digital stream in MPEG2-TS format into source packets, stores the MPEG2 digital stream converted into source packet format in an AV clip file, and stores the BD -Write to RE, BD-R, etc.
- the control unit of the recording apparatus performs processing for generating clip information and playlist information on the memory. Specifically, when a recording process is requested by the user, the control unit creates an AV clip stream file and a clip information file on the BD-RE and BD-R.
- the control unit of the recording apparatus is positioned at the head in this GOP.
- the PTS of the intra picture and the packet number of the source packet storing the head part of this GOP are acquired, and this combination of PTS and packet number is used as the EP_PTS entry and EP_SPN entry in the entry map of the clip information file. Append. Thereafter, each time a GOP is generated, a set of EP_PTS entries and EP_SPN entries is added to the entry map of the clip information file.
- the is_angle_change flag set to “ON” is added to the set of the EP_PTS entry and the EP_SPN entry. If the head of the GOP is not an IDR picture, the is_angle_change flag set to “off” is added to the set of EP_PTS entry and EP_SPN entry.
- the stream attribute information in the clip information file is set according to the stream attribute to be recorded. If AV clips and clip information are generated and written to BD-RE and BD-R as described above, playlist information that defines the playback path is generated via the basic entry map in the clip information. Write to BD-RE and BD-R. By executing such processing in the real-time recording technique, a hierarchical structure of AV clip-clip information-playlist information can be obtained on BD-RE and BD-R.
- the above is a recording apparatus that executes a recording method by real-time recording. Next, a recording method by preformat recording will be described.
- the recording method by preformat recording is an optical disc manufacturing method including an authoring process.
- FIG. 64 shows an optical disk recording method.
- FIG. 4A is a flowchart showing a recording method by preformat recording, and shows a processing procedure of an optical disc manufacturing method.
- the optical disk manufacturing method includes an authoring step, a signature step, a media key acquisition step, a media key encryption step, a physical formatting step, an identifier embedding step, a mastering step, and a replication step.
- the authoring step S201 creates a bit stream representing the entire image of the volume area of the optical disc.
- AACS ⁇ LA is an organization that manages licenses related to copyright protection technology for next-generation digital home appliances.
- the authoring site that authors an optical disc using the authoring device and the mastering site that performs mastering using the mastering device receive a license from AACS-LA. It also manages media keys and invalidation information. Then, a part of the bit stream signed from AACS LA is acquired.
- Media key acquisition step S203 acquires a media key from AACS LA.
- a unique media key is not always used, and a new one is updated when the number of optical discs manufactured so far reaches a certain number.
- By updating the media key it is possible to exclude a specific manufacturer or device, and even if the encryption key is broken, it can be invalidated by using the invalidation information.
- Media key encryption step S204 encrypts the key used for encryption of the bitstream using the media key acquired in the media key acquisition step.
- Physical format step S205 executes physical formatting on the bitstream.
- identifier embedding step S206 a unique identifier that cannot be detected by a general device is embedded in the bit stream recorded on the optical disc as a digital watermark. This can prevent mass production of pirated copies due to unauthorized mastering.
- Mastering step S207 creates an optical disc master.
- a photoresist layer is formed on a glass substrate, and the photoresist layer is exposed to a laser beam so as to correspond to a desired groove or pit, and is subjected to development processing.
- This groove or pit represents each bit value of the 8-16 modulated bit stream.
- an optical disc master is manufactured based on the photoresist in which irregularities corresponding to grooves and pits are formed by such laser cutting.
- Replication step S208 mass-produces an optical disc that is a duplicate of the master disc of the optical disc.
- FIG. 5B shows a case where a general user does not mass-produce an optical disc, but a general user uses a PC to record various files described in the above embodiments on a BD-R, BD-RE, etc.
- the processing procedure of the recording method by format recording is shown. Compared to (a) in the figure, in the recording method according to (b) in the figure, there is no physical format (step S205), mastering (step S207), and replication (step S208). Instead, writing of each file is performed. There is a process (step S209).
- FIG. 65 is a flowchart showing the processing procedure of the authoring process.
- reel sets for the main TS and sub-TS are defined.
- a “reel” is a file that stores data that is a material of an elementary stream, and exists in a drive on a local network in the authoring system. These reels include L images and R images captured by a 3D camera, audio recorded at the time of recording, audio recorded afterwards, subtitles for each language, and menu data.
- a “reel set” is a group of links to a material file that represents a set of elementary streams to be multiplexed in one TS.
- a reel set is defined for each of the main TS and sub-TS.
- step S102 the prototypes of the play item and the sub play item are defined, and the reproduction order of the play item and the sub play item is defined, thereby defining the prototype of the main path and the sub path.
- the definition of the prototype of the play item is made by accepting designation of a reel that should be permitted to be reproduced in the play item and designation of In_Time / Out_Time through the GUI in the planar output mode.
- the definition of the prototype of the sub play item is made by accepting, via the GUI, designation of a reel that should be permitted to be played on the play item corresponding to the sub play item and designation of In_Time / Out_Time in the stereoscopic output mode. .
- Specified reels that should be allowed to be played back consist of a GUI that checks the check box for the material file links in the reel set that should be allowed to play. At this time, a numerical value input field is displayed in association with each reel. Then, the priority order for each reel is received by this numerical value input field, and this is set as the priority order corresponding to the reel.
- a stream selection table and an extended stream selection table are generated from the above-described reel settings that should be permitted to be reproduced and priority settings.
- In_Time and Out_Time are specified on the GUI by displaying the time axis of the base view video stream or dependent view video stream as a graphic, and moving the slide bar along the graphic time axis. Is performed by the recording apparatus executing a process of accepting the position setting from the user.
- the definition of the play order of the play item and the sub play item is displayed on the GUI by displaying the picture of the play item In_Time as a thumbnail, and the recording apparatus accepts an operation of setting the play order for the thumbnail from the user. Made in
- a plurality of elementary streams are obtained by encoding the material file specified by the reel set.
- the plurality of elementary streams include a base-view video stream and a dependent-view video stream, and an audio stream, a PG stream, and an IG stream to be multiplexed with the base-view video stream and the dependent-view video stream. .
- step S104 the same elementary stream belonging to the same reel set as the base-view video stream among the elementary streams obtained by encoding is multiplexed with the base-view video stream to obtain one main TS.
- step S105 one sub-TS is obtained by multiplexing the elementary streams obtained by encoding that belong to the same reel set as the dependent-view video stream with the dependent-view video stream.
- step S106 a prototype of the clip information file is generated based on the parameters set at the time of encoding and multiplexing.
- step S107 play item information and sub play item information are generated based on a prototype of the play item, and main path information and sub path information are generated by defining a playback order in these play item information and sub play item information. Then, play list information is defined.
- the play item In the creation of play item information, the play item is set to be playable in the elementary stream multiplexed in the main TS, which is specified to be played in the planar view output mode in the basic structure of the play item.
- a stream selection table is generated in the information.
- In_Time and Out_Time defined in the above-described editing work are described in the play item information in order to define the playback section in the base-view video stream.
- the elementary stream multiplexed in the sub TS is set to playable in the basic structure of the play item that is specified to be played in the stereoscopic output mode.
- An extension stream selection table is generated in the extension data of the list information. Since the play item information and the sub play item information are defined based on the information in the clip information file, they are set based on the prototype of the clip information file.
- step S108 the main TS, the sub TS, the prototype of the clip information file, and the prototype of the playlist information are converted into a directory file group according to a predetermined application format.
- main TS, sub TS, clip information, play item information, and sub play item information are generated, main TS and sub TS are converted into independent stream files, and clip information is clip information.
- clip information is clip information.
- the offset sequence can be created in the process of encoding. .
- FIG. 66 shows the processing procedure of the AV file writing process.
- AV file writing is realized by the flowchart of this figure.
- step S401 xxxxx.ssif is created and a file entry is created on the memory of the recording device.
- step S402 it is determined whether or not an empty continuous sector area can be secured. If it can be secured, in step S403, the source packet sequence constituting the dependent view data block is converted to EXT2 in the empty continuous sector area. Only [i] is written, and then Steps S404 to S408 are executed. If it cannot be secured, an exception process is performed in step S409, and then the recording method is terminated.
- Steps S404 to S408 constitute a loop in which the processes of Steps S404 to S406 and Step S408 are repeated until Step S407 is determined to be No.
- step S405 only EXT1 [i] writes the source packet sequence constituting the base view data block in the empty continuous sector area.
- step S406 the start address in which the source packet sequence is written and the allocation identifier indicating the continuous length are added to the file entry and registered as an extent.
- extent start point information indicating the head source packet number of the written source packet sequence is added to the metadata in the clip base information and clip dependent information.
- Step S407 defines a loop termination condition, and determines whether or not an unwritten source packet exists in the base view data block and the dependent view data block. If it exists, the process proceeds to step S408 to continue the loop. If it does not exist, the process proceeds to step S410.
- Step S408 is a determination as to whether or not there is a continuous sector area. If it exists, the process proceeds to step S403, and if it does not exist, the process returns to step S402.
- step S410 xxxxx.ssif is closed and the file entry is written to the recording medium.
- step S411 xxxxx.m2ts is created, and a file entry of xxxxx.m2ts is generated in the memory.
- step S412 an allocation descriptor indicating the start address and continuous length of the base view data block unique to the file 2D is added to the file entry of xxxxx.m2ts.
- xxxxx.m2ts is closed and a file entry is written.
- Step S404 is a determination as to whether or not a long jump occurrence point exists within the range of EXTSS + EXT2D.
- the long jump occurrence point is assumed to be a layer boundary. If a layer boundary exists within the range of EXTSS + EXT2D, the base view data block is duplicated and a long jump is generated between the base view data block B [i] ss and the base view data block B [i] 2D in step S420. Writing is performed immediately before the point, and then the process proceeds to step S406. These become file 2D extents and file-based extents.
- EXT2D EXT1 [n]
- EXT2 [n] EXT2 [n]
- EXTss [n] EXTss
- the lower limit of EXT2D is determined so as not to cause a buffer underflow of the read buffer in the playback device during the jump period from each base view data block to the next base view data block during playback in 2D output mode.
- the jump from the nth base view data block to the (n + 1) th base view data block takes time Tjump2D (n), and each base view data block is read to the read buffer at the speed Rud2D, and
- the lower limit value of EXT2D is expressed by the following condition 1.
- the double buffer is composed of a read buffer 1 and a read buffer 2.
- the read buffer 1 is the same as the read buffer of the 2D playback device.
- the jump from the nth base view data block to the pth dependent view data block takes time TFjump3D (n), and from the pth dependent view data block ( Assume that the jump to the (n + 1) th base view data block takes time TBjump3D (n).
- each base view data block is read to the read buffer 1 at the speed Rud3D
- each dependent view data block is read to the read buffer 2 at the speed Rud3D
- the base view data is read from the read buffer 1 to the video decoder.
- the lower limit value of EXT1 [n] is expressed by the following condition 2.
- the continuous length of the big extent is set to this lower limit value or a value exceeding this lower limit value.
- the jump from the (n + 1) th base view data block to the (p + 1) th dependent view data block takes time TFjump3D (n + 1), and the read buffer 2 sends it to the decoder.
- the lower limit value of EXT2 [n] is expressed by the following condition 3.
- Tdiff means a delay time associated with the preloading of the dependent view data block in EXTss and the preloading of the dependent view data block in EXTssnext. The meaning of Tdiff is explained below.
- EXTss playback is delayed by the period required for preloading this dependent-view data block, so in EXTss, the time required for preloading the data block of the first dependent-view data block is the delay period of playback. "Become.
- the playback start of the video decoder may be delayed only during that time. That is, when EXTssnext is reproduced, the period during which the first dependent view data block is preloaded is a “grace period” during which the start of video decoder reproduction is suspended.
- Tdiff is derived as a value obtained by subtracting the delay period from the grace period of the dependent view data block. Specifically, it is calculated so as to satisfy the following formula.
- Tdiff ceil [(((S1stEXT1 [i] EXTSSnext] -S1stEXT1 [i] EXTSS) x1000x8) / Rud72]
- Tdiff means the difference between the reading period of S1stEXT2 [i] EXTss and the reading period of S1stEXT2 [i] EXTssnext, and S1stEXT2 [i] EXTss is the size of EXT2 [i] located at the beginning of EXTss.
- EXTssNEXT is the size of EXT2 [i] located at the beginning of EXTssNEXT.
- EXTssnext is an extent in a stereoscopic interleaved stream file, is located immediately after EXTss, and is reproduced seamlessly with EXTss.
- Rextss is the average bit rate of EXTss and is derived from the following formula.
- ATCDextss ceil [Nsp ⁇ 188 ⁇ 8 / (ATCDextss / 27000000)]
- ATCDextss ATCstart_extssnext -ATCstart_extss
- ATCDextss ATClast_extss-ATCstart_extss + ceil (27000000x188x8 / min (Rts1, Rts2)) ATCDextss is the ATC period of EXTss.
- ATCstart_EXTSS is the minimum ATC value indicated by the ATC field of the source packet sequence in EXTss.
- ATCstart_EXTssnext is the minimum ATC value indicated by the ATC field of the source packet sequence in EXTssnext.
- ATClast_EXTSS is the maximum ATC value indicated by the ATC field of the source packet sequence in EXTss.
- Nsp is the number of source packets in the main TS and sub-TS that have an ATC value corresponding to an ATC within the range of ATCDexss.
- Rts1 is the value of the TS recording rate in the main TS, and its maximum value is 48 Mbps.
- Rts2 is the TS recording rate value in the sub-TS, and its maximum value is 48 Mbps.
- EXTSs contains the first data byte of the ATC sequence used by the previous play item (play item 1) if two play items are present continuously.
- EXTss has a size that is equal to or larger than the minimum extent size defined in Condition 4.
- EXTss is the first data byte of the ATC sequence used in the previous play item
- EXTss contains the data byte of the ATC sequence used by the current play item (play item 2).
- EXTss has a size that is equal to or larger than the minimum extent size defined in Condition 4.
- EXTss is the last data byte of the ATC sequence used in play item 2
- the recording process is performed as follows.
- This boundary match divides the access unit delimiter of the GOP (i) first video access unit in the sub-TS as being the boundary of the dependent view data block, and the GOP (i) first video access in the main TS This is done by dividing the unit's access unit delimiter as being the boundary of the base view data block. In this division, the extent length restriction described above must be satisfied.
- the top video access of GOP (i) in the sub-TS is attempted by inserting a padding packet either immediately before the access unit delimiter of the unit or immediately before the access unit delimiter of the first video access unit of GOP (i) in the main TS.
- an entry pointing to the source packet number of the source packet storing the access unit delimiter of the first access unit of the dependent view GOP is added to the extension entry map.
- an entry for pointing the source packet of the source packet storing the access unit delimiter of the head access unit of the base view GOP is added to the basic entry map.
- the source packet storing the access unit delimiter of the first access unit of the dependent view GOP is located in the middle of the dependent view data block without boundary matching by padding packet insertion being successful, pointing to the source packet Do not add such entries to the extended entry map.
- the source packet storing the access unit delimiter of the first access unit of the base view GOP is located in the middle of the base view data block, an entry that points to the source packet is not added to the extension entry map.
- the process of generating an entry map after recording the base view data block and the dependent view data block detects the GOP head from the recorded stereoscopic interleaved stream file, and enters an entry pointing to this GOP head. This is realized by adding to the map.
- a procedure for creating a basic entry map and an extended entry map by GOP head detection and entry addition will be described with reference to FIG.
- FIG. 67 is a flowchart showing a procedure for creating a basic entry map and an extended entry map.
- step S601 a basic entry map and an extended entry map are generated on the memory, and the process proceeds to a loop of steps S602 to S610.
- the variable x indicates the GOP.
- the variable x is initialized to 1 (step S602), the head of GOP (x) is specified (step S603), and the head PTS of the GOP SPN (x) corresponding to (x) is specified (step S604). Thereafter, the determinations in step S605 and step S607 are executed.
- Step S605 determines whether SPN (x) is the head of EXT1 [i]. If it is not the head, steps S606 to S609 are skipped. If it is the head, in step S606, EXT2 [j] having SPN (y) corresponding to PTS (x) at the head is specified.
- Step S607 is a determination as to whether or not the variable i specifying EXT1 [i] matches the variable j specifying EXT2 [j]. If the variable i does not match, the subsequent processing is skipped. If they match, EP_entry (x) indicating the combination of PTS (x) -SPN (x) is added to the basic entry map (step S608), and EP_entry (x) indicating the combination of PTS (x) -SPN (y) Is added to the extended entry map (step S609).
- Step S610 is a determination of whether or not the variable x indicates the last GOP. If not, the variable x is incremented and the process proceeds to step S603.
- the index table described in the third embodiment can be generated as follows.
- the display frequency of each playlist to be recorded on the recording medium can be determined.
- the resolution and display frequency of the playlist used in the first play title, or the resolution and display frequency of the playlist designated by the title numbers 0 to 999 are displayed in the video format in the BDMV application information of the index table.
- FIG. 68 is a flowchart showing a procedure for creating a BD-J application, a BD-J object, a movie object, and an index table.
- a source program for instructing the playback device to generate a player instance for the playlist is created by object-oriented programming.
- the created source program is compiled and archived. Create a J application.
- Step S703 is a process of creating a BD-J object, and step S704 describes a movie object using a command for playing a playlist.
- step S705 an index table is created by describing the association between the title number and the BD-J object, and in step S706, a playlist to be used as the first play title is selected.
- step S707 BDMV application information indicating the video format and video rate of the playlist in the first play title is created.
- an index table including the title index and BDMV application information is created.
- the BD-J object, BD-J application, movie object, and index table are written to the recording medium.
- a recording medium that can be created with the above recording will be described.
- FIG. 69 shows the internal structure of a multilayered optical disc.
- the first row shows an example of a multilayered optical disk
- the second row is a drawing in which spiral tracks existing on each recording layer are stretched in the horizontal direction.
- the spiral tracks in these recording layers are treated as one continuous volume area.
- the volume area is composed of a lead-in located at the innermost circumference, a lead-out located at the outermost circumference, a recording area of the first recording layer, a recording area of the second recording layer, and a recording area of the third recording layer existing therebetween. Is done.
- the recording area of the first recording layer, the recording area of the second recording layer, and the recording area of the third recording layer constitute one continuous logical address space.
- the volume area is assigned a serial number in units of accessing the optical disc from the beginning, and this number is called a logical address.
- Reading data from the optical disk is performed by designating a logical address.
- a read-only disk such as a BD-ROM
- basically sectors having consecutive logical addresses are also continuous in physical arrangement on the optical disk. That is, the data of sectors with consecutive logical addresses can be read without seeking.
- continuous reading is not possible at the boundary of the recording layer even if logical addresses are continuous. Therefore, it is assumed that the logical address of the layer boundary is registered in advance in the recording apparatus.
- file system management information is recorded immediately after the lead-in area, followed by a partition area managed by the file system management information.
- a file system is a mechanism for expressing data on a disc in units called directories or files.
- the file system is recorded in UDF (Universal Disc Disc Format).
- UDF Universal Disc Disc Format
- FAT file system
- NTFS file system
- the fourth row shows the area allocation in the file system area managed by the file system.
- a non-AV data recording area exists on the inner circumference side.
- An AV data recording area exists immediately after the non-AV data recording area.
- the fifth row shows the recorded contents of these non-AV data recording areas and AV data recording areas.
- In the AV data recording area there are extents constituting the AV file.
- In the non-AV data recording area there are extents constituting non-AV files other than AV files.
- FIG. 70 shows an application format of an optical disc based on the file system.
- BDMV directory is a directory where data such as TS and management information handled by BD-ROM is recorded. Under the BDMV directory, there are five subdirectories called “PLAYLIST directory”, “CLIPINF directory”, “STREAM directory”, “BDJO directory”, and “JAR directory”. In the BDMV directory, “index.bdmv ”And“ MovieObject.bdmv ”are arranged.
- Index.bdmv (fixed file name) stores the index table.
- “MovieObject.bdmv (fixed file name)” stores one or more movie objects.
- a movie object is a program file that defines the control procedure to be performed by the playback device in an operation mode (HDMV mode) mainly controlled by a command interpreter.
- HDMV mode operation mode
- a user can use one or more commands, GUI menu calls, and title calls. Contains a mask flag that specifies whether to mask these calls.
- BDJO directory there is a program file (xxxxx.bdjo [“xxxxx” is variable, extension “bdjo” is fixed]) with the extension bdjo.
- This program file stores a BD-J object that defines a control procedure to be performed by the playback device in the BD-J mode.
- Java (registered trademark) application is actually the Java (registered trademark) archive file (YYYYY.jar) stored in the JAR directory under the BDMV directory.
- the application is, for example, a Java (registered trademark) application, and includes one or more xlet programs loaded in a heap area (also referred to as work memory) of the virtual machine.
- An application is configured from the xlet program and data loaded in the work memory.
- Extents constituting files existing in the above directories are recorded in the non-AV data area.
- the “STREAM directory” is a directory that stores stream files. This directory stores stream files in the format xxxxx.m2ts ([xxxxx is variable, extension “m2ts” is fixed]).
- Partition areas are "area where file set descriptor is recorded”, “area where end descriptor is recorded”, “ROOT directory area”, “BDMV directory area”, “JAR directory area”, “BDJO directory area” , “PLAYLIST directory area”, “CLIPINF directory area”, and “STREAM directory area”, which are areas accessed by the file system.
- ROI directory area "area where file set descriptor is recorded”
- end descriptor is recorded
- ROOT directory area "BDMV directory area”
- JAR directory area JAR directory area
- BDJO directory area BDJO directory area
- PROLAYLIST directory area “CLIPINF directory area”
- STREAM directory area which are areas accessed by the file system.
- File set descriptor includes a logical block number (LBN) indicating a sector in which a file entry of the ROOT directory is recorded in the directory area.
- LBN logical block number
- end descriptor indicates the end of the file set descriptor.
- the plurality of directory areas as described above all have a common internal configuration. That is, the “directory area” is composed of a “file entry”, a “directory file”, and a “file recording area for a lower file”.
- “File entry” includes “descriptor tag”, “ICB tag”, and “allocation descriptor”.
- Descriptor tag is a tag indicating that it is a file entry.
- Allocation descriptor includes a logical block number (LBN) indicating the recording position of the directory file. This completes the description of the file entry. Next, the details of the directory file will be described.
- LBN logical block number
- Database file includes “file identification descriptor for lower directory” and “file identification descriptor of lower file”.
- the “file identifier descriptor of the lower directory” is reference information for accessing the lower directory under its own, the identification information indicating the lower directory, the length of the directory name of the lower directory, and the lower directory It consists of a file entry address indicating in which logical block number the file entry is recorded and the directory name of the lower directory.
- the “file identification descriptor of the lower file” is reference information for accessing the file under its own, the identification information indicating the lower file, the length of the lower file name, and the file entry for the lower file Is composed of a file entry address indicating in which logical block number is recorded and a file name of a lower file.
- the file identifier descriptors in the directory files of these directories indicate in which logical block the file entries of the lower directory and the lower file are recorded. If this file identifier descriptor is followed, The file entry in the BDMV directory can be reached from the file entry in the ROOT directory, and the file entry in the PLAYLIST directory can be reached from the file entry in the BDMV directory. Similarly, file entries in the JAR directory, BDJO directory, CLIPINF directory, and STREAM directory can be reached. *
- the “file recording area of a lower file” is an area in which the substance of a lower file under a certain directory is recorded. A “file entry” and one or more “extents” for the lower file are included. It is recorded.
- the main stream file of the present application is a file recording area existing in the directory area of the directory to which the file belongs, and follows the file identification descriptor in the directory file and the allocation identifier in the file entry. Can be accessed.
- FIG. 71 shows the configuration of the 2D / 3D playback device.
- the 2D / 3D playback device includes a BD-ROM drive 1, a read buffer 2a, a read buffer 2b, a switch 3, a system target decoder 4, a plane memory set 5a, a plane synthesis unit 5b, an HDMI transmission / reception unit 6, a playback control unit 7, and a memory.
- Register set 203 program execution unit 11, program memory 12, HDMV module 13, BD-J platform 14, middleware 15, mode management module 16, user event processing unit 17, local storage 18, and nonvolatile memory 19.
- the BD-ROM drive 1 reads data from the BD-ROM disc based on a request from the playback control unit 7 in the same manner as the 2D playback device, but the AV clip read from the BD-ROM disc is read from the read buffer 2a. It is transferred to the buffer 2b.
- the reproduction control unit 7 sends a read request instructing to alternately read out the base view data block and the dependent view data block in extent units.
- the BD-ROM drive 1 reads the extents constituting the base view data block into the read buffer 2a, and reads the extents constituting the dependent view data block into the read buffer 2b.
- speed performance higher than the BD-ROM drive of the 2D playback device is required.
- the read buffer 2a is a buffer composed of a dual port memory or the like for storing data of the base view data block read by the BD-ROM drive 1.
- the read buffer 2b is a buffer composed of a dual port memory or the like for storing the data of the dependent view data block read by the BD-ROM drive 1.
- the switch 3 is a switch for switching the data input source for the read buffer to either the BD-ROM drive 1 or the local storage 18.
- the system target decoder 4 performs a demultiplexing process on the source packet read to the read buffer 2a and the source packet read to the read buffer 2b to perform a stream decoding process.
- the plane memory set 5a is composed of a plurality of plane memories.
- the plane memory includes a left view video plane, a right view video plane, a secondary video plane, an IG plane, and a PG plane.
- the plane synthesis unit 5b performs the plane synthesis described in the previous embodiments.
- a TV etc.
- it is necessary to play back the left eye image and right eye image alternately using shutter glasses it is output as it is.
- a temporary buffer is prepared and the left eye is transferred first. Store the image in a temporary buffer and output it simultaneously after the right-eye image is transferred.
- the HDMI transmitting / receiving unit 6 executes the negotiation phase described in the first embodiment in the HDMI standard (HDMI: High Definition Multimedia Interface), for example.
- HDMI High Definition Multimedia Interface
- the playback control unit 7 includes a playback engine 7a and a playback control engine 7b.
- the playback of the 3D playlist is instructed by the program execution unit 11 or the like, the base view data block of the play item to be played back in the 3D playlist And the dependent view data block of the sub play item of the 3D sub-pass that is played back in synchronization with the play item.
- it interprets the entry map of the corresponding clip information file, and based on the extent start type that indicates which extent is located first, the extent of the base view data block and the dependent view data from the playback start point It requests the BD-ROM drive 1 to read the block extents alternately.
- the first extent is read into the read buffer 2a or the read buffer 2b, and then transfer from the read buffer 2a and the read buffer 2b to the system target decoder 4 is started.
- the playback engine 7a executes an AV playback function.
- AV playback function is a group of functions followed from DVD playback device and CD playback device, playback start, playback stop, pause, release of pause, release of still image function, fast-forward that specifies the playback speed as immediate value, This is a process such as rewind, audio switching, secondary video picture data switching, and angle switching with the playback speed specified as an immediate value.
- the playback control engine 7b executes a playlist playback function in response to a function call from a command interpreter that is an HDMV mode operation subject and a Java platform that is an BD-J mode operation subject.
- the playlist playback function refers to performing playback start and playback stop according to current playlist information and current clip information constituting the current playlist among the AV playback functions described above.
- the memory is a memory for storing current playlist information and current clip information.
- Current playlist information refers to information that is currently processed among multiple playlist information that can be accessed from a BD-ROM, a built-in media drive, or a removable media drive.
- Current clip information refers to information that is currently processed among a plurality of clip information that can be accessed from a BD-ROM, a built-in media drive, or a removable media drive.
- the playback status / setting register (Player Status / Setting Register) set 10 includes general-purpose registers that can store arbitrary information used by the program file in addition to the playback status register and the playback setting register described in the above embodiments.
- the program execution unit 11 is a processor that executes a program stored in a BD program file. The operation is performed according to the stored program, and the following control is performed. (1) Command the playback control unit 7 to play a playlist. (2) Transfer PNG / JPEG for menu and game graphics to the system target decoder and display it on the screen. These can be performed freely according to the creation of the program, and how to control them depends on the programming process of the BD-J application by the authoring process.
- the program memory 12 is a memory that stores a current dynamic scenario and is used for processing by an HDMV module that is an HDMV mode operating entity and a Java platform that is an BD-J mode operating entity.
- the current dynamic scenario is an index.bdmv, BD-J object, or movie object currently recorded on the BD-ROM that is currently being executed.
- the program memory 12 includes a heap memory.
- the heap memory is a stack area in which the bytecode of the system application, the bytecode of the BD-J application, the system parameter used by the system application, and the application parameter used by the BD-J application are arranged.
- the HDMV module 13 is a command interpreter. And the HDMV mode control is executed by decoding and executing the navigation commands constituting the movie object.
- the BD-J platform 14 is a Java platform that is the main operating entity of the BD-J mode, and is fully loaded with Java2Micro_Edition (J2ME) Personal Basis Profile (PBP 1.0) and Globally Executable MHP specification (GEM1.0.2) for package media targets. It is implemented and consists of a class loader, a bytecode interpreter, and an application manager.
- J2ME Java2Micro_Edition
- PBP 1.0 Personal Basis Profile
- GEM1.0.2 Globally Executable MHP specification
- the class loader is one of the system applications, and loads the BD-J application by reading the byte code from the class file existing in the JAR archive file and storing it in the heap memory 31.
- Bytecode interpreter is a so-called Java virtual machine that converts the bytecodes that make up the BD-J application stored in the heap memory and the bytecodes that make up the system application into native codes, which are executed by the MPU.
- the application manager is one of the system applications. Based on the application management table in the BD-J object, the BD-J application application signaling, such as starting the BD-J application or ending the BD-J application I do. This completes the description of the internal configuration of the BD-J platform.
- the middleware 15 is an operating system for embedded software, and includes a kernel and device drivers.
- the kernel provides a playback device-specific function to the BD-J application in response to an application programming interface (API) call from the BD-J application.
- API application programming interface
- hardware control such as activation of an interrupt handler unit by an interrupt signal is realized.
- the mode management module 16 holds Index.bdmv read from the BD-ROM, built-in media drive, or removable media drive, and performs mode management and branch control.
- the mode management by the mode management module is a module assignment that determines whether the BD-J platform 22 or the HDMV module executes a dynamic scenario.
- the user event processing unit 17 requests the program execution unit 16 and the playback control unit 7 to execute processing in response to a user operation through the remote controller. For example, when a button is pressed with the remote controller, the program execution unit 16 is requested to execute a command included in the button. For example, when the fast forward / rewind button is pressed on the remote controller, the playback control unit 7 is instructed to execute fast forward / rewind processing for the AV clip of the playlist currently being played back.
- the local storage 18 includes a built-in media drive for accessing the hard disk and a removable media drive for accessing the semiconductor memory card, and is used for storing downloaded additional content and data used by the application.
- the storage area for additional content is divided for each BD-ROM, and the area that an application can use to hold data is divided for each application.
- the non-volatile memory 19 is a recording medium such as a readable / writable memory, and is a medium capable of holding recorded contents even when power is not supplied, such as a flash memory or FeRAM. This is used for backup of stored contents in the register set 203.
- FIG. 72 (a) is a diagram showing a form of usage of the recording medium according to the present invention.
- the recording medium according to the present invention is a BD-ROM 101.
- the BD-ROM 101 is used for supplying movie works to a home theater system formed by a playback device 102, a television 103, and a remote controller 104.
- FIG. 72 (b) is a diagram showing the structure of the BD-ROM.
- the BD-ROM 101 is shown on the fourth level of this figure, and tracks on the BD-ROM are shown on the third level.
- the track formed in a spiral shape from the inner periphery to the outer periphery of the BD-ROM 101 is drawn in the horizontal direction.
- the BD-ROM 101 has a spiral recording area from the inner circumference to the outer circumference like other optical disks such as DVD and CD, and can record logical data between the inner lead-in and the outer lead-out. It has a volume area.
- the volume area is assigned a serial number in units of accessing the optical disc from the beginning, and this number is called a logical address. Reading data from the optical disk is performed by designating a logical address.
- the logical addresses are defined as being continuous even in the physical arrangement on the optical disk. That is, data having consecutive logical addresses can be read without performing a seek.
- BCA Burt Cutting Area
- file system volume information is recorded from the beginning, followed by application data such as video data.
- the file system is a mechanism for expressing data on a disc in units called directories or files.
- directories or files In the case of the BD-ROM 101, it is recorded in UDF (Universal Disc Format).
- UDF Universal Disc Format
- FAT or NTFS file system
- This file system makes it possible to read logical data recorded in the same way as a normal PC using a directory and file structure.
- the directory and file structure on the BD-ROM 101 has a BDMV directory directly under the root directory (ROOT).
- the BDMV directory is a directory in which data such as AV content and management information handled by the BD-ROM 101 is recorded.
- Under the BDMV directory there are an index file (index.bdmv) in which an index table constituting a title is defined, a PLAYLIST directory, a CLIPINF directory, a STREAM directory, a BDJO directory, and a JAR directory.
- AV clip (XXX.M2TS) in which AV content such as video and audio is multiplexed and stored, clip information file (XXX.CLPI) in which AV clip management information is stored, and play that defines the logical playback path of the AV clip
- a list file (YYY.MPLS) and a BD program file (AAA.PROG) in which a program for defining a dynamic scenario is stored are placed under the STREAM directory, CLIPINF directory, and PLAYLIST directory, respectively.
- the index file has an index table shown in FIG.
- the index table is the highest layer table that defines the title structure such as all titles, top menus, and first play stored in the BD-ROM. In this table, all titles, top menus, and program files to be executed first from FirstPlay are designated.
- the BD-ROM player refers to the index table and executes a predetermined BD program file.
- FirstPlay is set by the content provider, and a BD program file that is automatically executed when the disc is inserted is set.
- the top menu specifies a movie object and a BDJ object to be called when a command such as “return to menu” is executed by a user operation on the remote controller.
- BD program file (AAA.PRG) stores the program to be specified and executed from the title. There are several program files, and the prefix (AAA) is used to identify the program.
- the program used here is a proprietary interpreter program called command navigation in the case of Blu-ray, but the language system is not the essence of the present invention, so it is a general-purpose programming language such as Java or JavaScrip. It does not matter.
- a playlist to be played back by this program is designated.
- AV clip is a digital stream in MPEG-2 transport stream format.
- FIG. 73 (a) is a diagram showing the configuration of an AV clip.
- an AV clip is obtained by multiplexing one or more of a video stream, an audio stream, a presentation graphics stream (PG), and an interactive graphics stream.
- the video stream indicates the main video and sub-video of the movie
- the audio stream (IG) indicates the main audio portion of the movie and the sub-audio mixed with the main audio
- the presentation graphics stream indicates the subtitles of the movie.
- the main video indicates a normal video displayed on the screen
- the sub-video is a video displayed on a small screen in the main video.
- the interactive graphics stream indicates an interactive screen created by arranging GUI parts on the screen.
- the video stream is encoded and recorded using a method such as MPEG-2, MPEG-4 AVC, or SMPTE VC-1.
- the audio stream is compressed and encoded by a method such as Dolby AC-3, Dolby Digital Plus, MLP, DTS, DTS-HD, or linear PCM.
- moving picture compression coding such as MPEG-2, MPEG-4 AVC, SMPTE VC-1, etc.
- data amount is compressed using redundancy in the spatial direction and temporal direction of moving images.
- inter-picture predictive coding is used as a method of using temporal redundancy.
- inter-picture predictive coding when a certain picture is coded, a picture that is forward or backward in display time order is used as a reference picture. Then, the amount of motion from the reference picture is detected, and the amount of data is compressed by removing the redundancy in the spatial direction from the difference value between the motion compensated picture and the picture to be coded.
- a picture that does not have a reference picture and performs intra-picture predictive coding using only a picture to be coded is called an I picture.
- a picture is a unit of encoding that includes both a frame and a field.
- a picture that is inter-picture prediction encoded with reference to one already processed picture is called a P picture
- a picture that is inter-picture predictively encoded with reference to two already processed pictures at the same time is called a B picture.
- a picture that is referred to by other pictures in the B picture is called a Br picture.
- a field having a frame structure and a field having a field structure are referred to as a video access unit here.
- Each stream included in the AV clip is identified by PID.
- PID For example, an array x1011 for a video stream used for movie images, an array x1100 to 0x111F for an audio stream, an array x1200 to 0x121F for a presentation graphics, and an array x1400 to 0x141F for an interactive graphics stream
- Arrays x1B00 to 0x1B1F are allocated to video streams used for sub-pictures
- arrays x1A00 to 0x1A1F are allocated to audio streams used for sub-audios to be mixed with main audio.
- FIG. 73 (b) is a diagram schematically showing how AV clips are multiplexed.
- a video stream 501 composed of a plurality of video frames and an audio stream 504 composed of a plurality of audio frames are converted into PES packet sequences 502 and 505, respectively, and converted into TS packets 503 and 506, respectively.
- the data of the presentation graphics stream 507 and the interactive graphics 510 are converted into PES packet sequences 508 and 511, respectively, and further converted into TS packets 509 and 512.
- the AV click step S513 is configured by multiplexing these TS packets into one stream.
- Fig. 74 shows in more detail how the video stream is stored in the PES packet sequence.
- the first level in the figure shows a video frame sequence of the video stream.
- the second level shows a PES packet sequence.
- multiple Video Presentation Unit I-pictures, B-pictures, and P-pictures in the video stream are divided for each picture and stored in the payload of the PES packet.
- Each PES packet has a PES header, and a PTS (Presentation Time-Stamp) that is a picture display time and a DTS (Decoding Time-Stamp) that is a picture decoding time are stored in the PES header.
- PTS Presentation Time-Stamp
- DTS Decoding Time-Stamp
- FIG. 74 (b) shows the format of the TS packet that is finally written in the AV clip.
- the TS packet is a 188-byte fixed-length packet composed of a 4-byte TS header having information such as a PID for identifying a stream and a 184-byte TS payload for storing data.
- the PES packet described above is divided into TS payloads.
- a 4-byte TP_Extra_Header is added to a TS packet, forms a 192-byte source packet, and is written to an AV clip.
- TP_Extra_Header describes information such as ATS (Arrival_Time_Stamp).
- ATS indicates the transfer start time of the TS packet to the PID filter of the system target decoder 1503 described later.
- Source packets are arranged in the AV clip as shown in the lower part of FIG. 74B, and the number incremented from the head of the AV clip is called SPN (source packet number).
- TS packets included in AV clips include PAT (Program Association Table), PMT (Program Map Table), PCR (Program Clock Reference), etc. in addition to video, audio, subtitles, and other streams.
- the PAT indicates what PMT PID is used in the AV clip, and is registered in the PAT array of the PAT itself.
- the PMT has the PID of each stream such as video, audio, and subtitles included in the AV clip, the stream attribute information corresponding to each PID, and various descriptors related to the AV clip.
- the descriptor includes copy control information for instructing permission / non-permission of copying of the AV clip.
- the PCR corresponds to the ATS in which the PCR packet is transferred to the decoder. Has STC time information.
- FIG. 75 (a) is a diagram for explaining the data structure of the PMT in detail.
- a PMT header describing the length of data included in the PMT is arranged at the head of the PMT. Behind that, a plurality of descriptors related to the AV clip are arranged.
- the copy control information described above is described as a descriptor.
- a plurality of pieces of stream information regarding each stream included in the AV clip are arranged after the descriptor.
- the stream information is composed of a stream descriptor in which a stream type, a stream PID, and stream attribute information (frame rate, aspect ratio, etc.) are described in order to identify a compression codec of the stream.
- the clip information file is AV clip management information as shown in FIG. 75 (b), has a one-to-one correspondence with the AV clip, and includes clip information, stream attribute information, and an entry map.
- Clip information is composed of a system rate, a playback start time, and a playback end time as shown in FIG. 75 (b).
- the system rate indicates the maximum transfer rate of the AV clip to the PID filter of the system target decoder described later.
- the ATS interval included in the AV clip is set to be equal to or less than the system rate.
- the playback start time is the PTS of the first video frame of the AV clip, and the playback end time is set by adding the playback interval of one frame to the PTS of the video frame at the end of the AV clip.
- attribute information about each stream included in the AV clip is registered for each PID.
- the attribute information has different information for each video stream, audio stream, presentation graphics stream, and interactive graphics stream.
- the video stream attribute information includes the compression codec used to compress the video stream, the resolution of the individual picture data constituting the video stream, the aspect ratio, and the frame rate. It has information such as how much it is.
- the audio stream attribute information includes the compression codec used to compress the audio stream, the number of channels included in the audio stream, the language supported, and the sampling frequency. With information. These pieces of information are used for initialization of the decoder before the player reproduces it.
- the entry map includes entry map header information 1101, a PTS indicating the display time of each I picture of the video stream included in the AV clip, and the SPN of the AV clip starting from each I picture.
- entry map header information 1101 a PTS indicating the display time of each I picture of the video stream included in the AV clip
- SPN the SPN of the AV clip starting from each I picture.
- entry point ID A value incremented for each entry point with the head set to 0 is referred to as an entry point ID (hereinafter referred to as EP_ID).
- entry map header information 1101 is stored at the top of the entry map, and information such as the PID and the number of entry points of the video stream pointed to by the corresponding entry map is stored here.
- the play list indicates the playback route of the AV clip.
- the play list includes one or more play items 1201, and each play item indicates a playback section for an AV clip.
- Each play item 1201 is identified by a play item ID, and is described in the order to be reproduced in the play list.
- the playlist includes an entry mark 1202 indicating a playback start point.
- the entry mark 1202 can be given within the playback section defined by the play item.
- the entry mark 1202 is attached to a position that can be a playback start point for the play item. Used. For example, it is possible to reproduce a chapter by adding an entry mark 1202 to the position at the beginning of a chapter in a movie title.
- a playback path of a series of play items is defined here as a main path 1205.
- the contents of the play item will be described with reference to FIG. 77 (b).
- the play item includes clip information 1301 to be reproduced, a reproduction start time 1302 and a reproduction end time 1303, a connection condition 1310, and a stream selection table 1305. Since the playback start time and playback end time are time information, the player refers to the entry map of the clip information file, acquires the SPN corresponding to the specified playback start time and playback end time, and specifies the read start position. To perform playback processing.
- the connection condition 1310 indicates the forward play item and the connection type.
- the connection condition 1310 of the play item is “1”, it indicates that the AV clip indicated by the play item does not guarantee seamless connection with the AV clip indicated by the play item before the play item.
- the play item connection condition 1310 is “5” or “6”, it is guaranteed that the AV clip indicated by the play item is seamlessly connected to the AV clip indicated by the play item before the play item.
- the connection condition 1310 is “5”, the continuity of STC between play items may be interrupted, that is, the AV clip of the play item after connection is longer than the video display time at the end of the AV clip of the play item before connection.
- the start time of the first video display time may be discontinuous.
- the AV clip of the play item before connection is input to the PID filter of the system target decoder 1503 to be described later, and the AV clip of the play item after connection is input to the PID filter of the system target decoder 1503 and played back, It is necessary to create an AV clip so that the decoding of the system target decoder 1503 does not fail.
- the connection condition 1310 is “6”
- the AV clip of the play item before connection and the AV clip of the play item after connection must be reproducible as one AV clip. That is, the STC continues and the ATC also continues between the AV clip of the play item before connection and the AV clip of the play item after connection.
- the stream selection table 1305 includes a plurality of stream entries 1309, and the stream entry 1309 includes a stream selection number 1306, stream path information 1307, and stream identification information 1308.
- the stream selection number 1306 is a number that is sequentially incremented from the beginning of the stream entry 1309 included in the stream selection table, and is used for stream identification in the player.
- the stream identification information 1308 is information such as PID, and indicates a stream multiplexed in the AV clip to be referred to.
- attribute information of each stream is recorded at the same time.
- the attribute information is information indicating the nature of each stream, and includes, for example, language attributes in the case of audio, presentation graphics, and interactive graphics.
- the playlist can have one or more sub-paths as shown in FIG. 77 (c).
- the subpaths are assigned IDs in the order of registration in the playlist, and are used as subpath IDs to identify the subpaths.
- the sub path indicates a series of reproduction paths that are reproduced in synchronization with the main path.
- the sub play item has clip information 1301 to be reproduced, a reproduction start time 1302 and a reproduction end time 1303 in the same manner as the play item, and the reproduction start time 1302 and the reproduction end time 1303 of the sub play item are on the same time axis as the main path. expressed.
- the presentation graphics 1 multiplexed in the AV clip of the sub play item # 2 to be played back in synchronization with the playback section is played back in the play item # 2 playback section.
- the sub play item has a field called SP connection condition having the same meaning as the connection condition of the play item.
- the AV clip at the sub play item boundary where the SP connection condition is “5” or “6” needs to follow the same conditions as the connection condition “5” or “6”.
- the above is the data structure for storing 2D video in the BD-ROM that is the recording medium according to the present invention.
- FIG. 78 (a) shows the configuration of the 2D playback device 1500.
- the 2D playback device 1500 includes a BD-ROM drive 1501, a read buffer 1502, a system target decoder 1503, a program recording memory 1504, a management information recording memory 1505, a program execution unit 1506, a playback control unit 1507, a player variable 1508, and a user event processing unit.
- Reference numeral 1509 denotes a plane adder 1510.
- the BD-ROM drive 1501 reads data from the BD-ROM disc based on a request from the playback control unit 1507. AV clips read from the BD-ROM disc are transferred to the read buffer 1502, index files, playlist files, and clip information files are transferred to the management information memory 1505, and movie object files are transferred to the program memory 1504.
- the read buffer 1502 stores data read using a BD-ROM drive
- the management information memory 1505 stores management information of index files, playlist files, and clip information files
- the program recording memory 1504 stores memory for movie object files. And so on.
- the system target decoder 1503 performs demultiplexing processing on the source packet read to the read buffer 1502 and performs stream decoding processing. Information such as codec type and stream attribute necessary for decoding the stream included in the AV clip is transferred from the playback control unit 1507.
- the system target decoder 1503 converts the decoded main video video stream, sub video video stream, interactive graphics stream, and presentation graphics stream into a main video plane, a sub video plane, and an interactive graphics plane (IG plane) as respective plane memories. Write to the presentation graphics plane (PG plane).
- the system target decoder 1503 mixes the decoded main audio stream and sub audio stream and outputs them to a speaker or the like.
- system target decoder 1503 performs a process of decoding graphics data such as JPEG and PNG to be displayed on a menu or the like transferred from the program execution unit 1506 and writing it to the image plane. Details of the system target decoder 1503 will be described later.
- the user event processing unit 1509 requests the program execution unit 1506 and the playback control unit 1507 to execute processing in response to a user operation through the remote controller. For example, when a button is pressed with the remote controller, the program execution unit 1506 is requested to execute a command included in the button. For example, when the fast forward / rewind button is pressed on the remote controller, the playback control unit 1507 is instructed to execute fast forward / rewind processing for the AV clip of the playlist currently being played back.
- the playback control unit 1507 has a function of controlling playback of AV clips by controlling the BD-ROM drive 1501 and the system target decoder 1503.
- the playback control unit 1507 interprets the playlist information and controls the playback processing of the AV clip based on the playback command from the program execution unit 1506 and the notification from the user event processing unit 1509. Also, the playback control unit 1507 sets and references the player variable 1508 and performs a playback operation.
- the player variable 1508 includes a system parameter (PSR) indicating a player state and a general parameter (GPRM) that can be used for general purposes.
- PSR system parameter
- GPRM general parameter
- FIG. 78 (b) is a list of system parameters (PSR).
- PSR0 Language code
- PSR1 Main audio stream number
- PSR2 Subtitle stream number
- PSR3 Angle number
- PSR4 Title number
- PSR5 Chapter number
- PSR6 Program number
- PSR7 Cell number
- PSR8 Selection key information
- PSR9 Navigation timer
- PSR10 Playback time information
- PSR11 Karaoke mixing mode
- PSR12 Parental country information
- PSR13 Parental level
- PSR14 Player settings (video)
- PSR15 Player setting value (audio)
- PSR16 Audio stream language code
- PSR17 Audio stream language code
- PSR18 Subtitle stream language code
- PSR19 Subtitle stream language code
- PSR20 Player region code
- PSR21 2D / 3D output mode user priority selection
- PSR22 Current 2D / 3D output mode
- PSR23 Display 3D video display capability
- PSR24 3D video playback capability
- PSR25 Spare PSR26: Spare PSR27: Spare PSR28: Spare
- the PSR16 audio stream language code and the PSR18 subtitle stream language code are items that can be set by the OSD of the player, and indicate the default language code of the player. For example, if the language code for the audio stream of PSR16 is English, when playing a playlist, find a stream entry with the same language code from the stream selection table of the play item, select that audio stream, and play it Functions such as can be put into the BD program file.
- the playback control unit 1507 plays back while checking the status of the system parameters.
- PSR1, PSR2, PSR21, and PSR22 indicate the audio stream number, subtitle stream number, sub-video stream, and sub-audio stream number, respectively, and these values correspond to the above-described stream selection number 606.
- the audio stream number PSR1 is changed by the program execution unit 1506.
- the playback control unit 1507 refers to the stream entry 609 that matches the stream selection number 606 in the stream selection table 605 of the play item currently being played back, and switches the audio stream to be played back. In this manner, which audio / subtitle / sub-picture stream is to be reproduced is switched.
- the program execution unit 1506 is a processor that executes a program stored in a BD program file. The operation is performed according to the stored program, and the following control is performed. (1) Command the playback control unit 1506 to play a playlist. (2) Transfer PNG / JPEG for menu and game graphics to the system target decoder and display it on the screen. These can be performed freely according to the creation of the program, and how to control them is determined by the programming process of the BD program file by the authoring process.
- the plane adder instantly superimposes the main video plane, sub-video plane, interactive graphics plane, presentation graphics plane, and image plane and displays them on a screen such as a TV.
- the source depacketizer interprets the source packet transferred to the system target decoder 1503, extracts the TS packet, and sends it to the PID filter.
- the input time to the decoder is adjusted according to the ATS of each source packet. Specifically, at the moment when the ATC value generated by the ATC counter becomes the same as the ATS value of the source packet, only the TS packet is transferred to the PID filter according to the recording rate of the AV clip.
- the PID filter is a TS packet output from the source depacketizer, and the TS packet PID matches the PID required for playback, according to the PID, the main video video decoder, sub-video video decoder, Transfer to IG decoder, PG decoder, audio decoder, sub audio decoder.
- the main video video decoder when the PID included in the TS packet is 0x1011, the main video video decoder is used.
- the subvideo video decoder is used. If the PID is 0x1400 to 0x141F, the PID is 0x1A00 to 0x1A1F. If the PID is 0x1200 to 0x121F, the PID is transferred to the IG decoder.
- the main video / video decoder is composed of TB (Transport Stream Buffer) 1701, MB (Multiplexing Buffer) 1702, EB (Elementary Stream Buffer) 1703, compressed video decoder 1704, and DPB (Decoded Picture Buffer) 1705.
- TB Transport Stream Buffer
- MB Multiplexing Buffer
- EB lementary Stream Buffer
- DPB Decoded Picture Buffer
- TB 1701 is a buffer that temporarily accumulates TS packets as they are when TS packets including a video stream are output from the PID filter 1702.
- the MB 1702 is a buffer for temporarily storing PES packets when outputting a video stream from the TB 1701 to the EB 1702.
- the TS header of the TS packet is removed.
- EB 1703 is a buffer that stores pictures in an encoded state (I picture, B picture, P picture).
- I picture, B picture, P picture pictures in an encoded state
- Compressed video decoder 1704 each video access unit of the video elementary stream is decoded at every predetermined decoding time (DTS) to create a frame / field image. Since there are MPEG2, MPEG4AVC, VC1 and the like in the compression encoding format of the video stream multiplexed into the AV clip, the decoding method of the compressed video decoder 1704 is switched according to the stream attribute.
- the compressed video decoder 1704 transfers the decoded frame / field image to the DPB 1705, and writes the corresponding frame / field image to the main video plane at the display time (PTS) timing.
- the DPB 1705 is a buffer that temporarily stores the decoded frame / field image.
- the compressed video decoder 1704 decodes a video access unit such as an inter-picture predictive encoded P picture or B picture, it is used to refer to the already decoded picture.
- the sub-video video decoder has the same configuration as the main video video decoder, decodes the input sub-video video stream, and writes the picture to the sub-video plane at the display time (PTS) timing.
- PTS display time
- the IG decoder extracts and decodes an interactive graphics stream from TS packets input from the source packetizer, and writes uncompressed graphics data to the IG plane at the display time (PTS) timing.
- PTS display time
- the PG decoder extracts and decodes the presentation graphics stream from TS packets input from the source packetizer, and writes uncompressed graphics data to the PG plane at the display time (PTS) timing.
- PTS display time
- the main audio decoder has a buffer buffer, and while storing data in the buffer, removes information such as TS header and PES header, decodes the audio stream, and obtains and plays back uncompressed LPCM audio data Output to the audio mixer at the time of the day (PTS). Since there are AC3, DTS, and the like in the compression encoding format of the audio stream multiplexed into the AV clip, the decoding method of the compressed audio is switched according to the stream attribute.
- the sub audio decoder has the same configuration as the main audio decoder, decodes the input sub audio stream, and outputs uncompressed LPCM audio data to the audio mixer at the display time. Since there are DolbyDigitalPlus, DTS-HD LBR, and the like as compression encoding formats of the audio stream multiplexed with the AV clip, the decoding method of the compressed audio is switched according to the stream attribute.
- the audio mixer mixes the uncompressed audio data output from the main audio decoder and the uncompressed audio data output from the sub audio decoder (superimposes the sound) and outputs the result to a speaker or the like.
- the image processor decodes the graphics data (PNG / JPEG) transferred from the program execution unit and outputs it to the image plane according to the display time specified by the program execution unit.
- the advantage of this method is that it is possible to realize stereoscopic viewing by simply preparing two viewpoint images for the right eye and left eye.
- Several technologies have been put into practical use from the viewpoint of being able to show only to the corresponding eyes.
- One is called the time separation method, and the user observes images that are alternately displayed in the time axis direction for the left eye and right eye on the display through continuous stereo glasses (with a liquid crystal shutter).
- the left and right scenes are overlapped by the afterimage reaction of the eyes and recognized as a stereoscopic image.
- the above-mentioned continuous stereo glasses transmit the liquid crystal shutter corresponding to the left eye and the liquid crystal shutter corresponding to the right eye shielded.
- the liquid crystal shutter corresponding to the right eye is made transparent while the liquid crystal shutter corresponding to the left eye is shielded from light.
- this method in order to display the left and right images alternately in the time axis direction, for example, in the case of a normal two-dimensional movie, 24 images are displayed per second, while the left and right images are combined. It is necessary to display 48 images per second. Therefore, this method is suitable for a display in which rewriting of one screen is relatively fast, but is not limited as long as the display can be rewritten a predetermined number of times per second.
- the left and right pictures were output alternately in the time axis direction in the previous time separation method, whereas the left eye picture and right eye picture in the vertical direction in one screen.
- the pixels that make up the picture for the left eye are focused on the left eye only, and the pixels that make up the picture for the right eye are focused only on the right eye
- the left and right eyes can see a picture with parallax and can be viewed as 3D.
- a device having the same function for example, a liquid crystal element may be used, and a vertically polarized filter is used for the left eye pixel, and a horizontally polarized filter is used for the right eye pixel.
- the viewer can view stereoscopically by viewing the display using polarized glasses with a vertically polarized filter for the left eye and a horizontally polarized filter for the right eye.
- the stereoscopic vision using this parallax image is already used generally in amusement park playground equipment and the like, and since it has been established technically, it can be said that it is the closest to practical use at home.
- various techniques such as a non-contact mode color separation method have been proposed as a method for stereoscopic viewing using a parallax image.
- a continuous separation method or a polarized glasses method is used as an example.
- the present invention is not limited to these two methods as long as parallax images are used.
- a method for storing a parallax image used for stereoscopic viewing in an information recording medium will be described.
- the video in which the screen for the left eye is stored is referred to as “left-eye video”
- the video in which the screen for the right eye is stored is referred to as “right-eye video”
- both videos are referred to as “3D video”.
- switching 2D / 3D display a data structure for storing 3D video in a BD-ROM that is a recording medium according to the present invention will be described.
- the basic part of the data structure is the same as the data structure for storing 3D video, and the explanation will focus on the extended or different parts.
- a playback device that can play back only 2D video is described as a 2D playback device
- a playback device that can play back both 2D video and 3D video is described as a 2D / 3D playback device.
- FIG. 98 is a diagram showing an example of an index file in the BD-ROM for reproducing a stereoscopic video.
- a 2D playlist 2601 describing a route for reproducing 2D video and a 3D playlist 2602 describing a route for reproducing stereoscopic video are prepared as playlists.
- the BD program file is displayed in the program when the playback device supports stereoscopic video playback or when the user supports stereoscopic video playback. Check whether video playback is selected and switch the playlist to be played back.
- 3D presence flag and “2D / 3D preference flag” are prepared in the index file.
- the 3D presence flag is a flag for identifying whether or not a playlist for playing 3D video exists on the title. When this is FALSE, the 2D / 3D playback device needs to prepare for stereoscopic video playback. Therefore, it is possible to perform processing at high speed by omitting processing such as HDMI authentication.
- the 2D / 3D preference flag is an identifier indicating whether the content provider wants to play back in 2D or 3D when the player and the television can play back in 2D or 3D. When this flag indicates “3D”, it is not necessary to make a transition to 2D, so HDMI authentication can be performed quickly.
- FIG. 99 shows a selection flow of 2D playlist and 3D playlist in the program of the BD program file.
- S2702 the user is asked whether the user desires 2D video playback or 3D video playback by displaying a menu screen. If 2D video playback is desired as a result of selection by the user's remote control or the like, the 2D playlist is played back. If 3D video playback is desired, the process proceeds to S2703.
- S2703 it is checked whether the display supports 3D video playback.
- the connection is made using HDMI, and the playback apparatus inquires of the display whether or not it supports 3D video playback. If the 3D video playback is not supported, the 2D playlist is played back, but the user may be informed on the menu screen that the television is not ready. When 3D video playback is supported, the 3D playlist is played back.
- the parental level can be set in the PSR13, and it is possible to control so that only a user of an appropriate age or older can play the corresponding BD-ROM disc.
- a 3D parental level PSR30 is prepared for the 2D / 3D playback device. Similar to PSR13, the 3D parental level stores the age of the user who uses the 2D / 3D playback device. It is determined whether or not the BD program file of the title of the BD-ROM disc is reproduced using this PSR30 in addition to PSR13. Since PSR30 is a parental level related to 3D video playback, the parental level is controlled by PSR13 in 2D video playback.
- the play list to be reproduced may be selected with reference to PSR30 after “YES” in S2703.
- the age information is stored in the PSR30 as in the case of the PSR13.
- the PSR30 may be configured to set whether or not to prohibit 3D video.
- information indicating “which user prefers to play back 2D / 3D” is similarly set in the system parameter (here, PSR31).
- PSR31 the user sets which of 2D / 3D is to be played back preferentially through the OSD of the 2D / 3D playback device. If the display supports 3D video playback, using the information in PSR31, if priority is given to 3D video, there is no need to transition to 2D video, so HDMI authentication can be performed quickly. Video playback processing can be performed quickly.
- the BD program refers to the PSR31 and determines whether 2D playback or 3D playback is performed, thereby enabling playback processing that suits the user's preference.
- the BD program may determine the default selection button of the menu displayed by the BD program with reference to the PSR31. For example, in the case of a menu that prompts the branch of “whether to play 2D video” or “play 3D video”, if the value of PSR31 is “2D”, place the cursor on the “Play 2D video” button If the value of PSR31 is “3D”, the cursor is moved to the “3D video playback” button.
- the user currently watching PSR31 may be set according to. For example, assuming that a 2D / 3D playback device is used by three family members (father, mother, and child), the 2D / 3D playback device may manage an account for each user. In this case, by updating the PSR31 according to the logged-in user, it is possible to perform control according to the preference of the person who is actually using it.
- information indicating “which user prefers to play back 2D / 3D” may be provided with a level instead of 2D or 3D. For example, it may be divided into four stages of “always 2D”, “somewhat 2D”, “somewhat 3D”, and “always 3D”. In this way, the playback process of the 2D / 3D playback device can be performed so as to better suit the user's preference. For example, in a case where PSR25, which is a system parameter indicating the state of the playback device, is set, if PSR31 is “always 2D”, PSR25 is always set to 2D mode, and if PSR31 is “always 3D”, PSR25 is always set. Also, the 3D mode (L / R mode or DEPTH mode) is set.
- the caption entry displayed as shown in FIG. 81 is composed of a plurality of caption data entries.
- Each caption data entry is composed of composition information, window information, palette information, and object information.
- the composition information is information that defines the screen configuration of the caption data.
- the cropping information of the object, the display position of the cropped object, the window ID that identifies the window to be referenced, the palette ID that identifies the palette to be referenced, and the palette that is referenced Is stored.
- the window information a window area that defines an area to be drawn by the decoder is stored together with the window ID.
- a graphics image is stored together with the object ID.
- the graphics image is image data composed of 256 index colors, and is compressed using a compression method such as run length compression.
- color table information (CLUT) used for the object is stored together with the palette ID.
- the table can store 256 colors, and each color can be referred to by a color ID.
- the color ID takes a value from 0 to 255, but 255 is assigned a fixed colorless and transparent.
- FIG. 82 is a diagram for explaining the flow of the presentation graphics decoding process.
- each caption data entry decodes a compressed graphics image designated by the reference object ID of the composition information.
- the graphics image is cropped using the cropping information of the composition information, and only necessary data is cut out.
- the display position on the graphics plane is determined according to the display position of the composition information.
- the object data is drawn on the graphics plane for the range of the window area of the window information specified by the reference window ID of the composition information.
- the graphics plane is colored and displayed using the palette information specified by the reference palette ID of the composition information. Note that the display timing is the PTS timing of the PES packet in which the composition information is stored.
- black frames that are not used for the main movie of the movie are used on the screen, the black frame regions are collected on the upper side or the lower side, and subtitle data is displayed there.
- the black frame inserted into the video stream is a useless area, subtitles may be displayed here. However, as shown in the right diagram of FIG. Therefore, it is a little small to put subtitle data. Therefore, as shown in FIGS. 4B and 4C, the main video is shifted up or down, and black is inserted in the shifted area, so that the upper and lower black frames are combined up or down. By doing so, it is possible to prepare a black frame large enough to contain subtitle data.
- FIG. 83 shows the structure of a play item of a 3D playlist.
- the stream additional information 1311 of the PG stream in the stream selection table includes “shift value when moving up video (PG_v_shift_value_for_Up)” and “shift value when moving up video ( PG_v_shift_value_for_Down) ”.
- “Shift value when moving up video (PG_v_shift_value_for_UP)” is the shift amount of the PG plane when the main video plane is shifted upward (when the black frame is gathered downward), and “Shift value when moving up video (PG_v_shift_value_for_Down) ) ”Is the shift amount of the PG plane when the main video plane is shifted downward (when the black frame is gathered on the lower side).
- the 2D / 3D playback device adjusts the shift amount of the PG plane based on this value. Details of the plane synthesis method will be described later.
- FIG. 84 shows the configuration of plane synthesis in the 2D / 3D playback device.
- a PG plane will be described as an example, but the present invention can be applied to any plane such as a sub-picture plane, an IG plane, and an image plane.
- a video plane cropping unit 9701 that performs cropping processing of the 2D / left-eye video plane and right-eye video plane, PSR32 that describes video shift mode, and cropping processing of the PG plane are performed.
- a PG plane cropping unit 9702 to perform, and PSR33 in which the shift amount of each plane such as a PG plane is described are included.
- PSR32 in (a) of FIG. 25 is a system parameter of the 2D / 3D playback device indicating the video shift mode (video_shift_mode).
- the video_shift_mode of PSR32 has three modes of Keep, Up, and Down. If the value of PSR32 is 0, it is Keep, if the value of PSR32 is 1, it indicates Up, and if the value of SRPM (32) is 2, it indicates Down. .
- the video plane cropping unit 9701 performs video plane cropping processing according to the video shift mode described in the PSR32.
- the value of PSR32 is set through an API or command of the BD program.
- the video plane cropping unit 9701 does not change the 2D / left-eye video plane and the right-eye video plane, and does not change the other planes. Proceed to superimposition processing.
- PSR32 indicates Up, as shown in (2) of FIG. 25 (b)
- the video plane cropping unit 9701 shifts the 2D / left-eye video plane and the right-eye video plane upward from each other. Then, the black frame in the upper region is cut out, the black frame is inserted in the lower region, and the process proceeds to the superimposing process with the plane. As a result, the black frame can be collected downward in the plane.
- PSR32 indicates Down, as shown in (3) of FIG.
- the video plane cropping unit 9701 shifts the 2D / left-eye video plane and right-eye video plane downward. Then, the black frame in the lower region is cut out, the black frame is inserted in the upper region, and the process proceeds to the superimposing process with the plane. As a result, the black frame can be collected upward in the plane.
- FIG. 36 shows system parameters (in this case, PSR33) indicating the shift amount of each plane in the vertical axis direction.
- the shift amount of PSR33 is prepared for each plane as a plane shift amount when the video shifts upward and a plane shift amount when the video shifts downward.
- PG_shift_value_for_UP and PG_shift_value_for_Down are prepared for the PG plane.
- the value of PSR33 is updated by the values (PG_v_shift_value_for_Up, PG_v_shift_value_for_Down) set in the playlist by stream switching.
- the PSR 33 may be set through an API or command of the BD program.
- the PG plane cropping unit 9702 shown in FIG. 84 performs plane shift according to the shift amount of the PG plane of PSR33.
- the shift process of the PG plane cropping unit 9702 and the synthesis process with the video plane are described in FIG. 11 and FIG.
- FIG. 11 when the video_shift_mode of PSR32 is Keep, the PG plane cropping unit 9702 performs the synthesis process with the video plane as it is without performing the process.
- the PG plane cropping unit 9702 uses the value of PG_shift_value_for_Up stored in the PSR33 to perform the PG plane shift process, Cut out and superimpose on video plane.
- the subtitle display position can be taken lower than in 2D playback, and the subtitle can be displayed at an appropriate position in a black frame area formed on the lower side.
- the PG plane cropping unit 9702 uses the value of PG_shift_value_for_Down stored in PSR33 to perform the PG plane shift process, Cut out and superimpose on video plane. By doing in this way, the display position of the subtitle can be brought to the upper side than during 2D playback, and the subtitle can be displayed at an appropriate position in the black frame area formed on the upper side.
- the offset processing (1 plane + offset method) in the horizontal axis direction for controlling popping out is omitted, but a mechanism of cropping processing by the offset value in the horizontal axis direction may be added. In this way, even a subtitle displayed in a black frame can be popped out.
- the offset value in the horizontal axis direction may be fixed.
- the shift amount in the x-axis direction is defined in the additional information of FIG. 83, the value is stored in PSR in the same way as PSR33, and the offset process in the horizontal axis direction is performed using that value. Also good. This makes it easy to create data.
- the shift amount in the Y-axis direction is stored in PSR33.
- the PG plane cropping unit 9702 directly refers to the playlist without providing system parameters. Good.
- video plane shift amount when video_shift_mode is Up or Down may be fixed at the size of the upper and lower black frames (131 pixels in the example of FIG. 4), but may be set as desired by the author or the user.
- a new system parameter may be prepared, the shift amount stored therein, and set by a BD program or a player OSD.
- the process of shifting the entire plane using the value stored in PSR33 has been described.
- the value added to the display position stored in the PG composition information It may be used as For example, when the display position of the PG composition information is (x, y) and video_shift_mode is Keep, the PG decoder displays the corresponding caption data entry at the position (x, y + PG_shift_value_for_UP). By configuring in this way, the processing can be made lighter than shifting the plane.
- PG_shift_value_for_UP may be stored in the composition information.
- PG restrictions include that the display position does not exceed the above area, that the size of the object to be displayed at the display position does not exceed the above area, that the window display position does not exceed the above area, The window size plus the display position does not exceed the above area. By restricting in this way, it is possible to prevent subtitles from being displayed.
- video_shift_mode may be added to the stream additional information 1311 of the stream selection information.
- the configuration of the plane synthesis process of the 2D / 3D playback device is as shown in FIG.
- PSR34 is added. This stores a flag indicating On / Off of whether or not to perform video shift. That is, if PSR34 is 1, it means that video shift is performed, and if it is 0, it means that video shift is not performed.
- the video shift On / Off flag PSR34 is controlled by the program execution unit using a menu or the like. It may be set by a user operation by the player's OSD or the like.
- the video shift mode is stored in the PSR 32, and this value is set based on the additional information of the selected subtitle stream by the selection of the PG stream.
- the video plane cropping unit 9701 performs a video plane cropping process based on the video_shift_mode set in the PSR 32.
- PSR34 is Off, the cropping process is not performed. With this configuration, it is possible to set an appropriate video_shift_mode for each subtitle.
- the remote control of the 2D / 3D playback device generally has a caption switching button, and each time the user presses the caption switching button, a user operation is defined that switches sequentially in the order of the caption streams registered in the stream selection table. Has been.
- the order of the PG streams registered in the stream selection table is arranged so that the same video_shift_mode is continuous as shown in FIG.
- the PG plane shift process is preferably applied after the video plane shift is completed.
- the method of dynamically collecting the black frame in the upper or lower area has been described.
- the main video is not located in the middle but is arranged slightly above to create a video stream.
- a large number of black frames may be arranged on the lower side, and this area may be used for subtitle display. In this way, the black frame does not have to be dynamically changed for subtitle display, and the video does not move up and down, so that the user does not feel uncomfortable.
- Black frame subtitles may be created.
- the value of PSR37 can be set on the menu screen of the BD program.
- FIG. 100 a configuration example (FIG. 100) realized by using the integrated circuit 3 will be described with respect to a reproducing apparatus that reproduces data having the structure described in the above-described embodiments.
- the medium IF unit 1 receives (reads) data from the medium and transfers it to the integrated circuit 3.
- the medium IF unit 1 receives data having the structure described in the above-described embodiment from the medium.
- the medium IF unit 1 includes, for example, a disk drive when the medium is an optical disk or a hard disk, a card IF when the medium is a semiconductor memory such as an SD card or a USB memory, and a CAN tuner when the medium is a broadcast wave including CATV. If the Si tuner or the medium is a network such as Ethernet, wireless LAN, or wireless public line, it is a network IF.
- the memory 2 is a memory for temporarily storing (reading) data received from a medium or temporarily storing data being processed in the integrated circuit 3.
- SDRAM Synchronous Dynamic Random Access Memory
- DDRx SDRAM Double -Date-Ratex
- Memory; x 1,2,3 ...) etc. are used. Any number of memories 2 may be provided, and a single or a plurality of memories 2 may be provided as necessary.
- the integrated circuit 3 is a system LSI that performs video / audio processing on the data transferred from the medium IF unit 1.
- the main control unit 6 has a processor core having a timer function and an interrupt function, and the processor core controls the entire integrated circuit 3 according to a program stored in a program memory or the like.
- the program memory or the like stores basic software such as an OS in advance.
- the stream processing unit 5 receives data transferred from the medium via the medium IF unit 1 under the control of the main control unit 6, and stores or receives the data in the memory 2 via the data bus in the integrated circuit 3.
- the obtained data is separated into video data and audio data.
- 2D / L AV clips containing left-view video streams and R AV clips containing right-view video streams are alternately arranged in several extents as described above. It is a thing. Therefore, when the integrated circuit 3 receives the left-eye data including the left-view stream, the main control unit 6 uses the right-eye video stream including the right-view video stream so that the data is stored in the first area of the memory 2. When data is received, control is performed so that the data is stored in the second area of the memory 2.
- the left-eye data belongs to the left-eye extent
- the right-eye data belongs to the right-eye extent.
- the first and second areas in the memory 2 may be a single memory logically divided into areas, or may be physically different memories.
- the description will be continued with the left-eye data including the left-view video stream as the main view data and the right-eye data including the right-view video stream as the sub-view data.
- the right-eye data is the main view data and the left-eye data.
- the data may be subview data.
- the graphics stream is multiplexed with one or both of main view data and subview data, or exists as a single stream.
- the signal processing unit 7 decodes the video system data and audio system data separated by the stream processing unit 5 by an appropriate method under the control of the main control unit 6.
- Video data is encoded and recorded using methods such as MPEG-2, MPEG-4 AVC, MPEG4-MVC, SMPTE VC-1, and audio data is Dolby AC-3, Dolby Digital Plus, Since the data is compressed, encoded and recorded by a method such as MLP, DTS, DTS-HD, or linear PCM, the signal processing unit 7 decodes the signal using a method corresponding to these.
- the model of the signal processing unit 7 corresponds to, for example, various decoders in FIG. 16 of the first embodiment.
- the memory control unit 9 arbitrates access to the memory 2 from each functional block in the integrated circuit 3.
- the AV output unit 8 superimposes the video data decoded by the signal processing unit 7 under the control of the main control unit 6, converts the format of the video system and data, and outputs the data outside the integrated circuit 3.
- FIG. 101 is a functional block diagram showing a typical configuration of the stream processing unit 5.
- the stream processing unit 5 includes a device stream IF unit 51, a demultiplexing unit 52, a switching unit 53, and the like.
- the device stream IF unit 51 is an interface for data transfer between the medium IF unit 1 and the integrated circuit 3.
- the medium is an optical disk or a hard disk
- SATA Serial Advanced Technology Attachment
- ATAPI Advanced Technology Attachment Packet Interface
- PATA Parallel Advanced Technology Attachment
- card IF if the medium is a semiconductor memory such as SD card or USB memory
- tuner IF if the medium is a broadcast wave including CATV, etc.
- the medium is Ethernet, wireless LAN, wireless
- a network such as a public line, it is a network IF.
- the device stream IF unit 51 may take over some of the functions of the medium IF unit 1, or the medium IF unit 1 may be built in the integrated circuit 3.
- the demultiplexing unit 52 separates playback data including video / audio transferred from the medium into video data and audio data.
- Each extent described above is composed of source packets such as video, audio, PG (caption), and IG (menu) (however, subview data may not include audio), and is included in each source packet.
- PID identifier
- it is separated into video and audio TS packets and transferred to the signal processing unit 7.
- the processed data is directly or once stored in the memory 2 and then transferred to the signal processing unit 7.
- the model of the demultiplexing unit 52 corresponds to, for example, the source depacketizer and the PID filter in FIG. 79 of the eighth embodiment.
- a graphics stream that exists as a single stream that is not multiplexed with the main view data or the subview data is transferred to the signal processing unit 7 without passing through the demultiplexing unit 52.
- the switching unit 53 stores data in the first area of the memory 2 when the device stream IF unit 51 receives the data for the left eye, and stores the data in the second area of the memory 2 when the data for the right eye is received.
- the output destination (storage destination) is switched so as to be stored.
- the switching unit 53 is, for example, a DMAC (Direct Memory Memory Access Controller).
- FIG. 102 is a conceptual diagram around the switching unit 53 when the switching unit 53 is a DMAC.
- the DMAC transmits the data received by the device stream IF unit and its data storage destination address to the memory control unit 9 under the control of the main control unit 6.
- the device stream IF unit receives left-eye data
- memory control is performed for address 1 (first storage area)
- address 2 second storage area
- the output destination is switched depending on the received data.
- the memory control unit 9 stores data in the memory 2 in accordance with the storage destination address sent from the DMAC.
- a dedicated circuit for controlling the switching unit 53 may be provided instead of the main control unit 6.
- the device stream IF unit 51, the demultiplexing unit 52, and the switching unit 53 have been described.
- a cryptographic engine unit that decrypts received encrypted data, key data, and the like may further include an execution control such as a device authentication protocol between the medium and the playback device, a secure management unit that holds a secret key, a controller for direct memory access, and the like. So far, when the data received from the medium is stored in the memory 2, the switching unit 53 switches the storage destination depending on whether the data for the left eye or the data for the right eye has been described. However, the data received from the medium is temporarily stored in the memory 2. When the data is transferred to the demultiplexing unit 52, the left-eye data and the right-eye data may be distributed.
- FIG. 103 is a functional block diagram showing a typical configuration of the AV output unit 8.
- the AV output unit 8 includes an image superimposing unit 81, a video output format converting unit 82, an audio / video output IF unit 83, and the like.
- the image superimposing unit 81 superimposes the decoded video data. Specifically, left-view video data or right-view video data, and PG (caption) and IG (menu) are superimposed on a picture basis.
- the model of the image superimposing unit 81 is, for example, the first embodiment, FIG. 20 to FIG. More specifically, the decoded video data and subtitle data are stored in an area for storing data drawn on each plane in the memory 2.
- the plane may be an area in the memory 2 or a virtual space.
- the image superimposing unit 81 superimposes a caption plane corresponding to the left view plane and a caption plane corresponding to the right view plane.
- the left view plane and the right view plane are superimposed on the caption data so that the caption data is superimposed on the display area of the caption data indicated by the area securing flag based on the area securing flag corresponding to the caption data (stream) to be superimposed.
- the first embodiment, FIG. 12, etc. That is, when the display area of the caption data indicated by the area reservation flag is at the upper end, the left view plane and the right view plane are shifted downward in the vertical coordinate and superimposed with the caption data, and the caption data indicated by the area reservation flag is displayed. When the display area is at the lower end, the left view plane and the right view plane are shifted upward in the vertical coordinates and superimposed with the caption data.
- the video output format conversion unit 82 performs resize processing for enlarging or reducing the decoded video data, IP conversion processing for converting the scanning method from one of the progressive method and the interlace method, and noise reduction for removing noise. Processing, frame rate conversion processing for converting the frame rate, and the like are performed as necessary.
- the audio / video output IF unit 83 performs encoding or the like according to the data transmission format with respect to the video data subjected to image superposition or format conversion and the decoded audio data. As will be described later, the audio / video output IF unit 83 may be partially provided outside the integrated circuit 3.
- FIG. 104 is a configuration example showing the AV output unit 8 or the data output part of the playback device in more detail.
- the integrated circuit 3 and the playback device in this embodiment are compatible with a plurality of video data and audio data transmission formats.
- 103 corresponds to the analog video output IF unit 83a, the analog audio output IF unit 83c, and the digital audio / output IF unit 83b.
- the analog video output IF unit 83a converts and encodes the video data subjected to the image superimposition process and the output format conversion process into an analog video signal format, and outputs it.
- a composite video encoder, encoder for S video signal (Y / C separation), encoder for component video signal, DAC (D / A converter), etc. corresponding to any of the three systems of NTSC, PAL and SECAM .
- the digital audio / video output IF unit 83b integrates the decoded audio system data with the video system data that has undergone image superimposition processing or output format conversion, further encrypts it, encodes it according to the data transmission standard, and outputs it .
- HDMI High-Definition Multimedia InterFace
- the analog audio output IF unit 83c corresponds to an audio DAC that D / A converts the decoded audio data and outputs analog audio data.
- the transmission format of these video data and audio data can be switched depending on the data reception device (data input terminal) supported by the display device / speaker 4 side, or the transmission format can be switched by user selection. It is. Furthermore, it is possible to transmit data corresponding to the same content not only in a single transmission format but also in a plurality of transmission formats in parallel.
- the image superimposing unit 81, the video output format converting unit 82, and the audio / video output IF unit 83 have been described as typical configurations of the AV output unit 8, but filter processing, screen synthesis, curve drawing, 3D display, etc.
- a graphics engine unit or the like for performing the graphics processing may be further included.
- the processing of the playback device in the present embodiment may be performed in the display device.
- the data received by the medium IF unit 1 is signal-processed by the integrated circuit 3
- the processed video data is output through the display panel 11 via the display driver 10
- the processed audio data is The sound is output through the speaker 12.
- the AV output unit 8 is configured as shown in FIG. 108, for example, and data transfer is performed through the audio output IF unit 85 and the video output IF unit 84 inside or outside the integrated circuit 3.
- a plurality of audio output IF units 85 and video output IF units 84 may be provided, or an IF unit common to video and audio may be provided.
- control bus and data bus paths in the integrated circuit 3 are arbitrarily arranged according to the processing procedure and processing contents of each processing block.
- the data bus is connected directly to each processing block.
- the data bus may be arranged so that the processing blocks are connected to each other via the memory 2 (memory control unit 9).
- the integrated circuit 3 may be a multi-chip module in which a plurality of chips are sealed in one package and apparently formed into one LSI.
- an FPGA Field Programmable Gate Array
- a reconfigurable processor that can reconfigure the connection and setting of circuit cells inside the LSI may be used.
- FIG. 109 is a flowchart simply showing a playback operation procedure in which data is received (read) from a medium, decoded, and then output as a video signal and an audio signal.
- S1 Receive (read) data from the medium (medium IF unit 1, stream processing unit 5).
- S2 The data received (read) in S1 is separated into various data (video data / audio data) (stream processing unit 5).
- S3 Various data separated in S2 are decoded in an appropriate format (signal processing unit 7).
- S6 The video data and audio data processed in S2 to S5 are output (AV output unit 8).
- FIG. 110 is a flowchart showing the playback operation procedure in more detail. Each operation / process is performed under the control of the main control unit 6.
- the device stream IF unit 51 of the stream processing unit 5 receives data (PLAYLIST, CLIPINF, etc.) necessary for reproducing data other than the reproduced data stored in the medium through the medium IF unit 1. (Read) and store in the memory 2 (medium IF unit 1, device IF unit 51, memory control unit 9, memory 2).
- data PLAYLIST, CLIPINF, etc.
- the main control unit 6 recognizes the compression format of the video data and audio data stored in the medium from the stream attribute information included in the received CLIPINF, and performs a corresponding decoding process so that the signal processing unit 7 can perform the corresponding decoding process. Is initialized (main control unit 6).
- the device stream IF unit 51 of the stream processing unit 5 receives (reads out) data to be reproduced such as video / audio stored in the medium through the medium IF unit 1, and causes the switching unit 53 and the memory control unit 9 to And stored in the memory 2.
- data is received (read) in extent units, and when the left-eye data is received (read), the first area is used. When the right-eye data is received (read), the second area is used.
- the main control unit 6 controls the switching unit 53 so that the data is stored, and the switching unit 53 switches the output destination (storage destination) of the data (the medium IF unit 1, the device IF unit 51, the main control unit 6, and the switching unit 53). , Memory control unit 9, memory 2).
- S104 The data stored in the memory 2 is transferred to the demultiplexing unit 52 of the stream processing unit 5, which demultiplexes the video system (main video, sub video, It recognizes whether it is PG (caption), IG (menu)), or audio system (audio, sub-audio), and transfers it to each corresponding decoder of the signal processing unit 7 in units of TS packets (demultiplexing unit 52).
- the video system main video, sub video, It recognizes whether it is PG (caption), IG (menu)), or audio system (audio, sub-audio
- Each decoder of the signal processing unit 7 performs a decoding process on the transferred TS packet by an appropriate method (signal processing unit 7).
- S106 The data corresponding to the left-view video stream and the right-view video stream among the video data decoded in the signal processing unit 7 is resized according to the display device 4 (video output format conversion unit 82).
- S107 The video stream resized in S106, PG (caption) / IG (menu), and text subtitle are superimposed (image superimposing unit 81).
- IP conversion which is conversion of the scanning method is performed on the video data superimposed in S107 (video output format conversion unit 82).
- S109 Encoding, D / A conversion, etc. according to the data output method of the display device / speaker 4 or the data transmission method to the display device / speaker 4 with respect to the video system data and audio system data that have been processed so far I do.
- the video data and audio data are processed to correspond to analog or digital output, respectively.
- analog output of video data composite video signals, S video signals, component video signals, and the like are supported.
- the digital output of video / audio data supports HDMI (audio / video output IF unit 83).
- S110 The video data and audio data processed in S109 are transmitted to and output to the display device / speaker 4 (audio / video output IF unit 83, display device / speaker 4).
- the above is the description of the operation procedure of the playback device in the present embodiment.
- the processing result may be temporarily stored in the memory 2 for each processing.
- the operation procedure is basically the same, and the functional blocks corresponding to the functional blocks of the playback device of FIG. 100 perform the same process.
- the process may be omitted if necessary, and other processes (noise reduction process, Frame rate conversion processing or the like may be performed. Further, the processing procedure may be changed for possible ones.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Television Signal Processing For Recording (AREA)
Abstract
Description
メインビュー映像とサブビュー映像との視差により立体視効果を実現する3D映像では、表示装置の画面サイズに応じて視差が変化してしまうため、画面サイズの違いによって画像の奥行きにも違いが生じる。そのため、画面サイズが大きい表示装置での視聴を想定して制作された3D映像が、画面サイズが小さい表示装置で視聴されると、製作者が意図したものより奥行き感の乏しい迫力のない映像となってしまう。逆に画面サイズが小さな表示装置での視聴を想定して制作された3D映像が、画面サイズが大きい表示装置で視聴されると、過剰な奥行き感が生じ眼精疲労を引き起こす恐れがある。
メインビュービデオストリームと、サブビュービデオストリームと、メタデータとが記録された記録媒体であって、前記メインビュービデオストリームは、立体視映像のメインビューを構成するピクチャデータを含み、前記サブビュービデオストリームは、立体視映像のサブビューを構成するピクチャデータを含み、前記メタデータは、複数の表示装置の画面サイズ情報に対応するオフセット補正値を含み、前記オフセット補正値は、前記メインビューを構成するピクチャデータが描画されるメインビュービデオプレーン及び前記サブビューを構成するピクチャデータが描画されるサブビュービデオプレーンの少なくとも一方を、水平座標の右方向及び左方向へ移動させるオフセットを規定することを特徴とする。
第1実施形態の特徴は、右目用のグラフィクスストリーム、左目用のグラフィクスストリームを用いた立体視を実現することができるかどうかのケーパビリティを、再生装置におけるレジスタに設定している点である。
本図の第2段目は、ベースビュービデオストリームの内部構成を示す。このストリームには、ピクチャタイプがI1,P2,Br3,Br4,P5,Br6,Br7,P9というビューコンポーネントが含まれている。これらのビューコンポーネントは、Decode Time Stamp(DTS)に従いデコードされる。第1段目は、左目画像を示す。そうしてデコードされたビューコンポーネントI1,P2,Br3,Br4,P5,Br6,Br7,P9をPTSに従い、I1,Br3,Br4,P2,Br6,Br7,P5の順序で再生することで、左目画像が再生されることになる。
インデックステーブルは記録媒体全体に関する管理情報であり、再生装置への記録媒体挿入後に、インデックステーブルが最初に読み出されることで、再生装置において記録媒体が一意に認識される。
動作モードオブジェクトのプログラムファイルは、再生装置を動作させるための制御プログラムを格納している。このプログラムには、コマンドによって記述されたものと、オブジェクト指向のコンパイラ言語によって記述されたものがある。前者のプログラムは、コマンドベースの動作モードにおいて、複数のナビゲーションコマンドをバッチジョブとして再生装置に供給し、これらナビゲーションコマンドに基づき再生装置を動作させる。このコマンドベースの動作モードを、"HDMVモード"と呼ぶ。
ストリームファイルは、ビデオストリーム、1つ以上のオーディオストリーム、グラフィクスストリームを多重化することで得られたトランスポートストリームを格納している。ストリームファイルには、2D専用のものと、2D-3D兼用のものとがある。2D専用のストリームファイルは、通常のトランスポートストリーム形式であり、2D-3D兼用のストリームファイルは、立体視インターリーブドストリームファイルのファイル形式を有する。
ストリーム情報ファイルは、ストリームファイルにおけるトランスポートストリーム内の任意のソースパケットに対するランダムアクセスや、他のトランスポートストリームとの途切れ無き再生を保障するストリーム情報ファイルである。このストリーム情報ファイルを通じることにより、ストリームファイルは"AVクリップ"として管理されることになる。ストリーム情報ファイルは、AVクリップにおけるストリームの符号化形式、フレームレート、ビットレート、解像度等の情報や、GOPの先頭位置のソースパケット番号を、フレーム期間のプレゼンテーションタイムスタンプと対応付けて示す基本エントリーマップをもっているので、ストリームファイルのアクセスに先立ち、このストリーム情報ファイルをメモリにロードしておけば、アクセスしようとするストリームファイル中のトランスポートストリームがどのようなものであるのかを把握することができるので、ランダムアクセスの実行を保障することができる。ストリーム情報ファイルには、2Dストリーム情報ファイルと、3Dストリーム情報ファイルとがあり、3Dストリーム情報ファイルは、ベースビューのためのクリップ情報(クリップベース情報)と、ディペンデントビューのためのクリップ情報(クリップディペンデント情報)と、立体視用の拡張エントリーマップとを含む。
プレイリスト情報ファイルは、再生装置にプレイリストを再生させるための情報を格納したファイルである。"プレイリスト"とは、トランスポートストリーム(TS)の時間軸上で再生区間を規定するとともに、この再生区間同士の再生順序を論理的に指定することで規定される再生経路であり、TSのうち、どれをどの部分だけ再生して、どのような順序でシーン展開してゆくかを規定する役割をもつ。プレイリスト情報は、かかるプレイリストの"型"を定義する。プレイリスト情報によって定義される再生経路は、いわゆる"マルチパス"である。マルチパスとは、主となるTSに対して定義された再生経路(メインパス)と、従となるストリームに対して定義された再生経路(サブパス)とを束ねたものである。このマルチパスにおいてベースビュービデオストリームの再生経路を規定し、サブパスにおいてディペンデントビュービデオストリームの再生経路を規定すれば、立体視を再生するためのビデオストリームの組合せを、好適に規定することができる。
図5(b)は、メインTSの内部構成を示し、同図(c)は、サブTSの内部構成を示す。同図(b)に示すように、メインTSは、1本のベースビュービデオストリームと、32本の左目PGストリーム、32本の左目用IGストリーム、32本のオーディオストリームを含むものとする。同図(c)に示すように、サブTSは、1本のディペンデントビュービデオストリームと、32本の右目PGストリーム、32本の右目用IGストリームを含むものとする。
ベースビュービデオストリームは、ピクチャインピクチャアプリケーションにおけるプライマリビデオストリームを構成する。ピクチャインピクチャアプリケーションは、このプライマリビデオストリームの他、セカンダリビデオストリームから構成される。プライマリビデオストリームとは、ピクチャインピクチャアプリケーションにおいて親画面となるピクチャデータから構成されるビデオストリームである。対照的に、セカンダリビデオストリームとは、ピクチャインピクチャにおいて子画面として、親画面の一部にはめ込まれるピクチャデータから構成されるビデオストリームである。
オーディオストリームには、プライマリオーディオストリーム、セカンダリオーディオストリームの2種類がある。プライマリオーディオストリームは、ミキシング再生を行う場合、主音声となるべきオーディオストリームであり、セカンダリオーディオストリームは、ミキシング再生を行う場合、副音声をとなるべきオーディオストリームである。セカンダリオーディオストリームは、このミキシングのためのダウンサンプリングのための情報、ゲイン制御のための情報が存在する。
PGストリームは、デコーダにパイプラインを採用することで、映像との緻密な同期を実現することができ、字幕表示に適したグラフィクスストリームであり、2DPGストリームと、立体視PGストリームという2つの種類がある。立体視PGストリームには、左目PGストリーム及び右目PGストリームという二種類のものがある。
IGストリームは、対話操作の情報を具備することで、ビデオストリームの再生進行に伴ってメニューを表示したり、またユーザ操作に従いポップアップメニューを表示することができるグラフィクスストリームである。
3D表示モードのどれが許容されるかは、ストリーム種別によって異なる。プライマリビデオストリームの3D表示モードには、B-Dプレゼンテーションモード、B-Bプレゼンテーションモードといった2つの出力モードが許容される。プライマリビデオストリームにおいて、B-Bプレゼンテーションモードが許容されるのは、ポップアップメニューがオンになっている場合のみである。B-Dプレゼンテーションモードで再生される場合におけるプライマリビデオストリームの類型を、"立体視B-D再生タイプ"という。B-Bプレゼンテーションモードで再生される場合におけるプライマリビデオストリームの類型を、立体視B-B再生タイプという。
3)プレイリストマーク情報は、再生区間固有のマークポイントを定義する情報であり、再生区間を示す参照子と、デジタルストリームの時間軸において、マークポイントが何処にあるかを示すタイムスタンプと、マークポイントの属性を示す属性情報とを含み、
前記属性情報は、プレイリストマーク情報により定義されたマークポイントが、リンクポイントであるか、エントリーマークであるかを示す。
<基本ストリーム選択テーブル(STreamNumber_table)>
前記基本ストリーム選択テーブルは、プレイリストを構成する複数のプレイアイテムのうち、その基本ストリーム選択テーブルを包含しているのものがカレントプレイアイテムになった際、マルチパスのメインパスにて参照されているAVクリップに多重化されているES、及び、マルチパスのサブパスにて参照されているAVクリップに多重化されているESのうち、どれの再生を許可するかを、複数のストリーム種別毎に規定するテーブルである。ここでのストリーム種別とは、ピクチャインピクチャにおけるプライマリビデオストリーム、ピクチャインピクチャにおけるセカンダリビデオストリーム、サウンドミキシングにおけるプライマリオーディオストリーム、サウンドミキシングにおけるセカンダリオーディオストリーム、PG_テキスト字幕ストリーム、IGストリームといった種別をいい、基本ストリーム選択テーブルは、これらのストリーム種別毎に、再生を許可すべきストリームを登録することができる。具体的には、基本ストリーム選択テーブルは、ストリーム登録の配列から構成される。ここでストリーム登録とは、基本ストリーム選択テーブルが帰属しているプレイアイテムがカレントプレイアイテムになった際、再生を許可すべきESがどのようなストリームであるかを、そのストリーム番号に対応付けて示すものであり、ストリーム登録は、論理的なストリーム番号に、ストリームエントリー及びストリーム属性の組合せを対応付けるというデータ構造になっている。
拡張ストリーム選択テーブルは、立体視出力モードにおいてのみ、ストリーム選択テーブルと共に使用されるストリーム選択テーブルであり、プレイアイテムの再生や、これに関連するサブパスが再生されている際、選択することができるESを定義する。
『B-Bモード時のオフセット値情報(IG_Plane_offset_value_during_BB_video)』は、B-BプレゼンテーションモードでポップアップメニューのユーザインターフェイスによってIGストリームが再生されている間、1plane+OffsetモードにおけるIGプレーンのオフセット値を画素単位で指示する。
エポックスタートのディスプレイセットとは、グラフィクスデコーダにおけるコンポジションバッファ、コードデータバッファ、グラフィクスプレーンをリセットして、メモリ管理を開始させる機能セグメントの集まりであり、画面構成に必要な機能セグメントを全て含んでいる。
ノーマルケースのディスプレイセットとは、グラフィクスデコーダにおけるコンポジションバッファ、コードデータバッファ、グラフィクスプレーンのメモリ管理を継続したまま画面構成を行うディスプレイセットであり、先行するディスプレイセットからの差分となる機能セグメントを含んでいる。
アクジッションポイントのディスプレイセットとは、画面構成に必要な機能セグメントを全て含むディスプレイセットであるが、グラフィクスデコーダにおけるコンポジションバッファ、コードデータバッファ、グラフィクスプレーンのメモリ管理をリセットさせないディスプレイセットである。このアクジッションポイントのディスプレイセットには、前のディスプレイセットとは異なる内容の機能セグメントが存在してもよい。
エポックコンティニューのディスプレイセットとは、PGストリームの再生を許可しているプレイアイテムと、その直前のプレイアイテムとの接続形態が、クリーンブレークを伴うシームレス接続(CC=5)である場合、再生装置におけるコンポジションバッファ、コードデータバッファ、オブジェクトバッファ、グラフィクスプレーンにおけるメモリ管理を、そのまま継続させる旨を示す。この際、オブジェクトバッファ、グラフィクスプレーン上に得られたグラフィクスオブジェクトは、廃棄されることなく、オブジェクトバッファ、グラフィクスプレーン上で存続する。
オブジェクト定義セグメントは、グラフィクスオブジェクトを定義する機能セグメントである。グラフィクス定義セグメントは、コード値と、そのコード値のランレングスとを用いることで、グラフィクスオブジェクトを定義している。
パレット定義セグメントは、各コード値と、輝度、赤色差・青色差との対応関係を示したパレットデータを含む。左目用グラフィクスストリームのパレット定義セグメントと、右目用グラフィクスストリームのパレット定義セグメントとでは、コード値と、輝度及び色差との対応関係が同一の内容に設定されている。
ウィンドゥ定義セグメントは、非圧縮のグラフィクスオブジェクトを画面上に展開するためのプレーンメモリにおいて、ウィンドゥと呼ばれる矩形枠を定義する機能セグメントである。グラフィクスオブジェクトの描画は、このプレーンメモリの内部で制限されており、このウィンドゥの外部では、グラフィクスオブジェクトの描画は行えない。
画面構成セグメントは、グラフィクスオブジェクトを用いた画面構成を規定する機能セグメントであり、グラフィクスデコーダにおけるコンポジションコントローラに対する複数の制御項目を含む。画面構成セグメントは、グラフィクスストリームにおけるディスプレイセットの詳細を規定すると共に、グラフィクスオブジェクトを用いた画面構成を規定する機能セグメントである。かかる画面構成には、Cut-In/Out、Fade-In/Out、Color Change、Scroll、Wipe-In/Outといったものがあり、画面構成セグメントによる画面構成を伴うことにより、ある字幕を徐々に消去しつつ、次の字幕を表示させるという表示効果が実現可能になる。
1つのディスプレイセットに属する複数の機能セグメントの最後尾に位置する機能セグメントである。再生装置は、画面構成セグメントからこのエンドセグメントまでが、1つのディスプレイセットを構成する機能セグメントであるとして解釈する。
PGデコーダは、PGストリームから読み出されう機能セグメントを格納する「コーデッドデータバッファ」と、画面構成セグメントをデコードしてグラフィクスオブジェクトを得る「ストリームグラフィクスプロセッサ」と、デコードにより得られたグラフィクスオブジェクトを格納する「オブジェクトバッファ」と、画面構成セグメントを格納する「コンポジションバッファ」と、コンポジションバッファに格納された画面構成セグメントを解読して、これらの画面構成セグメントにおける制御項目に基づき、オブジェクトバッファに得られたグラフィクスオブジェクトを用いてグラフィクスプレーン上で画面構成を行う「コンポジションコントローラ」とを含む。
テキスト字幕ストリームは、複数の字幕記述データから構成される。
左目用IGストリーム、右目用IGストリームは何れも複数のディスプレイセットを含み、各ディスプレイセットは、複数の機能セグメントを含む。ディスプレイセットには、PGストリームと同様、エポックスタートのディスプレイセット、ノーマルケースのディスプレイセット、アクジッションポイントのディスプレイセット、エポックコンティニューのディスプレイセットが存在する
これらのディスプレイセットに属する複数の機能セグメントには以下の種類がある。
このオブジェクト定義セグメントは、PGストリームのものと同じである但しIGストリームのグラフィクスオブジェクトは、ページのインエフェクト、アウトエフェクト、ボタン部材のノーマル状態、セレクテッド状態、アクティブ状態を定義するものである。オブジェクト定義セグメントは、ボタン部材の同じ状態を定義するもの同士、同じエフェクト映像を構成するもの同士、グループ化されている。同じ状態を定義するオブジェクト定義セグメントを寄せ集めたグループをグラフィクスデータ集合という。
パレット定義セグメントは、PGストリームのものと同じである。
対話制御セグメントは、複数のページ情報を含み、複数のページ情報は、マルチページメニューの画面構成を規定する情報であり、各ページ情報は、エフェクトシーケンスと、複数のボタン情報と、パレット識別子の参照値とを含む。
1つのディスプレイセットに属する複数の機能セグメントの最後尾に位置する機能セグメントである。対話制御セグメントからこのエンドセグメントまでが、1つのディスプレイセットを構成する機能セグメントであるとして解釈される。
ボタン近接情報は、あるボタンがセレクテッド状態になっていて、上下左右方向の何れかを指示するキー操作があった場合、どのボタンをセレクテッド状態にすべきかを指定する情報である。
セレクションタイムアウトタイムスタンプは、カレントページにおけるボタン部材を自動的にアクティベートして、セットボタンページコマンドを再生装置に実行させるためのタイムアウト時間を示す。
ユーザタイムアウトディレーションは、カレントページをファーストページに戻して、ファーストページのみが表示されている状態にするためのタイムアウト時間を示す。
コンポジションタイムアウト情報は、対話制御セグメントによる対話的な画面表示を終了させる時間を示す。IGストリームにおいてディスプレイセットの開始時点は、対話制御セグメントを格納したPESパケットのDTSによって特定され、ディスプレイセットの終了時点は、対話制御セグメントのコンポジションタイムアウト時刻によって特定される。左目用、右目用では、これらのDTSと、コンポジションタイムアウト時刻とは同一時点に設定される。
IGデコーダは、IGストリームから読み出されう機能セグメントを格納する「コーデッドデータバッファ」と、画面構成セグメントをデコードしてグラフィクスオブジェクトを得る「ストリームグラフィクスプロセッサ」と、デコードにより得られたグラフィクスオブジェクトを格納する「オブジェクトバッファ」と、画面構成セグメントを格納する「コンポジションバッファ」と、コンポジションバッファに格納された画面構成セグメントを解読して、これらの画面構成セグメントにおける制御項目に基づき、オブジェクトバッファに得られたグラフィクスオブジェクトを用いてグラフィクスプレーン上で画面構成を行う「コンポジションコントローラ」とを含む。
<カレントPGストリームの決定及びカレントPGストリームの再生タイプ決定>
PSR2のPG_テキスト字幕ストリーム番号によって参照されるPG_テキスト字幕ストリームは、出力モード(PSR22)、PSR24における立体視PGケーパビリティと、is_SS_PGを用いて選択される。
条件(A):カレントPG_テキスト字幕ストリーム番号から特定されるPGストリームをデコードする能力が再生装置に存在する。
条件(A):カレントPG_テキスト字幕ストリーム番号のテキスト字幕ストリームのキャラクターコードをビットマップに展開する能力が再生装置に存在すること。かかる再生能力は、PSRセット23におけるPSR30に示されている。
ここで条件(Z)とは、言語特性のサポートがないアンサポート言語の字幕を、再生することをユーザが意図していることであり、かかる意図は、レジスタセットにおけるPSR30に示されている。
図30は、セット立体視ストリームコマンド(SetstreamSSCommand)によってストリーム変化が要求された場合に実行すべきProcedure when stream change is requestedの処理手順を示すフローチャートである。
PSR0のIGストリーム番号によって参照されるIGストリームは、PSR22における出力モードと、PSR24における立体視PGケーパビリティと、is_SS_IGとを用いて選択される。
模式的に示したものである。
(変形例)
本実施の形態の変形例として、ピクチャデータと字幕とを合成する際に、ビデオプレーンメモリに格納されたピクチャデータのみならず、PGプレーンメモリに格納された字幕についても上方向又は下方向にシフトさせる方法について説明する。
本変形例ではPGストリームのストリーム登録情報に、『ビデオ上移動時PGシフト値(PG_v_shift_value_for_Up)』、及び『ビデオ下移動時PGシフト値(PG_v_shift_value_for_Down)』が追加されている。
(第2実施形態)
次に本発明の第2実施形態について説明する。
図42に示すように、その最適なサイズで映像を画面中央に表示させ、斜線で示す周りの領域には黒枠を表示するなどの処理を行うことができる。
(第3実施形態)
本実施形態では、立体視インターリーブドストリームファイルを内部構造についての改良である。
立体視インターリーブドストリームファイル(FileSS)は、2TSをインターリーブ形式にしたストリームファイル(2TSインターリーブファイル)であり、5桁の整数値と、立体視再生用のインターリーブド形式ファイルである旨を示す拡張子(ssif)とによって識別される。立体視インターリーブドストリームファイルは、エクステントSS[n]から構成され、エクステントSS[n](EXTSS[n])は、インデックス番号nによって特定される。インデックス番号nは、立体視インターリーブドストリームファイルの先頭から1つずつインクリメントされる番号である。
ファイルベース(FileBase)は、ファイル2Dに対応するクリップ情報におけるエクステントスタートポイント情報によって指示されるメインTSを"格納している"とされる仮想的なストリームファイルであり、少なくとも1つのエクステント1[i](EXT1[i]と呼ぶ)によって構成される。エクステント1[i]は、ファイルベースにおけるi番目のエクステントであり、iは、エクステントのインデックス番号であり、ファイルベースの先頭を0としてインクリメントされる。ファイルベースは、2TSファイルである立体視インターリーブドストリームファイルを、1TSファイルとして扱うための仮想的なストリームファイルであり、そのファイルエントリーを、再生装置のメモリ上で構築することで仮想的に生成される。
ファイルディペンデント(FileDependent)は、サブTSを"格納している"とされるストリームファイルであり、エクステント2[i](EXT2[i])によって構成される。EXT2[i]は、ファイルディペンデントにおけるi番目のエクステントであり、iは、エクステントのインデックス番号であり、ファイルディペンデントの先頭を0としてインクリメントされる。ファイルディペンデントは、2TSファイルである立体視インターリーブドストリームファイルを、サブTSを格納した1TSファイルとして扱うための仮想的なストリームファイルであり、そのファイルエントリーを、再生装置のメモリ上で構築することで仮想的に生成される。
ファイル2Dは、2D出力モードにおいて再生されるメインTSを格納している1TSのストリームファイルであり、エクステント2Dから構成される。ファイル2Dは、5桁の整数値と、立体視再生用のインターリーブド形式ファイルである旨を示す拡張子(ssif)とによって識別される。
ベースビューデータブロック(B[i])は、メインTSのi番目のデータブロックである。ここで、メインTSとは、カレントプレイアイテム情報のクリップ情報ファイル名情報(クリップ情報ファイルネーム情報)を通じて、メインパスの基軸として指定されているTSである。B[i]の"i"は、ファイルベースの先頭のデータブロックを0としてインクリメントされるインデックス番号である。
ディペンデントビューデータブロック(D[i])は、サブTSのi番目のデータブロックである。サブTSとは、カレントプレイアイテム情報に対応する拡張ストリーム選択テーブルのストリーム登録列におけるストリームエントリーにおいて、サブパスの基軸として指定されているTSである。D[i]の"i"は、ファイルディペンデントの先頭のデータブロックを0としてインクリメントされるインデックス番号である。
上述したように、ファイル2Dのエクステントには、ファイルベースのエクステントと共通のものと、ファイルベースと共通ではないものとがある。
エクステントスタートポイントについて説明する。
(2)別のエクステントスタートポイント情報テーブルは、サブパスタイプ=8のサブパスを含むプレイリストのサブプレイアイテムによって参照されるクリップ情報に格納されて再生装置に供給される。
(第4実施形態)
本実施形態は、立体視インターリーブドストリームファイルを構成するデータブロックから、ATCシーケンスを復元するための改良に関する。図55は、立体視インターリーブドストリームファイル構成するデータブロックからATCシーケンスがどのように復元されるかを示す。
ここで任意の時点からのランダムアクセスを行う際、ストリームファイル内のセクタサーチを行う必要がある。セクタサーチとは、任意の時点からのランダムアクセスを行う際、その時点に対応するソースパケットのソースパケット番号を特定して、そのソースパケット番号のソースパケットを含むセクタから、ファイルリードを行うという処理である。
(第5実施形態)
本実施の形態では、BD-ROMに格納された3Dコンテンツを再生する上でのHDIM出力の課題、超解像度方法、高フレームレート化方法について説明する。
(第6実施形態)
本実施形態では、これまでの実施形態に示した記録媒体の作り方、つまり、記録方法の形態について説明する。
EXT2Dの下限値 ≧(Rud2D×Rbext2D)/(Rud2D-Rbext2D)×Tjump2D(n)
ベースビューデータブロックB[n]ssに対応するエクステントをEXT1[n]であるものとする。この場合、EXT1[n]の下限値は、B-Dプレゼンテーションモードの再生時、各ベースビューデータブロックから次のディペンデントビューデータブロックまでのジャンプ期間と、当該ディペンデントビューデータブロックから次のベースビューデータブロックまでのジャンプ期間とを通して、ダブルバッファのアンダーフローを生じさせないように決定される。
EXT1[n]の下限値 ≧(Rud3D×Rbext3D)/(Rud3D-Rbext3D)
×(TFjump3D(n)+EXT2[n]/(Rud3D+TBjump3D(n)))
EXT2の下限値は、B-Dプレゼンテーションモードの再生時、各ディペンデントビューエクステントから次のベースビューエクステントまでのジャンプ期間と、当該ベースビューエクステントから次のディペンデントビューエクステントまでのジャンプ期間とを通して再生装置におけるダブルバッファにアンダーフローを生じさせないように決定されている。
EXT2[n]の下限値 ≧(Rud3D×Rbext3D)/(Rud3D-Rdext3D)
×(TBjump3D(n)+EXT1[n+1]/(Rud3D+TFjump3D(n+1)))
<EXTSSの具体的な値>
あるエクステントの読み出しから、次のエクステントへのジャンプにあたって、そのジャンプの直前のバッファ占有量は、充分なものでなければならない。そうすると、立体視インターリーブドストリームファイルの読み出し時にあたってリードバッファは、1つのエクステントによって充填される必要があり、バッファアンダーフローの発生を避けねばならない。
Tdiff=ceil[((S1stEXT1[i]EXTSSnext]-S1stEXT1[i]EXTSS)x1000x8)/Rud72]
ここでTdiffは、S1stEXT2[i]EXTssの読出期間と、S1stEXT2[i]EXTssnextの読出期間との差分を意味し、S1stEXT2[i]EXTssは、EXTssの最初に位置するEXT2[i]のサイズであり、S1stEXT2[i]EXTssNEXTは、EXTssNEXTの最初に位置するEXT2[i]のサイズである。EXTssnextは、立体視インターリーブドストリームファイルにおけるエクステントであって、EXTssの直後に位置し、EXTssとシームレスに再生されるものである。
このTdiffと、EXTssnextへのジャンプ時間(Tjump)とを用いれば、各エクステントにおける平均ビットレートに基づく最小エクステントサイズであるSextssは、以下の条件4を満たす値として算出される。
SextSS[Byte]≧ceil[(Tjump+Tdiff×Rud72)/(1000×8))×(Rextss×192)/(Rud72×188-Rextss×192)]
ここで、Rud72は、立体視出力モードにおけるBD-ROMドライブからのデータレートである。
ATCDextss=ATCstart_extssnext -ATCstart_extss
ATCDextss=ATClast_extss - ATCstart_extss + ceil(27000000x188x8/min(Rts1,Rts2))
ATCDextssは、EXTssのATC期間である。
EXTssは、カレントプレイアイテム(プレイアイテム2)によって使用されるATCシーケンスのデータバイトを含む。
<ベースビューデータブロック、ディペンデントビューデータブロックの記録の詳細>
メインTSにおけるGOP及びサブTSにおけるGOPを記録媒体に記録するにあたって、拡張エントリーマップにおける拡張エントリーマップのエントリーは、ディペンデントビューピクチャデータであって、同じ再生時刻で再生されるべきベースビューピクチャデータが、基本エントリーマップのエントリーによってポインティングされているもののみをポインティングするようにしている。
第3実施形態で述べたインデックステーブルは、以下のようにして生成することができる。図59のフローチャートにおいて、ベースビュービデオストリーム、ディペンデントビュービデオストリーム、クリップ情報ファイル、プレイリスト情報ファイルを作成することで、記録媒体に記録すべき各プレイリストの表示周波数が判明する。これらのうち、ファーストプレイタイトルで使用されるプレイリストの解像度・表示周波数、又は、タイトル番号0~999で指示されるタイトルのプレイリストの解像度・表示周波数を、インデックステーブルのBDMVアプリケーション情報におけるビデオフォーマット情報及びフレームレート情報に設定する。こうすることで、プレイリスト再生時に適用すべき解像度・表示周波数が、インデックステーブルにおいて設定されることになる。
アプリケーションは例えばJava(登録商標)アプリケーションであり、仮想マシンのヒープ領域(ワークメモリとも呼ばれる)にロードされた1つ以上のxletプログラムからなる。このワークメモリにロードされたxletプログラム、及び、データから、アプリケーションは構成されることになる。
(第7実施形態)
本実施形態では、これまでの実施形態で説明した再生装置の機能を統合した、2D/3D再生装置の内部構成について説明する。
HDMVモジュール13は、コマンドインタプリタを具備し、ムービーオブジェクトを構成するナビゲーションコマンドを解読して実行することでHDMVモードの制御を実行する。
(第8実施形態)
以降の実施形態は、本願の優先権主張の基礎となった特許出願の願書に添付した明細書及び図面に記載された発明と同一発明(以下、本発明と呼ぶ)を実施するための形態である。
PSR1 : 主音声ストリーム番号
PSR2 : 字幕ストリーム番号
PSR3 : アングル番号
PSR4 : タイトル番号
PSR5 : チャプタ番号
PSR6 : プログラム番号
PSR7 : セル番号
PSR8 : 選択キー情報
PSR9 : ナビゲーションタイマー
PSR10 : 再生時刻情報
PSR11 : カラオケ用ミキシングモード
PSR12 : パレンタル用国情報
PSR13 : パレンタルレベル
PSR14 : プレーヤ設定値(ビデオ)
PSR15 : プレーヤ設定値(オーディオ)
PSR16 : 音声ストリーム用言語コード
PSR17 : 音声ストリーム用言語コード
PSR18 : 字幕ストリーム用言語コード
PSR19 : 字幕ストリーム用言語コード
PSR20 : プレーヤリージョンコード
PSR21 : 2D/3D出力モードのユーザー優先選択
PSR22 : 現在の2D/3D出力モード
PSR23 : ディスプレイの3D映像表示能力
PSR24 : 3D映像再生能力
PSR25 : 予備
PSR26 : 予備
PSR27 : 予備
PSR28 : 予備
PSR29 : 予備
PSR30 : 予備
PSR31 : 予備
PSR10は、AVクリップに属する各ピクチャデータが表示される度に更新される。つまり再生装置が新たなピクチャデータを表示させれば、その新たなピクチャデータの表示時刻(PTS)を示す値にPSR10は更新される。このPSR10を参照すれば、現在の再生時点を知得することができる。
また、再生制御部1507は、システムパラメータの状態をチェックしながら再生する。また、PSR1、PSR2、PSR21、PSR22は、それぞれ音声ストリーム番号、字幕ストリーム番号、副映像ストリーム、副音声ストリームの番号を示し、これらの値は、前述したストリーム選択番号606に対応する。例えば、プログラム実行部1506によって、音声ストリーム番号PSR1が変更されたとする。再生制御部1507は、現在再生しているプレイアイテムのストリーム選択テーブル605の中から、ストリーム選択番号606と比較して、一致するストリームエントリ609を参照し、再生する音声ストリームを切り替える。このようにしてどの音声・字幕・副映像ストリームを再生するかどうかを切り替える。
次に図80を用いて家庭用ディスプレイで立体視を行う原理を説明する。立体視を実現する手法としては、ホログラフィ技術を用いる方法と、視差画像を用いる方式の大きく2つある。
2つ目の視差画像を用いた方式で、右目に入る映像と、左目に入る映像を各々用意し、それぞれの目に対応したピクチャーだけが入るようにして立体視を行う方法である。図80でユーザーが顔の中央の延長線上にある、比較的小さな立方体を見ている様子を上から見た図であり、右上の図は左目が見た場合の例を示しており、右下の図は同じ立方体を右目が見た場合の例を示している。
次に本発明に係る記録媒体であるBD-ROMの、3D映像を格納するためのデータ構造について説明する。
(第9実施形態)
本実施形態では、前述の実施形態において説明された構造のデータを再生する再生装置に関して、集積回路3を用いて実現した構成例(図100)について説明する。
(備考)
以上、本願の出願時点において、出願人が知り得る最良の実施形態について説明したが、以下に示す技術的トピックについては、更なる改良や変更実施を加えることができる。各実施形態に示した通り実施するか、これらの改良・変更を施すか否かは、何れも任意的であり、実施する者の主観によることは留意されたい。
各実施形態で説明したオフセットメタデータは、これまでに説明した形式に限らず他のデータ形式でも実現することができる。以下にオフセットメタデータの他のデータ形式を列挙する。
第5のデータ形式は、オフセットメタデータをプレイリスト情報ファイルに格納する更に他の形式であり、図93に示すように、同じオフセットメタデータを使うプレイアイテムへの参照ID(ref_playitem_id_of_same_offset_metadata)を追加している。2D/3D再生装置は、ref_playitem_id_of_same_offset_metadataがインバリッドを示す0xFFFFでない場合には、ref_playitem_id_of_same_offset_metadataが指すプレイアイテムと同じオフセットメタデータを適用する。このようにすることで、同じオフセットメタデータを持つ複数プレイアイテムに対して、一つのオフセットメタデータだけを定義すればよくなるためデータ量を圧縮できる。
第6のデータ形式は、オフセットメタデータをプレイリスト情報ファイルに格納するまた更に他の形式である。
このデータ形式では、図94に示すように、プレイアイテム単位でループするヘッダと、オフセットメタデータを格納する領域を分けている。プレイアイテムとオフセットメタデータは、offset_block_idで関連付けられている。これにより同じオフセットメタデータを使う複数のプレイアイテムがある場合に1つのオフセットメタデータを定義すればよくなるためデータ量を圧縮できる。また、ヘッダには該当のオフセットメタデータが格納されるファイルのアドレス値(start_address)を格納してもよい。このようにすることで、プレイアイテム単位でのアクセスが容易になる。
字幕の本数を減らしストリームの帯域増加を抑えるためには、1つのPGストリームを、1plane+Offset方式で使用するPGストリームと、2プレーンL/R方式で使用する左右何れか一方のPGストリームとして供用することが有効である。
図97の(a)は、FileBaseとFileDependentとのエクステントがインターリーブされている構成を示しており、ディスク上のデータ領域R[2]の先頭に付加されている三角形はFileDependentのエントリポイントのディスク上の位置、ディスク上のデータ領域L[2]の先頭に付加されている三角形はFileBaseエントリポイントのディスク上の位置を示す。このとき、2D/3D再生装置は該当のエントリポイントから飛び込み再生を行うような場合には、ディスク上のデータ領域であるR[2]のデータを読み込み、その後、L[2]を読みながらデコードを開始することになる。R[2]のデータ読み込みが完了するまで、次のL[2]を読み出せないためデコード開始できない。
付加情報は、以下の情報要素を含む拡張ストリーム選択テーブルであるとして、プレイリスト情報の拡張情報フィールド内に組み込んでもよい。
BD-ROMドライブは、半導体レーザ、コリメートレンズ、ビームスプリッタ、対物レンズ、集光レンズ、光検出器を有する光学ヘッドを備える。半導体レーザから出射された光ビームは、コリメートレンズ、ビームスプリッタ、対物レンズを通って、光ディスクの情報面に集光される。
各実施の形態における記録媒体は、光ディスク、半導体メモリーカード等、パッケージメディア全般を含んでいる。本実施の形態の記録媒体は予め必要なデータが記録された光ディスク(例えばBD-ROM、DVD-ROMなどの既存の読み取り可能な光ディスク)を例に説明をするが、これに限定される必要はなく、例えば、放送またはネットワークを経由して配信された本発明の実施に必要なデータを含んだ3Dコンテンツを光ディスクへ書き込む機能を有する端末装置(例えば左記の機能は再生装置に組み込まれていても良いし、再生装置とは別の装置であってもよい)を利用して書き込み可能な光ディスク(例えばBD-RE、DVD-RAMなどの既存の書き込み可能な光ディスク)に記録し、この記録した光ディスクを本発明の再生装置に適用しても本発明の実施は可能である。
各実施の形態で説明をしたデータ構造を半導体メモリーに記録する記録装置、及び、再生する再生装置の実施形態について説明する。
各実施形態で説明した再生装置は、本実施の形態で説明をしたデータに相応するデータ(配信データ)を電子配信サービスの配信サーバから受信し、半導体メモリカードに記録する端末装置としても実現することができる。
を行なう。これらのチェックを行なう順番は、どのような順序で行なってもよい。
200 再生装置
300 表示装置
400 3D眼鏡
Claims (11)
- 立体視画像を構成するビデオストリームと、プレイリスト情報と、複数の字幕ストリームとが記録された記録媒体であって、
プレイリスト情報は、ストリーム選択テーブルと、付加情報とを含み、
ストリーム選択テーブルは、平面視再生モードにおいて再生を許可すべき字幕ストリームのストリーム番号を、ストリームエントリー及びストリーム属性に対応付けて示し、
付加情報は、ストリーム選択テーブルにおけるストリーム番号に対応付けられており、領域確保フラグを含み、
領域確保フラグは、再生装置が立体視再生モードに設定されている場合の字幕の表示領域がビデオプレーンの上端であるか、又は、下端であるかを示し、
字幕の表示領域がビデオプレーン上端であることが領域確保フラグによって示される場合、ビデオストリームは再生装置によりビデオプレーンの下方向にシフトして描画され、
字幕の表示領域がビデオプレーン下端であることが領域確保フラグによって示される場合、ビデオストリームは再生装置によりビデオプレーンの上方向にシフトして描画される
ことを特徴とする記録媒体。 - ビデオプレーンの下方向又は上方向のシフト量を規定するシフト制御情報を含む
ことを特徴とする請求項1記載の記録媒体。 - 前記ストリーム選択テーブルにおけるストリームエントリーのストリーム番号は、複数の字幕ストリームを選択するにあたっての各字幕ストリームの優先順序を示す
ことを特徴とする請求項1記載の記録媒体。 - 字幕の表示領域がビデオプレーンの上端であることを示す領域確保フラグを含む複数の付加情報は、連続するストリーム番号に対応付けられており、
字幕の表示領域がビデオプレーンの下端であることを示す領域確保フラグを含む複数の付加情報は、他の連続するストリーム番号に対応付けられている
ことを特徴とする請求項3記載の記録媒体。 - 立体視画像を構成するビデオストリームと、プレイリスト情報と、複数の字幕ストリームとが記録された記録媒体を再生する再生装置であって、
プレイリスト情報は、ストリーム選択テーブルと、付加情報とを含み、
ストリーム選択テーブルは、平面視再生モードにおいて再生を許可すべき字幕ストリームのストリーム番号を、ストリームエントリー及びストリーム属性に対応付けて示し、
付加情報は、ストリーム選択テーブルにおけるストリーム番号に対応付けられており、領域確保フラグを含み、
領域確保フラグは、再生装置が立体視再生モードに設定されている場合の字幕の表示領域がビデオプレーンの上端であるか、又は、下端であるかを示し、
前記再生装置は、
現在の再生モードが平面視再生モード、立体視再生モードのどちらであるかを示すモード記憶部と、
カレントストリーム番号を格納しているストリーム番号レジスタと、
カレントストリーム番号に対応する付加情報の領域確保フラグが、字幕の表示領域はビデオプレーンの上端であることを示しているか、又は、下端であることを示しているか、を格納している領域確保レジスタと、
ビデオストリームをデコードするビデオデコーダと、
デコードによって得られたピクチャを格納するビデオプレーンと、
複数の字幕ストリームのうち、カレントストリーム番号に対応するものをデコードし、デコードにより得られた字幕をビデオプレーンに合成する字幕デコーダとを備え、
領域確保レジスタが、字幕の表示領域はビデオプレーン上端であることを示している場合、ビデオストリームをビデオプレーンの下方向にシフトして描画し、
領域確保レジスタが、字幕の表示領域はビデオプレーン下端であることを示している場合、ビデオストリームをビデオプレーンの上方向にシフトして描画し、
前記字幕デコーダは、
カレントストリーム番号に対応する付加情報の領域確保フラグが示す表示領域に字幕ストリームのデコードによって得られた字幕を書き込む
ことを特徴とする再生装置。 - メインビュービデオストリームと、サブビュービデオストリームと、メタデータとが記録された記録媒体であって、
前記メインビュービデオストリームは、立体視映像のメインビューを構成するピクチャデータを含み、
前記サブビュービデオストリームは、立体視映像のサブビューを構成するピクチャデータを含み、
前記メタデータは、複数の表示装置の画面サイズ情報に対応するオフセット補正値を含み、
前記オフセット補正値は、前記メインビューを構成するピクチャデータが描画されるメインビュービデオプレーン及び前記サブビューを構成するピクチャデータが描画されるサブビュービデオプレーンの少なくとも一方を、水平座標の右方向及び左方向へ移動させるオフセットを規定する
ことを特徴とする記録媒体。 - 前記メインビュービデオプレーン及び前記サブビュービデオプレーンの少なくとも一方には、グラフィクスデータが描画されたグラフィクスプレーンが重畳されており、
前記オフセット補正値は、グラフィクスプレーンが重畳されたメインビュービデオプレーン及び前記サブビュービデオプレーンの少なくとも一方を、水平座標の右方向及び左方向へ移動させるオフセットを規定する
ことを特徴とする請求項6記載の記録媒体。 - メインビュービデオストリームと、サブビュービデオストリームと、メタデータとが記録された記録媒体を再生する再生装置であって、
前記メインビュービデオストリームは、立体視映像のメインビューを構成するピクチャデータを含み、
前記サブビュービデオストリームは、立体視映像のサブビューを構成するピクチャデータを含み、
前記メタデータは、複数の表示装置の画面サイズ情報に対応するオフセット補正値を含み、
前記オフセット補正値は、前記メインビューを構成するピクチャデータが描画されるメインビュービデオプレーン及び前記サブビューを構成するピクチャデータが描画されるサブビュービデオプレーンの少なくとも一方を、水平座標の右方向及び左方向へ移動させるオフセットを規定し、
前記再生装置は、
前記メインビュービデオストリーム及び前記サブビュービデオストリームをデコードすることにより、メインビューを構成するピクチャデータ及びサブビューを構成するピクチャデータを得るビデオデコーダと、
前記メインビューを構成するピクチャデータが描画されるメインビュービデオプレーンと、
前記サブビューを構成するピクチャデータが描画されるサブビュービデオプレーンと、
を有し、
前記メタデータから、接続されている表示装置の画面サイズに基づいたオフセット補正値を抽出し、メインビュービデオプレーン及びサブビュービデオプレーンの少なくとも一方を、水平座標の右方向及び左方向へ移動させることを特徴とする再生装置。 - 前記メインビュービデオプレーン及び前記サブビュービデオプレーンの少なくとも一方には、グラフィクスデータが描画されたグラフィクスプレーンが重畳されており、
グラフィクスプレーンが重畳されたメインビュービデオプレーン及びサブビュービデオプレーンの少なくとも一方を、水平座標の右方向及び左方向へ移動させることを特徴とする請求項8記載の再生装置。 - 前記再生装置は、接続されている表示装置から取得した画面サイズを格納する設定レジスタを備えることを特徴とする請求項8記載の再生装置。
- 平面視映像の再生に利用されるメインビュービデオストリームと、前記メインビュービデオストリームと組み合わされて立体視映像の再生に利用されるサブビュービデオストリームと、プレイリスト情報と、複数の字幕ストリームを含んだデータを受信し、映像信号処理を行う半導体集積回路であって、
前記プレイリスト情報は、各前記複数の字幕ストリームに対応した領域確保フラグを含み、
領域確保フラグは、字幕の表示領域がビデオプレーンの上端であるか、又は、下端であるかを示し、
前記メインビュービデオストリームは、メインビュートランスポートストリームとして多重化された後、複数のメインビューデータ群に分割され、
前記サブビュービデオストリームは、サブビュートランスポートストリームとして多重化された後、複数のサブビューデータ群に分割され、
前記データにおいて、前記メインビューデータ群と前記サブビューデータ群は交互に配置されており、
前記データにおいて、前記各複数の字幕ストリームは、前記メインビュートランスポートストリームもしくは前記サブビュートランスポートストリームに多重化されている、もしくは単独のストリームで存在し、
前記半導体集積回路は、
前記半導体集積回路の制御を行う主制御部と、
前記データを受信し、前記半導体集積回路の内部もしくは外部に設けられたメモリに一旦格納した後、ビデオデータと字幕データに多重分離するストリーム処理部と、
前記ビデオデータと、多重分離された字幕データもしくは単独で存在する字幕ストリーム中の字幕データとをそれぞれデコードする信号処理部と、
デコードされた前記ビデオデータを出力するAV出力部とを備えており、
前記ストリーム処理部は、受信した前記データのうち、前記メインビューデータ群と前記サブビューデータ群が交互に配置されたデータの格納先を前記メモリ内の第1の領域と第2の領域との間で切り替える切替部を備えており、
前記主制御部は、前記メインビューデータ群に属しているデータを、前記第1の領域に格納するように前記切替部を制御し、前記サブビューデータ群に属しているデータを、前記第2の領域に格納するように前記切替部を制御し、
前記デコードされたビデオデータのうち前記メインビューデータ群に属しているデータは、前記メモリ内の第3の領域に格納され、前記第3の領域は前記メインビュービデオプレーンに対応しており、
前記デコードされたビデオデータのうち前記サブビューデータ群に属しているデータは、前記メモリ内の第4の領域に格納され、前記第4の領域は前記サブビュービデオプレーンに対応しており、
前記デコードされた字幕データは、前記メモリ内の第5の領域に格納され、前記第5の領域は字幕プレーンに対応しており、
前記AV出力部は、前記デコードされたビデオデータと前記デコードされた字幕データとを重畳する画像重畳部を備えており、
前記画像重畳部は、前記領域確保フラグが示す字幕の表示領域がビデオプレーンの上端である場合は、前記メインビュービデオプレーン及び前記サブビュービデオプレーンを垂直座標の下方向にシフトさせて、前記領域確保フラグが示す字幕データの表示領域に字幕が重畳されるように、前記デコードされたビデオデータと前記デコードされた字幕データを重畳し、
前記画像重畳部は、前記領域確保フラグが示す字幕の表示領域がビデオプレーンの下端である場合は、前記メインビュービデオプレーン及び前記サブビュービデオプレーンを垂直座標の上方向にシフトさせて、前記領域確保フラグが示す字幕データの表示領域に字幕が重畳されるように、前記デコードされたビデオデータと前記デコードされた字幕データを重畳する
ことを特徴とする半導体集積回路。
Priority Applications (10)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
MX2011003076A MX2011003076A (es) | 2009-06-17 | 2010-06-16 | Medio de grabacion de informacion y dispositivo de reproduccion para reproducir imagenes en tercera dimension. |
JP2010546132A JP4733785B2 (ja) | 2009-06-17 | 2010-06-16 | 3d映像を再生するための情報記録媒体、記録媒体の記録方法、再生装置、及び記録媒体再生システム |
ES10789226.7T ES2634941T3 (es) | 2009-06-17 | 2010-06-16 | Medio de registro de información para reproducir vídeo de 3D, y dispositivo de reproducción |
EP10789226.7A EP2445224B1 (en) | 2009-06-17 | 2010-06-16 | Information recording medium for reproducing 3d video, and reproduction device |
BRPI1005171A BRPI1005171A2 (pt) | 2009-06-17 | 2010-06-16 | meio de gravação de informações e dispositivo de reprodução destinado à reprodução de imagens em 3d |
DK10789226.7T DK2445224T3 (en) | 2009-06-17 | 2010-06-16 | INFORMATION RECORDING MEDIUM FOR 3D VIDEO RENDERING AND REPLACEMENT DEVICE |
RU2011120907/08A RU2520325C2 (ru) | 2009-06-17 | 2010-06-16 | Носитель записи информации и устройство воспроизведения для воспроизведения 3d изображений |
CN2010800027752A CN102172033B (zh) | 2009-06-17 | 2010-06-16 | 再生装置、记录方法、记录介质再生系统 |
US13/036,616 US8121460B2 (en) | 2009-06-17 | 2011-02-28 | Information recording medium and playback device for playing back 3D images |
ZA2011/01979A ZA201101979B (en) | 2009-06-17 | 2011-03-15 | "information recording medium and playback device playing back 3d images" |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2009144568 | 2009-06-17 | ||
JP2009-144568 | 2009-06-17 |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/036,616 Continuation US8121460B2 (en) | 2009-06-17 | 2011-02-28 | Information recording medium and playback device for playing back 3D images |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010146847A1 true WO2010146847A1 (ja) | 2010-12-23 |
Family
ID=43356178
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2010/003998 WO2010146847A1 (ja) | 2009-06-17 | 2010-06-16 | 3d映像を再生するための情報記録媒体、及び再生装置 |
Country Status (13)
Country | Link |
---|---|
US (1) | US8121460B2 (ja) |
EP (1) | EP2445224B1 (ja) |
JP (3) | JP4733785B2 (ja) |
KR (1) | KR20120036791A (ja) |
CN (1) | CN102172033B (ja) |
BR (1) | BRPI1005171A2 (ja) |
DK (1) | DK2445224T3 (ja) |
ES (1) | ES2634941T3 (ja) |
MX (1) | MX2011003076A (ja) |
RU (1) | RU2520325C2 (ja) |
TW (1) | TW201130287A (ja) |
WO (1) | WO2010146847A1 (ja) |
ZA (1) | ZA201101979B (ja) |
Cited By (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2012017603A1 (ja) * | 2010-08-06 | 2012-02-09 | パナソニック株式会社 | 再生装置、集積回路、再生方法、プログラム |
WO2012060182A1 (ja) * | 2010-11-05 | 2012-05-10 | 富士フイルム株式会社 | 画像処理装置、画像処理プログラム、画像処理方法、及び記憶媒体 |
WO2012108393A1 (ja) * | 2011-02-09 | 2012-08-16 | ソニー株式会社 | 電子機器、電子機器における立体画像情報送信方法、および電子機器における立体画像情報受信方法 |
JP2012178706A (ja) * | 2011-02-25 | 2012-09-13 | Toshiba Corp | 映像表示方法および映像表示装置 |
US20120235992A1 (en) * | 2011-03-14 | 2012-09-20 | Kenjiro Tsuda | Stereoscopic image processing apparatus and stereoscopic image processing method |
JP2012220888A (ja) * | 2011-04-13 | 2012-11-12 | Nikon Corp | 撮像装置 |
JP2012227687A (ja) * | 2011-04-19 | 2012-11-15 | Toshiba Corp | 電子機器および映像表示方法 |
CN103067680A (zh) * | 2011-10-21 | 2013-04-24 | 康佳集团股份有限公司 | 一种2d转3d视频格式下的osd显示方法及系统 |
WO2013061523A1 (ja) * | 2011-10-28 | 2013-05-02 | パナソニック株式会社 | 旧来のフォーマットとの互換を維持しつつも、記録内容を編集することができる記録媒体、再生装置、記録装置、再生方法、記録方法 |
JP5501458B2 (ja) * | 2010-06-11 | 2014-05-21 | 富士フイルム株式会社 | 立体画像表示装置、立体撮像装置、及び方法 |
KR101530826B1 (ko) * | 2014-08-18 | 2015-06-24 | 정관호 | 360도 공간영상 재생방법 및 그 시스템 |
WO2016059760A1 (ja) * | 2014-10-17 | 2016-04-21 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 記録媒体、再生方法、および再生装置 |
WO2016079925A1 (ja) * | 2014-11-17 | 2016-05-26 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 記録媒体、再生方法、および再生装置 |
JP2016100039A (ja) * | 2014-11-17 | 2016-05-30 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 記録媒体、再生方法、および再生装置 |
US10542285B2 (en) | 2011-09-23 | 2020-01-21 | Velos Media, Llc | Decoded picture buffer management |
Families Citing this family (39)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8737810B2 (en) | 2002-11-15 | 2014-05-27 | Thomson Licensing | Method and apparatus for cropping of subtitle elements |
US8074248B2 (en) | 2005-07-26 | 2011-12-06 | Activevideo Networks, Inc. | System and method for providing video content associated with a source image to a television in a communication network |
US9031381B2 (en) * | 2006-07-20 | 2015-05-12 | Panopto, Inc. | Systems and methods for generation of composite video from multiple asynchronously recorded input streams |
US9826197B2 (en) | 2007-01-12 | 2017-11-21 | Activevideo Networks, Inc. | Providing television broadcasts over a managed network and interactive content over an unmanaged network to a client device |
WO2008088741A2 (en) | 2007-01-12 | 2008-07-24 | Ictv, Inc. | Interactive encoded content system including object models for viewing on a remote device |
US20110013888A1 (en) * | 2009-06-18 | 2011-01-20 | Taiji Sasaki | Information recording medium and playback device for playing back 3d images |
TW201130289A (en) * | 2009-07-14 | 2011-09-01 | Panasonic Corp | Image reproducing apparatus |
US9025933B2 (en) | 2010-02-12 | 2015-05-05 | Sony Corporation | Information processing device, information processing method, playback device, playback method, program and recording medium |
KR20120017649A (ko) * | 2010-08-19 | 2012-02-29 | 삼성전자주식회사 | 디스플레이장치 및 그 제어방법 |
US8942543B1 (en) * | 2010-10-06 | 2015-01-27 | Verint Video Solutions Inc. | Systems, methods, and software for improved video data recovery effectiveness |
AU2011315950B2 (en) | 2010-10-14 | 2015-09-03 | Activevideo Networks, Inc. | Streaming digital video between video devices using a cable television system |
US9516379B2 (en) | 2011-03-08 | 2016-12-06 | Qualcomm Incorporated | Buffer management in video codecs |
WO2012138660A2 (en) | 2011-04-07 | 2012-10-11 | Activevideo Networks, Inc. | Reduction of latency in video distribution networks using adaptive bit rates |
US20120281073A1 (en) * | 2011-05-02 | 2012-11-08 | Cisco Technology, Inc. | Customization of 3DTV User Interface Position |
CN102521805B (zh) * | 2011-11-30 | 2013-07-24 | 华中科技大学 | 基于帧间信息的视频文字处理方法 |
EP2815582B1 (en) * | 2012-01-09 | 2019-09-04 | ActiveVideo Networks, Inc. | Rendering of an interactive lean-backward user interface on a television |
KR101872859B1 (ko) * | 2012-01-20 | 2018-06-29 | 엘지전자 주식회사 | 디스플레이 장치 및 그 제어방법 |
US9800945B2 (en) | 2012-04-03 | 2017-10-24 | Activevideo Networks, Inc. | Class-based intelligent multiplexing over unmanaged networks |
US9123084B2 (en) | 2012-04-12 | 2015-09-01 | Activevideo Networks, Inc. | Graphical application integration with MPEG objects |
WO2013183947A1 (ko) | 2012-06-05 | 2013-12-12 | 엘지전자 주식회사 | 3d 방송 서비스를 위한 방송 신호 처리 방법 및 장치 |
EP2862360A1 (en) * | 2012-06-19 | 2015-04-22 | Thomson Licensing | Improved playlist content selection system and method |
US9756282B2 (en) * | 2012-11-20 | 2017-09-05 | Sony Corporation | Method and apparatus for processing a video signal for display |
KR20140099146A (ko) * | 2013-02-01 | 2014-08-11 | 한국전자통신연구원 | 다중 영상 기반 방송 서비스를 위한 다중 영상의 기준 pcr 결정 방법 및 그 장치 |
KR101361108B1 (ko) * | 2013-02-05 | 2014-02-13 | 한국과학기술원 | Dibr 3d영상 저작권 보호를 위한 워터마킹 시스템 및 그 방법 |
WO2014145921A1 (en) | 2013-03-15 | 2014-09-18 | Activevideo Networks, Inc. | A multiple-mode system and method for providing user selectable video content |
US9294785B2 (en) | 2013-06-06 | 2016-03-22 | Activevideo Networks, Inc. | System and method for exploiting scene graph information in construction of an encoded video sequence |
US9219922B2 (en) | 2013-06-06 | 2015-12-22 | Activevideo Networks, Inc. | System and method for exploiting scene graph information in construction of an encoded video sequence |
EP3005712A1 (en) | 2013-06-06 | 2016-04-13 | ActiveVideo Networks, Inc. | Overlay rendering of user interface onto source video |
CN103699584B (zh) * | 2013-12-06 | 2017-11-24 | 福建星网视易信息系统有限公司 | 一种终端播放方法及系统 |
US10475227B1 (en) * | 2014-02-28 | 2019-11-12 | Ansys, Inc. | Systems and methods for three dimensional computation and visualization using a parallel processing architecture |
US9788029B2 (en) | 2014-04-25 | 2017-10-10 | Activevideo Networks, Inc. | Intelligent multiplexing using class-based, multi-dimensioned decision logic for managed networks |
CN104965693A (zh) * | 2014-08-29 | 2015-10-07 | 腾讯科技(深圳)有限公司 | 一种视频处理的方法和系统 |
WO2016038791A1 (ja) | 2014-09-10 | 2016-03-17 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 記録媒体、再生装置および再生方法 |
WO2016039172A1 (ja) * | 2014-09-12 | 2016-03-17 | ソニー株式会社 | 再生装置、再生方法、情報処理装置、情報処理方法、プログラム、および記録媒体 |
CN107431835B (zh) * | 2015-04-13 | 2020-09-11 | 索尼公司 | 发送装置、发送方法、再现装置以及再现方法 |
GB2557194A (en) * | 2016-11-29 | 2018-06-20 | Sony Corp | A decoder, encoder, computer program and method |
US10958890B2 (en) * | 2017-03-31 | 2021-03-23 | Samsung Electronics Co., Ltd. | Method and apparatus for rendering timed text and graphics in virtual reality video |
PL3610654T3 (pl) * | 2017-04-11 | 2022-02-14 | Dolby Laboratories Licensing Corporation | Wrażenia z warstwowej rozszerzonej rozrywki |
CN110383811A (zh) * | 2018-06-29 | 2019-10-25 | 深圳市大疆创新科技有限公司 | 图像处理方法及装置、第一电子设备和图像处理系统 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2002016884A (ja) * | 2000-06-29 | 2002-01-18 | Matsushita Electric Ind Co Ltd | 映像信号再生装置 |
WO2004095837A1 (ja) * | 2003-04-22 | 2004-11-04 | Matsushita Electric Industrial Co., Ltd. | 再生装置、プログラム。 |
JP3935507B2 (ja) | 1996-02-28 | 2007-06-27 | 松下電器産業株式会社 | 高解像度および立体映像記録用光ディスク、光ディスク再生装置、光ディスク記録装置 |
WO2008044191A2 (en) * | 2006-10-11 | 2008-04-17 | Koninklijke Philips Electronics N.V. | Creating three dimensional graphics data |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPH08331456A (ja) * | 1995-05-31 | 1996-12-13 | Philips Japan Ltd | 字幕移動装置 |
JPH10150608A (ja) * | 1996-11-19 | 1998-06-02 | Sanyo Electric Co Ltd | デジタル放送システムのユーザ側端末 |
US20050259147A1 (en) * | 2002-07-16 | 2005-11-24 | Nam Jeho | Apparatus and method for adapting 2d and 3d stereoscopic video signal |
JP2004112169A (ja) * | 2002-09-17 | 2004-04-08 | Victor Co Of Japan Ltd | 色調整装置及び色調整方法 |
ATE517413T1 (de) * | 2003-04-09 | 2011-08-15 | Lg Electronics Inc | Aufzeichnungsmedium mit datenstruktur zur verwaltung der wiedergabe von textuntertiteldaten und verfahren und vorrichtungen zum aufzeichnen und wiedergeben |
JP2005123775A (ja) * | 2003-10-15 | 2005-05-12 | Sony Corp | 再生装置、再生方法、再生プログラムおよび記録媒体 |
KR100619053B1 (ko) * | 2003-11-10 | 2006-08-31 | 삼성전자주식회사 | 서브 타이틀을 기록한 정보저장매체 및 그 처리장치 |
CN1700329B (zh) * | 2004-01-29 | 2010-06-16 | 索尼株式会社 | 再现装置、再现方法、再现程序和记录介质 |
RU2351997C2 (ru) * | 2004-02-10 | 2009-04-10 | ЭлДжи ЭЛЕКТРОНИКС ИНК. | Носитель информации с наличием структуры данных для управления различными потоками данных и способы и устройства записи и воспроизведения |
DE602005023285D1 (de) * | 2004-03-26 | 2010-10-14 | Lg Electronics Inc | Speichermedium,verfahren, und gerät zum wiedergeben von untertiteldatenströmen |
KR20060047266A (ko) * | 2004-04-26 | 2006-05-18 | 엘지전자 주식회사 | 기록매체, 기록매체의 재생방법과 재생장치 |
JP4692950B2 (ja) * | 2004-06-11 | 2011-06-01 | ソニー株式会社 | データ処理装置およびデータ処理方法、プログラムおよびプログラム記録媒体、並びにデータ記録媒体 |
JP2007004896A (ja) * | 2005-06-23 | 2007-01-11 | Toshiba Corp | 情報記憶媒体、情報転送方法、情報再生方法、及び情報記録方法 |
JP2007316613A (ja) * | 2006-04-26 | 2007-12-06 | Matsushita Electric Ind Co Ltd | 字幕表示制御装置 |
WO2009057298A1 (ja) | 2007-11-01 | 2009-05-07 | Panasonic Corporation | 記録媒体、再生装置、記録装置、再生方法、及び記録方法 |
JP2009135686A (ja) * | 2007-11-29 | 2009-06-18 | Mitsubishi Electric Corp | 立体映像記録方法、立体映像記録媒体、立体映像再生方法、立体映像記録装置、立体映像再生装置 |
JP4946874B2 (ja) * | 2008-01-09 | 2012-06-06 | ソニー株式会社 | 再生装置および再生方法 |
-
2010
- 2010-06-16 BR BRPI1005171A patent/BRPI1005171A2/pt not_active Application Discontinuation
- 2010-06-16 WO PCT/JP2010/003998 patent/WO2010146847A1/ja active Application Filing
- 2010-06-16 ES ES10789226.7T patent/ES2634941T3/es active Active
- 2010-06-16 KR KR1020117007329A patent/KR20120036791A/ko not_active Application Discontinuation
- 2010-06-16 EP EP10789226.7A patent/EP2445224B1/en active Active
- 2010-06-16 DK DK10789226.7T patent/DK2445224T3/en active
- 2010-06-16 MX MX2011003076A patent/MX2011003076A/es active IP Right Grant
- 2010-06-16 JP JP2010546132A patent/JP4733785B2/ja active Active
- 2010-06-16 CN CN2010800027752A patent/CN102172033B/zh active Active
- 2010-06-16 RU RU2011120907/08A patent/RU2520325C2/ru not_active IP Right Cessation
- 2010-06-17 TW TW099119670A patent/TW201130287A/zh unknown
-
2011
- 2011-02-28 US US13/036,616 patent/US8121460B2/en active Active
- 2011-03-04 JP JP2011047671A patent/JP5451667B2/ja not_active Expired - Fee Related
- 2011-03-04 JP JP2011047670A patent/JP5491437B2/ja not_active Expired - Fee Related
- 2011-03-15 ZA ZA2011/01979A patent/ZA201101979B/en unknown
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3935507B2 (ja) | 1996-02-28 | 2007-06-27 | 松下電器産業株式会社 | 高解像度および立体映像記録用光ディスク、光ディスク再生装置、光ディスク記録装置 |
JP2002016884A (ja) * | 2000-06-29 | 2002-01-18 | Matsushita Electric Ind Co Ltd | 映像信号再生装置 |
WO2004095837A1 (ja) * | 2003-04-22 | 2004-11-04 | Matsushita Electric Industrial Co., Ltd. | 再生装置、プログラム。 |
WO2008044191A2 (en) * | 2006-10-11 | 2008-04-17 | Koninklijke Philips Electronics N.V. | Creating three dimensional graphics data |
Non-Patent Citations (1)
Title |
---|
See also references of EP2445224A4 * |
Cited By (63)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9369699B2 (en) | 2010-06-11 | 2016-06-14 | Fujifilm Corporation | Stereoscopic image display device, stereoscopic imaging device, and methods |
JP5501458B2 (ja) * | 2010-06-11 | 2014-05-21 | 富士フイルム株式会社 | 立体画像表示装置、立体撮像装置、及び方法 |
US8737811B2 (en) | 2010-08-06 | 2014-05-27 | Panasonic Corporation | Playback device, integrated circuit, playback method, and program |
JP5728649B2 (ja) * | 2010-08-06 | 2015-06-03 | パナソニックIpマネジメント株式会社 | 再生装置、集積回路、再生方法、プログラム |
WO2012017603A1 (ja) * | 2010-08-06 | 2012-02-09 | パナソニック株式会社 | 再生装置、集積回路、再生方法、プログラム |
WO2012060182A1 (ja) * | 2010-11-05 | 2012-05-10 | 富士フイルム株式会社 | 画像処理装置、画像処理プログラム、画像処理方法、及び記憶媒体 |
US9143764B2 (en) | 2010-11-05 | 2015-09-22 | Fujifilm Corporation | Image processing device, image processing method and storage medium |
JP2012169702A (ja) * | 2011-02-09 | 2012-09-06 | Sony Corp | 電子機器、電子機器における立体画像情報送信方法、および電子機器における立体画像情報受信方法 |
CN102972034A (zh) * | 2011-02-09 | 2013-03-13 | 索尼公司 | 电子设备、发送电子设备中的立体图像信息的方法和接收电子设备中的立体图像信息的方法 |
RU2604987C2 (ru) * | 2011-02-09 | 2016-12-20 | Сони Корпорейшн | Электронное устройство, способ передачи информации стереоскопического изображения электронным устройством и способ приема информации стереоскопического изображения электронным устройством |
US9628872B2 (en) | 2011-02-09 | 2017-04-18 | Sony Corporation | Electronic device, stereoscopic image information transmission method of electronic device and stereoscopic information receiving method of electronic device |
WO2012108393A1 (ja) * | 2011-02-09 | 2012-08-16 | ソニー株式会社 | 電子機器、電子機器における立体画像情報送信方法、および電子機器における立体画像情報受信方法 |
JP2012178706A (ja) * | 2011-02-25 | 2012-09-13 | Toshiba Corp | 映像表示方法および映像表示装置 |
US20120235992A1 (en) * | 2011-03-14 | 2012-09-20 | Kenjiro Tsuda | Stereoscopic image processing apparatus and stereoscopic image processing method |
JP2012220888A (ja) * | 2011-04-13 | 2012-11-12 | Nikon Corp | 撮像装置 |
JP2012227687A (ja) * | 2011-04-19 | 2012-11-15 | Toshiba Corp | 電子機器および映像表示方法 |
US11490119B2 (en) | 2011-09-23 | 2022-11-01 | Qualcomm Incorporated | Decoded picture buffer management |
US10856007B2 (en) | 2011-09-23 | 2020-12-01 | Velos Media, Llc | Decoded picture buffer management |
US10542285B2 (en) | 2011-09-23 | 2020-01-21 | Velos Media, Llc | Decoded picture buffer management |
CN103067680A (zh) * | 2011-10-21 | 2013-04-24 | 康佳集团股份有限公司 | 一种2d转3d视频格式下的osd显示方法及系统 |
JP5363685B2 (ja) * | 2011-10-28 | 2013-12-11 | パナソニック株式会社 | 旧来のフォーマットとの互換を維持しつつも、記録内容を編集することができる記録媒体、再生装置、記録装置 |
US9218109B2 (en) | 2011-10-28 | 2015-12-22 | Panasonic Corporation | Recording medium, playback device, recording device, playback method and recording method for editing recorded content while maintaining compatibility with old format |
WO2013061523A1 (ja) * | 2011-10-28 | 2013-05-02 | パナソニック株式会社 | 旧来のフォーマットとの互換を維持しつつも、記録内容を編集することができる記録媒体、再生装置、記録装置、再生方法、記録方法 |
KR101530826B1 (ko) * | 2014-08-18 | 2015-06-24 | 정관호 | 360도 공간영상 재생방법 및 그 시스템 |
US10205969B2 (en) | 2014-08-18 | 2019-02-12 | Gwan Ho JEONG | 360 degree space image reproduction method and system therefor |
JP2018014158A (ja) * | 2014-10-17 | 2018-01-25 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 記録媒体、再生方法、および再生装置 |
JP2022136168A (ja) * | 2014-10-17 | 2022-09-15 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 再生装置、及び、再生方法 |
JP2017168180A (ja) * | 2014-10-17 | 2017-09-21 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生方法、および再生装置 |
JP2017168181A (ja) * | 2014-10-17 | 2017-09-21 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 記録媒体 |
JP2017191632A (ja) * | 2014-10-17 | 2017-10-19 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 記録媒体 |
US11984139B2 (en) | 2014-10-17 | 2024-05-14 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
JP2017162544A (ja) * | 2014-10-17 | 2017-09-14 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生方法、および再生装置 |
WO2016059760A1 (ja) * | 2014-10-17 | 2016-04-21 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 記録媒体、再生方法、および再生装置 |
US10026440B2 (en) | 2014-10-17 | 2018-07-17 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
JP7139543B1 (ja) | 2014-10-17 | 2022-09-20 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 再生装置、及び、再生方法 |
JP2017168182A (ja) * | 2014-10-17 | 2017-09-21 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生方法、および再生装置 |
JP2016081553A (ja) * | 2014-10-17 | 2016-05-16 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 記録媒体、再生方法、および再生装置 |
JP7113107B2 (ja) | 2014-10-17 | 2022-08-04 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 再生装置、及び、再生方法 |
JP2019079591A (ja) * | 2014-10-17 | 2019-05-23 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生装置、及び、再生方法 |
US11386930B2 (en) | 2014-10-17 | 2022-07-12 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
US10418060B2 (en) | 2014-10-17 | 2019-09-17 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
US11328748B2 (en) | 2014-10-17 | 2022-05-10 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
JP2021090215A (ja) * | 2014-10-17 | 2021-06-10 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生装置、及び、再生方法 |
US10998000B2 (en) | 2014-10-17 | 2021-05-04 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
US10902881B2 (en) | 2014-10-17 | 2021-01-26 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
US10679663B2 (en) | 2014-10-17 | 2020-06-09 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
US10679664B2 (en) | 2014-10-17 | 2020-06-09 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
JP2020098661A (ja) * | 2014-10-17 | 2020-06-25 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生装置、及び、再生方法 |
US10867631B2 (en) | 2014-10-17 | 2020-12-15 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
US10102879B2 (en) | 2014-11-17 | 2018-10-16 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
US10770115B2 (en) | 2014-11-17 | 2020-09-08 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
JP2020047358A (ja) * | 2014-11-17 | 2020-03-26 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生方法および再生装置 |
JP2020021531A (ja) * | 2014-11-17 | 2020-02-06 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生方法および再生装置 |
US10553253B2 (en) | 2014-11-17 | 2020-02-04 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
US11133034B2 (en) | 2014-11-17 | 2021-09-28 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
US11138999B2 (en) | 2014-11-17 | 2021-10-05 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
WO2016079925A1 (ja) * | 2014-11-17 | 2016-05-26 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカ | 記録媒体、再生方法、および再生装置 |
JP2019079590A (ja) * | 2014-11-17 | 2019-05-23 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生方法および再生装置 |
US10236030B2 (en) | 2014-11-17 | 2019-03-19 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
JP2016100039A (ja) * | 2014-11-17 | 2016-05-30 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 記録媒体、再生方法、および再生装置 |
JP2018129115A (ja) * | 2014-11-17 | 2018-08-16 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生装置 |
US9997202B2 (en) | 2014-11-17 | 2018-06-12 | Panasonic Intellectual Property Corporation Of America | Recording medium, playback method, and playback device |
JP2017216038A (ja) * | 2014-11-17 | 2017-12-07 | パナソニック インテレクチュアル プロパティ コーポレーション オブ アメリカPanasonic Intellectual Property Corporation of America | 再生方法および再生装置 |
Also Published As
Publication number | Publication date |
---|---|
BRPI1005171A2 (pt) | 2019-07-02 |
JP5491437B2 (ja) | 2014-05-14 |
CN102172033B (zh) | 2013-07-17 |
CN102172033A (zh) | 2011-08-31 |
JPWO2010146847A1 (ja) | 2012-11-29 |
ES2634941T3 (es) | 2017-09-29 |
DK2445224T3 (en) | 2017-08-21 |
EP2445224B1 (en) | 2017-05-03 |
RU2011120907A (ru) | 2013-07-27 |
RU2520325C2 (ru) | 2014-06-20 |
JP2011120301A (ja) | 2011-06-16 |
EP2445224A4 (en) | 2013-12-04 |
US8121460B2 (en) | 2012-02-21 |
EP2445224A1 (en) | 2012-04-25 |
KR20120036791A (ko) | 2012-04-18 |
MX2011003076A (es) | 2011-04-19 |
TW201130287A (en) | 2011-09-01 |
US20110150421A1 (en) | 2011-06-23 |
ZA201101979B (en) | 2012-04-25 |
JP5451667B2 (ja) | 2014-03-26 |
JP4733785B2 (ja) | 2011-07-27 |
JP2011120302A (ja) | 2011-06-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP5451667B2 (ja) | 3d映像を再生するための半導体集積回路 | |
JP5497679B2 (ja) | 半導体集積回路 | |
JP5038543B2 (ja) | 受信装置 | |
JP5336436B2 (ja) | 集積回路 | |
WO2010143439A1 (ja) | 再生装置、集積回路、記録媒体 | |
WO2010095410A1 (ja) | 記録媒体、再生装置、集積回路 | |
US20110013888A1 (en) | Information recording medium and playback device for playing back 3d images |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 201080002775.2 Country of ref document: CN |
|
ENP | Entry into the national phase |
Ref document number: 2010546132 Country of ref document: JP Kind code of ref document: A |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 10789226 Country of ref document: EP Kind code of ref document: A1 |
|
REEP | Request for entry into the european phase |
Ref document number: 2010789226 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010789226 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 1021/KOLNP/2011 Country of ref document: IN |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2011/003076 Country of ref document: MX |
|
ENP | Entry into the national phase |
Ref document number: 20117007329 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2011120907 Country of ref document: RU |
|
ENP | Entry into the national phase |
Ref document number: PI1005171 Country of ref document: BR Kind code of ref document: A2 Effective date: 20110415 |