WO2010038365A1 - 3d映像に係る記録媒体、再生装置、システムlsi、再生方法、眼鏡、表示装置 - Google Patents
3d映像に係る記録媒体、再生装置、システムlsi、再生方法、眼鏡、表示装置 Download PDFInfo
- Publication number
- WO2010038365A1 WO2010038365A1 PCT/JP2009/004554 JP2009004554W WO2010038365A1 WO 2010038365 A1 WO2010038365 A1 WO 2010038365A1 JP 2009004554 W JP2009004554 W JP 2009004554W WO 2010038365 A1 WO2010038365 A1 WO 2010038365A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- stream
- playback
- view
- file
- information
- Prior art date
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/003—Details of a display terminal, the details relating to the control arrangement of the display terminal and to the interfaces thereto
- G09G5/005—Adapting incoming signals to the display format of the display terminal
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G5/00—Control arrangements or circuits for visual indicators common to cathode-ray tube indicators and other visual indicators
- G09G5/003—Details of a display terminal, the details relating to the control arrangement of the display terminal and to the interfaces thereto
- G09G5/006—Details of the interface to the display terminal
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B20/00—Signal processing not specific to the method of recording or reproducing; Circuits therefor
- G11B20/10—Digital recording or reproducing
- G11B20/12—Formatting, e.g. arrangement of data block or words on the record carriers
- G11B20/1217—Formatting, e.g. arrangement of data block or words on the record carriers on discs
- G11B20/1251—Formatting, e.g. arrangement of data block or words on the record carriers on discs for continuous data, e.g. digitised analog information signals, pulse code modulated [PCM] data
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B27/00—Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
- G11B27/02—Editing, e.g. varying the order of information signals recorded on, or reproduced from, record carriers
- G11B27/031—Electronic editing of digitised analogue information signals, e.g. audio or video signals
- G11B27/034—Electronic editing of digitised analogue information signals, e.g. audio or video signals on discs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/178—Metadata, e.g. disparity information
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/172—Processing image signals image signals comprising non-image signal components, e.g. headers or format information
- H04N13/183—On-screen display [OSD] information, e.g. subtitles or menus
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/189—Recording image signals; Reproducing recorded image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/30—Image reproducers
- H04N13/332—Displays for viewing with the aid of special glasses or head-mounted displays [HMD]
- H04N13/341—Displays for viewing with the aid of special glasses or head-mounted displays [HMD] using temporal multiplexing
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/426—Internal components of the client ; Characteristics thereof
- H04N21/42646—Internal components of the client ; Characteristics thereof for reading from or writing on a non-volatile solid state storage medium, e.g. DVD, CD-ROM
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/432—Content retrieval operation from a local storage medium, e.g. hard-disk
- H04N21/4325—Content retrieval operation from a local storage medium, e.g. hard-disk by playing back content from the storage medium
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/434—Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/434—Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
- H04N21/4343—Extraction or processing of packetized elementary streams [PES]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/434—Disassembling of a multiplex stream, e.g. demultiplexing audio and video streams, extraction of additional data from a video stream; Remultiplexing of multiplex streams; Extraction or processing of SI; Disassembling of packetised elementary stream
- H04N21/4347—Demultiplexing of several video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/816—Monomedia components thereof involving special video data, e.g 3D video
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/91—Television signal processing therefor
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2370/00—Aspects of data communication
- G09G2370/10—Use of a protocol of communication by packets in interfaces along the display data pipeline
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09G—ARRANGEMENTS OR CIRCUITS FOR CONTROL OF INDICATING DEVICES USING STATIC MEANS TO PRESENT VARIABLE INFORMATION
- G09G2370/00—Aspects of data communication
- G09G2370/12—Use of DVI or HDMI protocol in interfaces along the display data pipeline
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B20/00—Signal processing not specific to the method of recording or reproducing; Circuits therefor
- G11B20/10—Digital recording or reproducing
- G11B20/10527—Audio or video recording; Data buffering arrangements
- G11B2020/10537—Audio or video recording
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B20/00—Signal processing not specific to the method of recording or reproducing; Circuits therefor
- G11B20/10—Digital recording or reproducing
- G11B20/10527—Audio or video recording; Data buffering arrangements
- G11B2020/10537—Audio or video recording
- G11B2020/10592—Audio or video recording specifically adapted for recording or reproducing multichannel signals
- G11B2020/10611—3D video data
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B20/00—Signal processing not specific to the method of recording or reproducing; Circuits therefor
- G11B20/10—Digital recording or reproducing
- G11B20/12—Formatting, e.g. arrangement of data block or words on the record carriers
- G11B20/1217—Formatting, e.g. arrangement of data block or words on the record carriers on discs
- G11B2020/1218—Formatting, e.g. arrangement of data block or words on the record carriers on discs wherein the formatting concerns a specific area of the disc
- G11B2020/1224—Formatting, e.g. arrangement of data block or words on the record carriers on discs wherein the formatting concerns a specific area of the disc extent, i.e. a set of sectors which numbers form a continuous ascending sequence
-
- G—PHYSICS
- G11—INFORMATION STORAGE
- G11B—INFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
- G11B2220/00—Record carriers by type
- G11B2220/20—Disc-shaped record carriers
- G11B2220/25—Disc-shaped record carriers characterised in that the disc is based on a specific recording technology
- G11B2220/2537—Optical discs
- G11B2220/2541—Blu-ray discs; Blue laser DVR discs
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/23602—Multiplexing isochronously with the video sync, e.g. according to bit-parallel or bit-serial interface formats, as SDI
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/23605—Creation or processing of packetized elementary streams [PES]
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/20—Servers specifically adapted for the distribution of content, e.g. VOD servers; Operations thereof
- H04N21/23—Processing of content or additional data; Elementary server operations; Server middleware
- H04N21/236—Assembling of a multiplex stream, e.g. transport stream, by combining a video stream with other content or additional data, e.g. inserting a URL [Uniform Resource Locator] into a video stream, multiplexing software data into a video stream; Remultiplexing of multiplex streams; Insertion of stuffing bits into the multiplex stream, e.g. to obtain a constant bit-rate; Assembling of a packetised elementary stream
- H04N21/2365—Multiplexing of several video streams
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/765—Interface circuits between an apparatus for recording and another apparatus
- H04N5/775—Interface circuits between an apparatus for recording and another apparatus between a recording apparatus and a television receiver
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/76—Television signal recording
- H04N5/84—Television signal recording using optical recording
- H04N5/85—Television signal recording using optical recording on discs or drums
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/8042—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components involving data reduction
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/804—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components
- H04N9/806—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal
- H04N9/8063—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback involving pulse code modulation of the colour picture signal components with processing of the sound signal using time division multiplex of the PCM audio and PCM video signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N9/00—Details of colour television systems
- H04N9/79—Processing of colour television signals in connection with recording
- H04N9/80—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback
- H04N9/82—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only
- H04N9/8205—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal
- H04N9/8227—Transformation of the television signal for recording, e.g. modulation, frequency changing; Inverse transformation for playback the individual colour picture signal components being recorded simultaneously only involving the multiplexing of an additional signal and the colour video signal the additional signal being at least another television signal
Definitions
- the present invention belongs to the technical field of 3D video and 2D video recording technology.
- 2D video is an image expressed by pixels on the X-Y plane when the display screen of the display device is regarded as the X-Y plane, and is also called a planar view image.
- 3D video is an image in which the depth in the Z-axis direction is added to the pixels on the XY plane on the screen of the display device.
- 3D video can be viewed by the user by playing both the left-view video that should be viewed with the left eye and the right-view video that should be viewed with the right eye, and exhibiting the stereoscopic effect on these left-view video and right-view video. To be served.
- the user feels that the pixel having the positive Z-axis coordinate is in front of the screen of the display device, and feels that the pixel having the negative Z-axis coordinate exists behind the screen.
- 2D playback device When storing 3D video on an optical disc, playback compatibility with a playback device that can play back only the optical disc storing 2D video (hereinafter referred to as “2D playback device”) is required. If the 2D playback device cannot play back 3D video as 2D video from an optical disc that contains 3D video, it is necessary to produce two types of 3D discs and 2D discs for the same content, resulting in high costs. . Therefore, an optical disc containing 3D video is played back as 2D video on a 2D playback device, and as 2D video or 3D video on a playback device that can play back 2D video and 3D video (hereinafter referred to as “2D / 3D playback device”). It must be reproducible.
- Patent Document 1 As a prior art of a technology for ensuring reproduction compatibility with an optical disc storing 3D video, there is one described in Patent Document 1 below.
- a general recording format is to multiplex a left-view video stream and a right-view video stream at the TS packet level and record them as one transport stream. The bit rate that can be assigned to the stream is reduced. As a result, image quality may be degraded.
- the idea is to store the left-view video stream and the right-view video stream in separate transport stream files, supply the left-view video stream from the optical disc, and supply the right-view video stream from the hard disk.
- TS packets can be supplied from both the optical disk and the hard disk, a certain bit rate can be secured for each of the left-view video stream and the right-view video stream.
- this idea is suitable for use in which a left-view video stream is supplied by an optical disc, a right-view video stream is supplied from a network, and these are combined to be played back by the user.
- Video streams and right-view video streams cannot be stored. Therefore, this idea is not suitable for the business form of recording the left-view video stream and the right-view video stream on one optical disc and selling them as one product or renting them at the store. The industry does not seem to be very welcome.
- the left-view video stream and the right-view video stream are converted into one interleaved format as is realized by so-called multi-angle playback.
- multi-angle playback There is an idea of converting to a port stream file and recording it on an optical disc.
- the extents that make up the left-view video stream and the extents that make up the right-view video stream Since the value of Arrival Time Stamp (ATS) is not continuous, during playback, the ATS increases and decreases, and the irregular change that ATS increases and decreases is repeated. This is different from the change in the ATS of a normal video stream, that is, the change in which the ATS increases monotonously, so that the interleaved transport stream file is used for playback by the 2D playback device by the 2D playback device. There is a risk that normal operation of the 2D playback device cannot be guaranteed.
- ATS Arrival Time Stamp
- An object of the present invention is to provide a recording medium that can ensure reproduction by both the 3D playback device and the 2D playback device.
- a recording medium is a recording medium in which playlist information and a stream file are recorded, and the playlist information includes one or more playback section information,
- the playback section information includes file reference information specifying the stream file storing a video stream,
- the stream file is an interleaved transport stream file and a normal-format transport stream file,
- the interleaved transport stream file is obtained by dividing a transport stream storing a right-view video stream and a plurality of divided parts obtained by dividing a transport stream storing a left-view video stream.
- Each of the plurality of divided parts is configured by alternately arranging, and is identified by the same identification number as the file reference information and an extension indicating that the file is interleaved
- the normal-format transport stream file is one of the left-view video stream and the right-view video stream, and stores a base-view video stream that can be played back independently. It is characterized by being specified by the same identification number and an extension indicating that the format is normal.
- an interleaved transport stream file is specified by the same identification number as the file reference information and an extension indicating that the interleaved transport stream file is an interleaved transport stream file.
- the extent of the interleaved transport stream file identified from the file reference information in the playlist information and the extension indicating that the transport stream file is in interleaved format is read.
- the extent of an interleaved transport stream file can be read only when the output mode is set to the stereoscopic playback mode.
- the extent of the interleaved transport stream file is not read by the 2D playback device, so the unique ATS change of the interleaved transport stream, that is, the ATS increases or decreases, and the ATS increases. Repeated irregular changes such as decreasing will not cause malfunction or instability of the 2D playback device.
- playlist information in which certain file reference information is described is created, at the time of 3D playback, the file name of the file reference information, an extension indicating that it is an interleaved transport stream, An interleaved stream file with a file name will be read and played, and in 2D playback, it has a file name of the file reference information and an extension indicating that it is a normal format transport stream file The transport stream file is read and played. By doing so, it is not necessary to create 3D playlist information and 2D playlist information separately, so the effort of authoring is reduced.
- the extent of the base-view video stream includes file reference information described in the playlist information and an extension indicating that it is a normal format transport stream file. By using it, it can be accessed. Even if the transport stream for planar view is not recorded separately from the transport stream file in the interleave format, both stereoscopic viewing on the 3D playback device and planar view on the 2D playback device can be realized. -It becomes possible to record 3D movie works on a single ROM and supply them to users. There is no need to record a transport stream file for planar viewing separately from an interleaved transport stream file, so a 3D video recording medium and a 2D recording medium are bundled and sold. There is no need to sell the recording medium for 3D video and the recording medium for 2D video as separate products. The distribution industry and the inventory management cost at retail and wholesale stores will not increase, so the movie industry will be able to handle existing 2D movie works unchanged.
- FIG. 1 It is a figure which shows the form about the usage act of a recording medium, a reproducing
- the PES packet sequence shows how the video stream is stored and converted into TS packets and source packets. It is a figure which shows typically how an AV clip is multiplexed.
- the internal structure of the extent obtained by the recording method is shown. It is a figure which shows matching with an extent and a transport stream file. It shows a method of coupling an interleaved transport stream file and a left view transport stream file.
- It is a figure which shows the internal structure of a clip information file.
- a playlist in which 2D play items and 3D play items are not mixed is shown.
- a playlist obtained by adding another sub path to the 3D playlist of FIG. 17 is shown.
- a stream selection table is shown.
- FIG. 18 is a diagram in which left view / right view identification information is added to the 3D playlist of FIG. 17.
- the left view image, the right view image, and the center image indicate two pieces of playlist information that are configured separately.
- the structure of the 2D / 3D playback device is shown. It is a figure which shows the internal structure of the system target decoder 4 and the plane memory set 5a.
- FIG. 19 is a diagram illustrating what events are output to the BD-J application when the state of the playback control engine is switched from 2D playlist playback to 3D playlist playback.
- Fig. 19 is a diagram illustrating what events are output to the BD-J application when a target stream is switched during playback of a 3D playlist by the playback control engine.
- FIG. 3 is a diagram showing an internal configuration of a display device 300 and 3D glasses 400. The display contents in the 3D mode, the left view state of the glasses, and the right view state are shown.
- the display contents in the multi-channel mode, the left view state, and the right view state when the shutters of two glasses are switched instead of switching the left and right shutters are shown.
- the relationship between the pixel number which shows the difference of right and left, and the distance on the screen of a display apparatus is shown.
- FIG. 10 is a diagram in which a 3D playback method supported by a display device is reflected in a player setting register. The relationship between the index table and the movie object is shown. It is a flowchart which shows the process sequence of a stream selection. It is a figure which shows the internal structure of a recording device.
- the right eye and the left eye have a slight difference in appearance between the image seen from the right eye and the image seen from the left eye due to the difference in position. Using this difference, a human can recognize a visible image as a solid. In the case of stereoscopic display, human parallax is used to make a planar image look as if it is stereoscopic.
- the right-eye image and the left-eye image have a difference corresponding to a difference in appearance corresponding to human parallax, and these images are switched at short time intervals. Display, it looks as if a three-dimensional display has been made.
- the short time interval may be a time enough to give an illusion that a person looks three-dimensional by the switching display described above.
- a method for realizing stereoscopic vision there are a method using a holography technique and a method using a parallax image.
- the advantage of this method is that it is possible to realize stereoscopic viewing by simply preparing two viewpoint images for the right eye and left eye.
- the time-separation method is a method in which left-eye video and right-eye video are alternately displayed in the time axis direction, and left and right scenes are superimposed in the brain by an afterimage reaction of the eyes to be recognized as a stereoscopic video.
- FIG. 1A is a diagram showing a form of usage of a recording medium, a playback device, a display device, and glasses.
- the BD-ROM 100 and the playback device 200 as an example of a recording medium constitute a home theater system together with a television 300, 3D glasses 400, and a remote controller 500, and are used by a user.
- BD-ROM 100 supplies, for example, movie works to the home theater system.
- the playback device 200 is connected to the television 300 and plays back the BD-ROM 100.
- the TV 300 provides a user with an interactive operation environment by displaying a playback image of a movie work or displaying a menu or the like.
- the display device 300 realizes stereoscopic vision by wearing 3D glasses 400 by a user.
- the 3D glasses 400 are not necessary.
- the left-eye picture and the right-eye picture are alternately arranged in the vertical direction in the screen at the same time, and the left-eye picture is passed through the lens on the ridge called a lenticular lens on the screen surface of the display device.
- the pixels constituting the image are formed only on the left eye and the pixels forming the picture for the right eye are formed only on the right eye, so that a picture with parallax is shown on the left and right eyes, and stereoscopic viewing is realized.
- the 3D glasses 400 are provided with a liquid crystal shutter, and allow the user to view parallax images by the sequential separation method or the polarization glasses method.
- a parallax image is a set of videos composed of a video image that enters the right eye and a video image that enters the left eye, and allows stereoscopic viewing so that only pictures corresponding to each eye enter the user's eyes.
- FIG. 1B shows the display time of the left-eye video.
- the aforementioned 3D glasses 400 transmit the liquid crystal shutter corresponding to the left eye and shield the liquid crystal shutter corresponding to the right eye.
- FIG. 4C shows the time when the right-eye video is displayed.
- the liquid crystal shutter corresponding to the right eye is made transparent, and the liquid crystal shutter corresponding to the left eye is shielded from light.
- the remote controller 500 is a device that accepts an operation on a hierarchical GUI from a user. For accepting such an operation, the remote controller 500 moves a menu key for calling a menu constituting the GUI and a focus of a GUI component constituting the menu.
- An arrow key, a determination key for performing a confirmation operation on a GUI component constituting the menu, a return key for returning a hierarchical menu to a higher level, and a numerical key are provided.
- the parallax image method is a method of performing stereoscopic viewing so that a video entering the right eye and a video entering the left eye are prepared, and only a picture corresponding to each eye enters.
- Fig. 2 shows the user's face on the left side, and the right side shows an example when the dinosaur skeleton, the target object, is viewed from the left eye, and an example when the dinosaur skeleton, the target object, is viewed from the right eye. ing. If it repeats from the translucency and shading of the right eye and the left eye, the left and right scenes are overlapped by the afterimage reaction of the eyes in the user's brain, and it can be recognized that there is a stereoscopic image on the extension line in the center of the face. .
- an image entering the left eye is referred to as a left eye image (L image), and an image entering the right eye is referred to as a right eye image (R image).
- a moving image in which each picture is an L image is referred to as a left view video
- a moving image in which each picture is an R image is referred to as a right view video.
- a video stream obtained by digitizing and compression-coding the left-view video and the right-view video is referred to as a left-view video stream and a right-view video stream.
- FIG. 3 is a diagram illustrating an example of an internal configuration of a left-view video stream and a right-view video stream for stereoscopic viewing.
- the second row in the figure shows the internal structure of the left-view video stream.
- This stream includes picture data of picture data I1, P2, Br3, Br4, P5, Br6, Br7, and P9. These picture data are decoded according to Decode Time Stamp (DTS).
- DTS Decode Time Stamp
- the first row shows a left eye image.
- the decoded picture data I1, P2, Br3, Br4, P5, Br6, Br7, P9 are reproduced in the order of I1, Br3, Br4, P2, Br6, Br7, P5 according to PTS, and the left-eye image Will be played.
- a picture that does not have a reference picture and performs intra-picture predictive coding using only a picture to be coded is called an I picture.
- a picture is a unit of encoding that includes both a frame and a field.
- a picture that is inter-picture prediction encoded with reference to one already processed picture is called a P picture
- a picture that is inter-picture predictively encoded with reference to two already processed pictures at the same time is called a B picture.
- a picture that is referred to by other pictures in the B picture is called a Br picture.
- a field having a frame structure and a field having a field structure are referred to as a video access unit here.
- the fourth row shows the internal structure of the left-view video stream.
- This left-view video stream includes picture data P1, P2, B3, B4, P5, B6, B7, and P8. These picture data are decoded according to DTS.
- the third row shows a right eye image. The right-eye image is reproduced by reproducing the decoded picture data P1, P2, B3, B4, P5, B6, B7, P8 in the order of P1, B3, B4, P2, B6, B7, P5 according to PTS. Will be played.
- the fifth row shows how the state of the 3D glasses 400 is changed. As shown in the fifth row, the right-eye shutter is closed when the left-eye image is viewed, and the left-eye shutter is closed when the right-eye image is viewed.
- left-view video stream and right-view video stream are compressed by inter-picture prediction encoding using correlation characteristics between viewpoints in addition to inter-picture prediction encoding using temporal correlation characteristics.
- Pictures in the right-view video stream are compressed with reference to pictures at the same display time in the left-view video stream.
- the first P picture of the right-view video stream refers to the I picture of the left-view video stream
- the B picture of the right-view video stream refers to the Br picture of the left-view video stream
- two of the right-view video streams The P picture of the eye refers to the P picture of the left view video stream.
- MVC Multiview Video Coding
- JVT Joint Video Team
- MVC Multiview Video Coding
- the left-view video stream and right-view video stream that have been compression-encoded by MVC, one that can be decoded independently is referred to as a “base-view video stream”.
- the left-view video stream and the right-view video stream are compressed and encoded based on the inter-frame correlation characteristics with the individual picture data constituting the base-view video stream, and the base-view video stream is decoded.
- a video stream that can be decoded is referred to as a “dependent view stream”.
- a mode for producing a recording medium that is, a mode of production of the recording medium will be described.
- FIG. 4 shows the internal structure of a multilayered optical disc.
- the first row shows a BD-ROM that is a multilayered optical disc
- the second row is a drawing in which spiral tracks existing on each recording layer are stretched in the horizontal direction.
- the spiral tracks in these recording layers are treated as one continuous volume area.
- the volume area is composed of a lead-in located at the innermost circumference, a lead-out located at the outermost circumference, a recording area of the first recording layer, a recording area of the second recording layer, and a recording area of the third recording layer existing therebetween. Is done.
- the recording area of the first recording layer, the recording area of the second recording layer, and the recording area of the third recording layer constitute one continuous logical address space.
- the volume area is assigned a serial number in units of accessing the optical disc from the beginning, and this number is called a logical address. Reading data from the optical disk is performed by designating a logical address.
- the sectors where the logical addresses are continuous are also continuous in the physical arrangement on the optical disk. That is, the data of sectors with consecutive logical addresses can be read without seeking.
- the logical address is not continuous in a portion where continuous reading is impossible, such as the boundary of the recording layer.
- file system management information is recorded immediately after the lead-in area, followed by a partition area managed by the file system management information.
- the file system is a mechanism for expressing data on a disc in units called directories or files.
- the file system is recorded in UDF (Universal Disc Disc Format).
- UDF Universal Disc Disc Format
- FAT or NTFS file system
- This file system makes it possible to read out logical data recorded in the same way as a normal PC using a directory and file structure.
- AV file a file storing an AV stream obtained by multiplexing a video stream and an audio stream
- non-AV file a file storing general data other than an AV stream
- PES streams such as video streams and audio streams are multiplexed after being converted into a pack sequence. It's called a file.
- AV files recorded on BD-ROM, BD-RE, and BD-R are the former transport stream files.
- AV files recorded on DVD-Video, DVD-RW, DVD-R, and DVD-RAM are the latter system stream files and are also called Video objects.
- the fourth level shows area allocation in the file system area managed by the file system.
- a non-AV data recording area exists on the inner circumference side.
- An AV data recording area exists immediately after the non-AV data recording area.
- the fifth row shows the recorded contents of these non-AV data recording areas and AV data recording areas.
- In the AV data recording area there are extents constituting the AV file.
- In the non-AV data recording area there are extents constituting non-AV files other than AV files.
- FIG. 5 is a diagram showing an application format of an optical disc based on the file system.
- the BDMV directory is a directory in which data such as AV content and management information handled by the BD-ROM is recorded.
- BDMV directory Under the BDMV directory, there are five subdirectories called “PLAYLIST directory”, “CLIPINF directory”, “STREAM directory”, “BDJO directory”, and “JAR directory”. In the BDMV directory, “index.bdmv” ”And“ MovieObject.bdmv ”are arranged.
- Index.bdmv (file name fixed) associates the title number of multiple titles that can be played back on a BD-ROM with a program file that defines each title, that is, a BD-J object or a movie object.
- An index table is stored.
- the index table is management information relating to the entire BD-ROM, and after the disc is inserted into the playback device, the index.bdmv is first read, whereby the disc is uniquely recognized by the playback device.
- the index table is the highest layer table that defines the title structure such as all titles, top menus, and first play stored in the BD-ROM.
- a program file to be executed first from a general title, a top menu title, and a FirstPlay title is specified.
- the BD-ROM playback device refers to the index table each time a title or menu is called, and executes a predetermined program file.
- the FirstPlay title is set by the content provider, and a program file that is automatically executed when the disc is inserted is set.
- the top menu title specifies a movie object or BD-J object that is called when a command such as “return to menu” is executed by a user operation on the remote controller.
- Index.bdmv has Initial_output_mode information as information relating to stereoscopic viewing. This Initial_output_mode information is information that defines what the initial state of the output mode of the playback device should be when Index.bdmv is loaded, and the output mode desired by the producer is specified in this Initial_output_mode information Can be kept.
- “MovieObject.bdmv (fixed file name)” stores one or more movie objects.
- a movie object is a program file that defines the control procedure to be performed by the playback device in an operation mode (HDMV mode) mainly controlled by a command interpreter.
- HDMV mode operation mode
- a user can use one or more commands, GUI menu calls, and title calls. Contains a mask flag that specifies whether to mask these calls.
- BDJO directory there is a program file (xxxxx.bdjo [“xxxxx” is variable, extension “bdjo” is fixed]) with the extension bdjo.
- This program file stores a BD-J object that defines a control procedure to be performed by a playback device in an operation mode (BD-J mode) mainly controlled by a Java virtual machine that is a bytecode interpreter. Yes.
- This BD-J object includes an “application management table”.
- the “application management table” in the BD-J object is a table for causing the playback apparatus to perform application signaling with the title as a life cycle.
- the application management table includes an “application identifier” for specifying an application to be operated and a “control code” when the title corresponding to the BD-J object becomes the current title.
- a BD-J application whose life cycle is defined by the application management table is particularly referred to as a “BD-J application”.
- the control code indicates that this application is loaded into the heap memory and then automatically started.
- this application is loaded into the heap memory and then another application Wait for a call from, indicating that it should be activated.
- some BD-J applications do not end their operations even if the title ends. Such an application is referred to as a “title boundary application”.
- Java (registered trademark) application is actually the Java (registered trademark) archive file (YYYYY.jar) stored in the JAR directory under the BDMV directory.
- the application is, for example, a Java (registered trademark) application, and includes one or more xlet programs loaded in the heap area (also called work memory) of the virtual machine.
- An application is configured from the xlet program and data loaded in the work memory.
- a “playlist” is a playback path that is defined by specifying playback sections on the time axis of an AV stream and logically specifying the playback order between the playback sections. , Which part plays back only what part and in what order the scene is developed.
- the playlist information file stores playlist information that defines the “type” of the playlist.
- AV playback can be started by a Java TM application for playback control instructing the Java TM virtual machine to generate a JMF player instance that plays back this playlist information.
- a JMF (Java Media Frame work) player instance is actual data generated on a heap memory of a virtual machine based on a JMF player class.
- Extents constituting files existing in the above directories are recorded in the non-AV data area.
- the “STREAM directory” is a directory that stores transport stream files. This directory contains transport streams in the format xxxxx.m2ts ([xxxxx ”is variable, extension“ m2ts ”is fixed]). A file is stored.
- the transport stream file in the STREAM directory stores AV clips.
- An “AV clip” is a “snip piece” of an AV stream, and is an aggregate of packets storing divided parts of a plurality of types of PES streams such as a video stream, an audio stream, and a graphics stream, and has a time stamp. It is continuous and enables seamless AV playback for a certain period.
- the AV clip is sufficient if it guarantees playback for a limited time on the time axis, regardless of whether it is 1 second, 5 seconds, 1 minute, or the like.
- the AV clip also includes packet management information (PCR, PMT, PAT) defined in the European digital broadcasting standard as information for managing and controlling a plurality of types of PES streams.
- packet management information PCR, PMT, PAT
- PCR Program # Clock # Reference
- ATC Arriv Time Clock
- STC System Time Clock
- PMT Program_map_table
- PID of each stream such as video / audio / graphics included in the transport stream file and attribute information of the stream corresponding to each PID, and has various descriptors related to AV clips.
- the descriptor includes copy control information for instructing permission / non-permission of copying of the transport stream file.
- a PMT header describing the length of data included in the PMT is arranged. Behind that, a plurality of descriptors related to the AV clip are arranged.
- the copy control information described above is described as a descriptor.
- a plurality of pieces of stream information regarding each stream included in the transport stream file are arranged after the descriptor.
- the stream information is composed of a stream descriptor in which a stream type, a stream PID, and stream attribute information (frame rate, aspect ratio, etc.) are described in order to identify a compression codec of the stream.
- PAT Program Association Table
- PCR, PMT, and PAT have the role of defining a partial transport stream that constitutes one broadcast program (Program) in the European digital broadcasting standard, and the playback device is one broadcast program in the European digital broadcasting standard.
- the AV clip can be subjected to processing by the decoder as if the partial transport stream that constitutes. This is intended to be compatible with European digital broadcasting standard terminal devices and BD-ROM playback devices.
- a packet storing a segment of a left-view video stream a packet storing a segment of a graphics stream for left view, a packet storing a segment of an audio stream to be played back with these, etc.
- this left-view AV clip includes a base-view video stream and can be reproduced in plan view, the left-view AV clip is referred to as a “2D / left-view AV clip”.
- the left-view video stream is a base-view video stream
- the left-view AV clip including the left-view video stream is a 2D / left-view AV clip.
- a source packet storing a divided part of a right-view video stream a source packet storing a divided part of a graphics stream for right-view view, and a source storing a divided part of an audio stream to be played back with these
- the clip information file stored in the CLIPINF directory is information that shows details of AV clips for each AV clip, such as what kind of packet a left-view AV clip or a right-view AV clip is. This information is read into the memory prior to the playback of the corresponding AV clip and is used as a reference in the playback device while the playback of the AV clip continues.
- the recording method according to the present embodiment is not limited to real-time recording in which AV files and non-AV files as described above are created in real time and directly written in the AV data recording area and non-AV data recording area, but also in the volume area.
- the recording medium according to the present embodiment is also specified by a recording method using real-time recording and a recording method using preformatting recording.
- FIG. 6 is a flowchart showing the processing procedure of the recording method.
- Step S301 is a process for determining the title structure of the BD-ROM.
- title structure information is information that defines a relationship between playback units in the BD-ROM, for example, a relationship between a title, a movie object, a BD-J object, and a playlist, using a tree structure.
- the title structure information includes a node corresponding to the “disc name” of the BD-ROM to be produced, a node corresponding to the “title” that can be played back from Index.bdmv on the BD-ROM, Specify the nodes corresponding to “movie objects and BD-J objects” that make up the title, and “playlist” nodes that are played from the movie objects and BD-J objects.
- the relationship between the title, movie object, BD-J object, and playlist is defined.
- Step S302 is a process of importing video, audio, still image, and caption information used for the title.
- Step S303 is a process of creating BD-ROM scenario data by performing editing processing on the title structure information in accordance with a user operation via the GUI.
- the BD-ROM scenario data is information that causes a playback device to perform playback in units of titles when playing back an AV stream.
- information defined as an index table, movie object, and playlist is included. It corresponds to a scenario.
- the BD-ROM scenario data includes material information that constitutes a stream, playback sections, information indicating playback paths, menu screen layout, transition information from menus, and the like.
- Step S304 is an encoding process, in which encoding is performed based on the BD-ROM scenario data to obtain a PES stream.
- Step S305 is a multiplexing process according to the BD-ROM scenario data. By this process, the PES stream is multiplexed to obtain an AV clip.
- step S306 a database for recording on the BD-ROM is obtained.
- the database here is a generic term for the index table, movie object, play list, BD-J object, etc. defined in the BD-ROM.
- step S307 the Java (TM) program, the AV clip obtained in the multiplexing process, and the BD-ROM database are input, and an AV file and a non-AV file are created in a file system format compliant with BD-ROM.
- Step S308 is a non-AV file writing process of data to be recorded on the BD-ROM
- step S309 is an AV file writing process.
- the video stream, the audio stream, and the graphics stream are converted into PES streams, and then converted into a transport stream.
- Individual TS packets constituting the transport stream are converted into source packets.
- a source packet sequence that includes a second conversion process to be converted and that constitutes a moving image, audio, and graphics is multiplexed.
- the source packet sequence is written to the continuous area of the recording medium as an extent of the AV file.
- the stream to be written is as follows.
- Video stream The video stream shows the primary video and secondary video of the movie.
- the primary video indicates a normal video displayed as a parent image in picture-in-picture
- the secondary video is a video displayed on a small screen in picture-in-picture.
- primary video left-view video and right-view video
- secondary video also includes two types of left-view video and right-view video.
- the video stream is encoded and recorded using a method such as MPEG-2, MPEG-4AVC, or SMPTE VC-1 in addition to the MVC described above.
- Audio stream indicates the main audio part of the movie.
- the audio stream is compressed and encoded and recorded by a method such as Dolby AC-3, Dolby Digital Plus, MLP, DTS, DTS-HD, or linear PCM.
- the primary audio stream is an audio stream that should be a main sound when mixing reproduction is performed
- the secondary audio stream is an audio stream that should be a sub sound when mixing reproduction is performed.
- Presentation Graphics stream is a graphics stream that shows graphics that should be closely synchronized with pictures such as movie subtitles and characters. Streams for multiple languages such as English, Japanese, and French Exists.
- PG stream consists of a series of functional segments: PCS (Presentation Control Segment), PDS (Pallet Definition Segment), WDS (Window Definition Segment), ODS (Object Definition Segment).
- PCS Presentation Control Segment
- PDS Pallet Definition Segment
- WDS Window Definition Segment
- ODS ODS
- ODS Object Define Segment
- WDS Window Definition Segment
- PDS Pulse Definition Segment
- PCS Presentation / Control / Segment
- Such page control includes Cut-In / Out, Fade-In / Out, Color Change, Scroll, Wipe-In / Out, etc., accompanied by page control by PCS, gradually erasing certain subtitles The display effect of displaying the next subtitle can be realized.
- the graphics decoder When playing back a graphics stream, the graphics decoder decodes the ODS belonging to a certain display unit and writes the graphics object to the object buffer, and the graphics object obtained by decoding the ODS belonging to the preceding display unit to the plain memory. By executing the writing process in the pipeline and fully driving the hardware, the above-described precise synchronization is realized.
- -Streams that are not multiplexed in the transport stream file but that display subtitles include text subtitle (textST) streams in addition to PG streams.
- the textST stream is a stream that represents the content of subtitles in character code.
- the combination of the PG stream and the textST stream is called “PGTextST stream” in the BD-ROM standard.
- the Interactive Graphics stream (IG stream) is a graphics stream that realizes interactive control through a remote controller.
- the dialog control defined by the IG stream is a dialog control compatible with the dialog control on the DVD playback device.
- the IG stream is composed of functional segments called ICS (Interactive Composition Segment), PDS (Palette Definition Segment), and ODS (Object Definition Segment).
- ODS Object Definition Segment
- ODS is a functional segment that defines a graphics object. A plurality of graphics objects are gathered to draw a button on the dialogue screen.
- a PDS (Palette Definition Segment) is a functional segment that defines color development when drawing a graphics object.
- the ICS is a functional segment that realizes a state change in which a button state is changed in accordance with a user operation.
- the ICS includes a button command to be executed when a confirmation operation is performed on the button.
- the interactive graphics stream indicates an interactive screen created by arranging GUI parts on the screen.
- FIG. 7A shows in more detail how the video stream is stored in the PES packet sequence.
- the first level in the figure shows a video frame sequence of the video stream.
- the second level shows a PES packet sequence.
- the third level shows a TS packet sequence obtained by converting these PES packet sequences.
- I picture, B picture, and P picture which are a plurality of video presentation units in the video stream, are divided for each picture and stored in the payload of the PES packet. .
- Each PES packet has a PES header, and a PTS (Presentation Time-Stamp) that is a picture display time and a DTS (Decoding Time-Stamp) that is a picture decoding time are stored in the PES header.
- PTS Presentation Time-Stamp
- DTS Decoding Time-Stamp
- FIG. 7B shows the format of the TS packet that is finally written in the AV clip.
- the first level shows a TS packet sequence
- the second level shows a source packet sequence.
- the third level shows an AV clip.
- the TS packet is a fixed-length packet divided into a 4-byte “TS header” having information such as a PID for identifying a stream and a 184-byte “TS payload” for storing data.
- the described PES packet is divided and stored in the TS payload.
- TP_Extra_Header is assigned to the TS packet, and the TS packet is converted into a 192-byte source packet and written to the AV clip.
- TP_Extra_Header describes information such as ATS (Arrival_Time_Stamp).
- ATS indicates the transfer start time of the TS packet to the PID filter.
- SPN source packet number
- FIG. 8 is a diagram schematically showing how left-view AV clips are multiplexed.
- the left-view video stream and the audio stream (first stage) are converted into PES packet sequences (second stage) and converted into source packet sequences (third stage).
- the left-view presentation graphics stream and the left-view interactive graphics (seventh level) are each converted into a PES packet sequence (sixth level) and further converted into a source packet sequence (5th level).
- the source packets constituting the video, audio, and graphics obtained in this way are arranged in the order of the ATS. This is because the source packet should be read into the read buffer according to its ATS.
- a left-view AV clip is obtained.
- This left-view AV clip is set to a size that does not cause the read buffer to underflow, and is a target for recording on a recording medium.
- ATC sequence A collection of source packets in which ATS continues on the Arrival Time Clock (ATC) time axis is called an ATC sequence.
- STC System Time Clock
- DTS decode time stamp
- PTS presentation time stamp
- STC sequence A group of source packets that are used is called an STC sequence.
- FIG. 9 shows the extension obtained by the recording method.
- the first level shows extents EXT_L [i], EXT_L [i + 1], EXT_R [i], and EXT_R [i + 1] constituting the AV file.
- the second row shows a source packet sequence belonging to each extent.
- the extent in the first row is an interleaved arrangement of a group of source packets making up a right-view AV clip and a group of source packets making up a left-view AV clip.
- the interleaved arrangement means that a source packet set constituting a right-view AV clip and a source packet set constituting a left-view AV clip are regarded as one extent as “right view”, “left view”, and “right”. It is recorded with the regularity of “view”, “left view”...
- variables i and i + 1 in parentheses indicate the number of extents to be played back. From this notation, the two extents indicated by the variable i, that is, EXT_L [i], EXT_R [i] are reproduced simultaneously, and the two extents indicated by the variable i + 1, that is, EXT_L [i + 1 ], EXT_R [i + 1] are reproduced simultaneously.
- SEXT_L and SEXT_R are alternately read into two buffers, a right-view read buffer and a left-view read buffer, in the playback device, and provided to the video decoder.
- the sizes of SEXT_L and SEXT_R need to be determined in consideration of the time required for buffering the right-view read buffer and the left-view read buffer.
- the capacity of the read buffer for right view must be determined so as to satisfy the relationship.
- jumping is synonymous with disc seeking. This is because the continuous area that can be secured for recording in the BD-ROM is limited, and the left-view video stream and the right-view video stream are not necessarily recorded side by side, but may be recorded in a skipped area. It is possible.
- TS packet accumulation in the left-view read buffer is performed at a transfer rate of Rud-Rmax2. This means the difference between the output rate Rmax2 from the left-view read buffer and the input rate Rud to the left-view read buffer. Then, the time for filling the left-view read buffer becomes RB2 / (Rud-Rmax2).
- the transfer rate of the right-view read buffer is Rmax1
- all the source packets in the right-view read buffer must be output at the transfer rate of Rmax1 in the accumulation time of the left-view read buffer described above.
- the size of the view read buffer RB1 is RB1 ⁇ Rmax1 ⁇ [2 ⁇ Tjump + RB2 / (Rud-Rmax2)] become.
- the left view read buffer capacity RB2 is calculated.
- RB2 ⁇ Rmax2 ⁇ [2 ⁇ Tjump + RB1 / (Rud-Rmax1)] become.
- the specific values of the memory size of the right-view read buffer and the left-view read buffer are 1.5 Mbytes or less.
- the extent sizes SEXT_R and SEXT_L are the right-view read buffer and the left-view read. It is set to the same size as the buffer size or approximately the same size. This completes the description of how left-view AV clips and right-view AV clips are recorded.
- the internal configuration of the left view AV clip and the right view AV clip will be described.
- the internal configuration of the extents EXT_R [i] and extents EXT_L [i] will be described with reference to the first row in FIG.
- Extent EXT_L [i] is composed of the following source packets.
- the source packet having a packet ID of 0x0100 constitutes a PMT
- the TS packet having a packet ID of 0x1001 constitutes a PCR
- a source packet with a packet ID of 0x1011 is a left-view video stream
- Source packets with packet IDs from 0x1220 to 123F are left-view PG streams
- Source packets having packet IDs of 0x1420 to 143F are left-view IG streams.
- Source packets having PIDs of 0x1100 to 0x111F constitute an audio stream.
- Extent EXT_R [i] is composed of the following source packets.
- Ox1012 TS packets make up a right-view video stream
- Ox1240-125F source packets make up a right-view PG stream
- Ox1440-145F source packets make up a right-view IG stream.
- Partition areas are "area where file set descriptor is recorded”, “area where end descriptor is recorded”, “ROOT directory area”, “BDMV directory area”, “JAR directory area”, “BDJO directory area” , “PLAYLIST directory area”, “CLIPINF directory area”, and “STREAM directory area”, which are areas accessed by the file system.
- these areas will be described.
- File set descriptor includes a logical block number (LBN) indicating a sector in which a file entry of the ROOT directory is recorded in the directory area.
- LBN logical block number
- end descriptor indicates the end of the file set descriptor.
- the plurality of directory areas as described above all have a common internal configuration. That is, the “directory area” is composed of a “file entry”, a “directory file”, and a “file recording area for a lower file”.
- “File entry” includes “descriptor tag”, “ICB tag”, and “allocation descriptor”.
- Descriptor tag is a tag indicating that it is a file entry.
- Allocation descriptor includes a logical block number (LBN) indicating the recording position of the directory file. This completes the description of the file entry. Next, the details of the directory file will be described.
- LBN logical block number
- Database file includes “file identification descriptor for lower directory” and “file identification descriptor of lower file”.
- the “file identifier descriptor of the lower directory” is reference information for accessing the lower directory under its own, the identification information indicating the lower directory, the length of the directory name of the lower directory, and the lower directory It consists of a file entry address indicating in which logical block number the file entry is recorded and the directory name of the lower directory.
- the “file identification descriptor of the lower file” is reference information for accessing the file under its own, the identification information indicating the lower file, the length of the lower file name, and the file entry for the lower file Is composed of a file entry address indicating in which logical block number is recorded and a file name of a lower file.
- the file identifier descriptors in the directory files of these directories indicate in which logical block the file entries of the lower directory and the lower file are recorded. If this file identifier descriptor is followed, The file entry in the BDMV directory can be reached from the file entry in the ROOT directory, and the file entry in the PLAYLIST directory can be reached from the file entry in the BDMV directory. Similarly, file entries in the JAR directory, BDJO directory, CLIPINF directory, and STREAM directory can be reached.
- the “file recording area of a lower file” is an area in which the substance of a lower file under a certain directory is recorded. A “file entry” and one or more “extents” for the lower file are included. It is recorded.
- “File entry” includes “descriptor tag”, “ICB tag”, and “allocation descriptor”.
- Descriptor tag is a tag indicating that it is a file entry. There are types of tags such as a file entry descriptor and a space bitmap descriptor. In the case of a file entry, “261” indicating a file entry is described as a descriptor tag.
- the “allocation descriptor” includes a logical block number (LBN) that indicates the recording position of the extent that constitutes a lower file under a certain directory.
- the allocation descriptor includes data indicating the extent length and a logical block number indicating the recording position of the extent. However, the upper 2 bits of the data indicating the extent length are set to “0” to indicate that they are allocated and recorded extents, and are set to “1” to indicate allocated and unrecorded extents. It shows that it is. When set to “0”, it indicates that the extent is a continuation of the allocation identifier.
- the file entry has a plurality of allocation descriptors for each extent.
- the transport stream file that is the main subject of this application is a file recording area that exists in the directory area of the directory to which the file belongs, and traces the file identifier descriptor in the directory file and the allocation identifier in the file entry. You can access it by going.
- the playback device should make a system call to open the file. These can be read out into the memory.
- File open refers to the process of searching the directory by the file name given at the time of the system call, securing the FCB (File Control Block) if the file exists, and returning the file handle number.
- the FCB is generated by copying the contents of the directory entry of the target file onto the memory.
- the transport stream file with the extension m2ts is specified by the file path using the STREAM directory.
- the transport stream file with the extension ssif is specified by a file path using the STEAM directory and the SSIF directory. This is because these are arranged in the STEAM directory and the SSIF directory.
- FIG. 10 is a diagram showing the correspondence between extents and transport stream files.
- the first row shows 0001.m2ts and 00002m2ts, which are normal transport stream format transport stream files, and the second row shows right-view extents and left-view extents.
- the third level shows an interleaved transport stream file 00001.ssif.
- Dotted arrows h1, h2, h3, and h4 indicate the attribution relationship by the allocation identifier that indicates which file the extents EXT_L [i] and EXT_L [i] belong to. According to the belonging relationship indicated by the arrows h1 and h2, it can be seen that the extents EXT_L [i] and EXT_L [i + 1] are registered as extents of 00001.m2ts.
- extents EXT_R [i], EXT_L [i], EXT_R [i + 1], and EXT_L [i + 1] are registered as extents of 00001.ssif. You can see that As described above, it is understood that the extents EXT_L [i] and EXT_L [i + 1] have the duality that they belong to 00001.ssif and at the same time belong to 00001.m2ts.
- the extension “ssif” is an acronym for StereoScopic Interleave File, and indicates that it is in an interleaved format for stereoscopic playback.
- FIG. 11 shows a method for coupling the corresponding left-view AV clip and right-view AV clip.
- the 2D playback device is the same as the conventional 2D playback method, and only needs to be able to access the 2D playlist and 2D AV clips, so that the 2D playback device can recognize the left-view video stream. Stored in file format.
- the first method is to use the same file name as the 2D playback method so that the left view can be used in 2D playback, and the extension of the interleaved transport stream file is changed.
- 00001.m2ts and 00001.ssif are coupled by the same file name “00001” while one is a 2D system and the other is a 3D system.
- playlists refer only to AV clips in the left view
- existing 2D playback devices can only play back 2D AV clips.
- a 3D-compatible playback device only refers to AV clips with a left view in the playlist, but if there is a file with the same identification number and a different extension, the file is found and used for 3D video. And the left view and right view are output.
- the second method is to divide folders.
- the left view is stored in a folder with an existing folder name (eg STREAM), but the extension right view has the same file name “00001” in a folder with a 3D specific name (eg SSIF) Store it.
- a playlist refers to a file
- the 2D playback device refers only to the file in the “STREAM” folder
- the 3D playback device refers to a file with the same name from the “STREAM” and “SSIF” folders at the same time.
- the left view and the right view can be associated with each other.
- the third method is based on an identification number.
- the identification number of the left view is “00001”
- the identification number of the right view is a constant obtained by adding “1” to the identification number of the left view, that is, an identification number of “0002”.
- This is a method of associating according to a rule.
- the left view identification number plus “1” is used as the right view identification number, but the left view identification number plus “10000” is adopted as the right view identification number. Also good.
- the video to be played back by the existing 2D playback device is described as being a left view, but in practice, either may be used, and in order to identify which video is being played as the prescribed video in the playlist There may be information.
- FIG. 12 is a flowchart showing the processing procedure of the AV file writing process.
- step S401 xxxxx.ssif is created and a file entry is created on the memory of the recording device.
- Step S402 is a determination as to whether or not an empty continuous sector area can be secured. If it can be secured, steps S403 to S408 are executed. If it cannot be secured, an exception process is performed in step S409, and then the recording method is terminated.
- Steps S403 to S408 constitute a loop in which the processes of Steps S403 to S406 and Step S408 are repeated until Step S407 is determined to be No.
- step S403 only SEXT_L [i] is written in the source packet sequence constituting the left-view AV clip in the empty continuous sector area, and in step S404, an allocation identifier indicating the start address and the continuous length in which the source packet sequence has been written is filed. Add to the entry and register as an extent.
- step S405 only SEXT_R [i] is written in the source packet sequence constituting the right-view AV clip in the free continuous sector area, and in step S406, an allocation identifier indicating the start address and the continuous length in which the source packet sequence has been written is filed. Add to the entry and register as an extent.
- Step S407 defines a loop end condition, and determines whether or not an unwritten source packet exists in the right-view AV clip and the left-view AV clip. If it exists, the process moves to step S408 to engage the loop. If it does not exist, the process proceeds to step S410.
- Step S408 is a determination as to whether or not there is a continuous sector area. If it exists, the process proceeds to step S403, and if it does not exist, the process returns to step S402.
- step S410 xxxxx.ssif is closed and the file entry is written to the recording medium.
- step S411 xxxxx.m2ts is created, and a file entry of xxxxx.m2ts is generated in the memory.
- step S412 an allocation descriptor indicating the start address and the continuous length of the extent of the left-view AV clip and the right-view AV clip including the base-view video stream is added to the file entry xxxxx.m2ts.
- step S412 xxxxx.m2ts is closed and a file entry is written.
- FIG. 13 shows the internal structure of the clip information file.
- the clip information file is AV clip management information as shown in the figure, and has a one-to-one correspondence with the AV clip.
- a lead line ch1 shows a close-up of the internal structure of the clip information file.
- the clip information file is composed of “clip information”, “stream attribute information”, “entry map table”, and “3D metadata”.
- the clip information is composed of “system rate”, “reproduction start time”, and “reproduction end time” as indicated by the lead line ch2.
- the system rate indicates a maximum transfer rate for transferring TS packets constituting an AV clip to a PID filter of a system target decoder described later.
- the ATS interval included in the AV clip is set to be equal to or less than the system rate.
- the playback start time is the PTS of the first video frame of the AV clip, and the playback end time is set by adding the playback interval of one frame to the PTS of the video frame at the end of the AV clip.
- FIG. 14 is a diagram showing stream attribute information in the clip information file.
- the lead line ah1 in the figure shows a close-up of the internal structure of the stream attribute information.
- the stream attribute information indicates what kind of attribute is included.
- attribute information about each stream included in the AV clip is registered for each PID.
- FIG. 15 shows an entry map table in the clip information file. *
- FIG. 1 shows the schematic structure of the entry map table.
- a leader line eh1 shows a close-up of the internal structure of the entry map table.
- Entry map header information stores information such as the PID of the video stream indicated by the entry map and the number of entry points.
- “Extent start type” indicates which one of the left-view video stream and the right-view video stream is located first.
- Entry map for PID 0x1011
- Entry map for PID 0x1012
- the is_angle_change flag is a flag indicating whether or not decoding can be performed independently from the entry point. If the video stream is encoded with MVC or MPEG4-AVC and an IDR picture exists at the entry point, this flag is set to ON. If a Non-IDR picture exists at the entry point, this flag is set to off.
- FIG. 16 shows registration of entry points using an entry map.
- the first level shows a time axis defined by the STC sequence.
- the second level shows an entry map in clip information.
- the third level shows a source packet sequence constituting the STC sequence.
- the entry map specifies a source packet of n1 in the ATC sequence
- the entry map specifies a source packet of n21 in the ATC sequence
- the playback device can perform arbitrary access on the time axis of the video stream.
- the file position of the AV clip corresponding to the point can be specified. For example, at the time of special playback such as fast-forward / rewind, it is possible to efficiently perform processing without analyzing an AV clip by specifying, selecting and playing back an I picture registered in the entry map.
- 3D metadata is a metadata group that defines various information necessary for stereoscopic playback, and includes multiple offset entries. Each offset entry is associated with a plurality of PIDs and a plurality of display times. When a PES stream of a certain PID is played back, it is possible to specify for each PID and each display time how many offsets should be used at a plurality of display times of that PES stream to realize stereoscopic viewing. .
- FIG. 17 shows a playlist in which 2D play items and 3D play items are not mixed. By eliminating the mixing, the playback environment of the playback device is not switched.
- the playlist in this figure is composed of a “main path” and one or more “sub paths”.
- the main path consists of one or more play items.
- the play item is composed of play items # 1, # 2, and # 3.
- Sub path indicates a series of playback paths that are played together with the main path, and IDs (sub path IDs) are assigned in the order of registration in the playlist.
- the sub path ID is used to identify the sub path.
- the sub play item is composed of one or more pieces of sub play item information.
- “play item” includes a stream selection table.
- the stream selection table is information indicating the stream numbers of elementary streams that are permitted to be played in the play item and the sub play item. Details of the play list information, play item information, sub play item information, and stream selection table are left to the description of the later embodiment.
- AV clips # 1, # 2, and # 3 are AV clips that are played back as 2D video, and AV clips that are played back as a left view when played back as 3D video.
- AV clips # 4, # 5, and # 6 are AV clips that are played back as a right view when played back as 3D video.
- the main path of the 2D playlist refers to AV clips # 1, # 2, and # 3 that store left-view AV clips as indicated by reference numerals rf1, 2, and 3.
- a sub-path for storing a right view is prepared together with a main path including a play item referring to a left view AV clip as indicated by reference numerals rf4, rf5, and rf5.
- AV clips # 4, # 5, and # 6 that store the right view are referenced as indicated by reference numerals rf7, 8, and 9.
- This sub path is set to synchronize with the main path on the time axis.
- the AV clip storing the left view can be shared between the 2D playlist and the 3D playlist, and in the 3D playlist, the left view and the right view can be associated in synchronization on the time axis.
- the play item information # 1 to # 3 refer to AV clips # 1 to # 3 that are common AV clips, so these 3D playlist and 2D playlist A common description is sufficient for describing playlist information that defines a list (see symbols df1 and df2). Therefore, if playlist information that realizes this 3D playlist is described, it functions as a 3D playlist when the output mode of the playback device is the stereoscopic output mode, and the output mode of the playback device is the 2D output mode. When it is, it will function as a 2D playlist.
- the 2D playlist and 3D playlist in this figure are interpreted as 2D playlists and 3D playlists according to the output mode of the playback device that interprets one playlist information. It is possible to reduce the time and effort of the person who performs the authoring.
- FIG. 18 shows a playlist obtained by adding another sub path to the 3D playlist of FIG.
- AV clips # 7, # 8, and # 9 are referenced.
- Multiple light views defined by providing two or more sub-path information are multiple light views with different angles for viewing the subject from the right eye, and AV clips are prepared for the number of angles for each angle.
- a sub path is provided.
- AV clips # 1, # 2, and # 3 and AV clips # 4, # 5, and # 6 both store the right view, but the angle at which the subject is viewed from the right eye is different.
- the sub-path with the sub-path ID “0” refers to the AV clips # 4, # 5, and # 6 as indicated by the symbols rf7, 8, and 9, and the sub-path with the sub-path ID “1” has the AV clips # 7, # 8, Reference # 9 is made as indicated by reference numerals rf10, rf11, and rf12.
- a stereoscopic video can be displayed using a parallax image that is comfortable for the user by switching the sub-path to be reproduced in synchronization with the main path for storing the left view. It becomes possible.
- Playlist information that realizes this 3D playlist also functions as a 3D playlist when the output mode of the playback device is the stereoscopic output mode, and when the output mode of the playback device is the 2D output mode, the 2D playlist Will function as. If the 2D playlist and 3D playlist in FIG. 18 describe one piece of playlist information, it is interpreted as a 2D playlist and 3D playlist according to the output mode of the playback device that interprets the playlist information. Since the optimum output mode is made, it is possible to reduce the trouble for the author.
- FIG. 19 is a diagram showing the data structure of playlist information.
- the playlist information includes “main path information”, “sub path information table”, “extension data”, “ Mark information ”.
- a lead line mp1 shows a close-up of the internal configuration of the main path information.
- MainPath is defined from a plurality of PlayItem information # 1,..., #N as indicated by an arrow mp1.
- PlayItem information defines one logical playback section that constitutes the MainPath.
- the structure of the PlayItem information is closed up by a lead line mp2.
- the PlayItem information includes “Clip_Information_file_name” indicating the file name of the playback section information of the AV clip to which the IN point and Out point of the playback section belong, and “Clip_codec_identifier” indicating the encoding method of the AV clip, “Is_multi_angle” indicating whether or not the PlayItem constitutes a multi-angle, “connection_condition” indicating the connection state between this PlayItem (current PlayItem) and the previous PlayItem (previousPlayItem), and this PlayItem “Ref_to_STC_id [0]” uniquely indicating the STC_Sequence that is present, time information “In_time” indicating the start point of the playback section, time information “Out_time” indicating the end point of the playback section, and user operation to be masked in this PlayItem “UO_mask_table”, “STN_table”, “left view / right view identification information”, “multi_clip_entry”
- STN_table (STream Number_table) is a table that assigns a logical stream number to a combination of a stream entry including a packet ID and a stream attribute. The order of combinations of stream entries and stream attributes in the STN_table indicates the priority order of the corresponding streams. This STN_table is for 2D playback, and there is another STN_table for 3D playback.
- “Left-view / right-view identification information” is base-view video stream designation information for designating which of the left-view video stream and the right-view video stream is a base-view video stream. The stream indicates that it is a base-view video stream, and if it is 1, it indicates that the right-view video stream is a right-view video stream.
- connection_condition indicates the forward play item and the connection type.
- connection_condition of the play item is “1”
- the AV clip indicated by the play item indicates that seamless connection with the AV clip indicated by the play item before the play item is not guaranteed.
- connection_condition of the play item is “5” or “6”, it is guaranteed that the AV clip indicated by the play item is seamlessly connected to the AV clip indicated by the play item before the play item.
- connection_condition When the connection_condition is “5”, the continuity of STC between play items may be interrupted, that is, the start of the AV clip at the beginning of the AV item of the play item after connection than the video display time at the end of the AV clip of the pre-connection play item
- the video display time start time may be discontinuous.
- the AV clip of the pre-connection play item is input to the PID filter of the system target decoder, which will be described later, and the AV clip of the post-connection play item is input to the PID filter of the system target decoder and played, the system target It is necessary to create an AV clip so that the decoding of the decoder does not fail.
- the audio end frame of the AV clip of the pre-connection play item and the audio start frame of the post-connection play item must overlap on the playback time axis.
- connection_condition is “6”
- the AV clip of the pre-connection play item and the post-connection play item AV clip must be reproducible as a single AV clip. That is, the STC continues and the ATC also continues between the AV clip of the play item before connection and the AV clip of the play item after connection.
- Multi_clip_entries is information for specifying which AV clip represents each angle video when a multi-angle section is formed by a play item.
- FIG. 20 shows the internal structure of the subpath information table.
- a lead line su1 shows a close-up of the internal configuration of the subpath information.
- the subpath information table includes a plurality of subpath information 1, 2, 3,. These subpath information is a plurality of instances derived from one class structure, and the internal configuration is common.
- a lead line su2 shows a close-up of a common internal configuration of Subpath information.
- each Subpath information includes SubPath_type indicating the type of subpath and one or more subplay item information (... sub play item information # 1 to VOB # m ).
- a lead line su3 shows a close-up of the internal structure of SubPlayItem.
- the sub play item information includes “Clip_information_file_name”, “Clip_codec_identifier”, “ref_to_STC_id [0]”, “SubPlayItem_In_time”, “SubPlayItem_Out_time”, “sync_PlayItem_id”, and “sync_start_PTS_of_PlayItem”.
- the internal configuration of SubPlayItem will be described.
- “Clip_information_file_name” is information for uniquely specifying a SubClip corresponding to a SubPlayItem by describing the file name of the clip information.
- “Clip_codec_identifier” indicates the encoding method of the AV clip.
- SubPlayItem_In_time is information indicating the start point of SubPlayItem on the playback time axis of SubClip.
- SubPlayItem_Out_time is information indicating the end point of SubPlayItem on the playback time axis of SubClip.
- Sync_PlayItem_id is information that uniquely designates the PlayItems constituting the MainPath that should be synchronized with the SubPlayItem.
- SubPlayItem_In_time exists on the playback time axis of Play Item specified by this sync_PlayItem_id.
- Sync_start_PTS_of_PlayItem indicates where the start point of the SubPlayItem specified by SubPlayItem_In_time exists on the playback time axis of Play Item specified by sync_PlayItem_id with a time accuracy of 45 KHz.
- FIG. 21 shows what playback sections are defined for the left view and the right view. This figure is drawn on the basis of FIG. 16, and In_Time and Out_Time of PlayItem are drawn on the time axis of the second stage of this figure. In_Time and Out_Time of SubPlayItem are drawn on the time axis of the first row.
- the third to fifth stages are the same as the third to fifth stages in FIG.
- the left view and right view I pictures are at the same point in time. This completes the description of the data structure of the playlist information.
- Entry mark information can be given within the playback section defined by the play item, is attached to a position that can be a playback start point for the play item, and is used for cue playback.
- chapter reproduction can be performed by adding an entry mark to a position at the beginning of a chapter in a movie title.
- Extension data is an extension part unique to 3D playlists that is not compatible with 2D playlists, and STN_table_SS # 1 to #N are stored here.
- STN_table_SS is a table that corresponds to each of a plurality of pieces of play item information, and assigns logical stream numbers to combinations of stream entries and stream attributes for 3D playback. The order of the combination of stream entry and stream attribute in STN_table_SS indicates the priority order of the corresponding stream.
- a stream selection table is configured by combining the STN_table in the play item information and the STN_table_SS in the extension data.
- FIG. 22A shows a stream selection table.
- the stream selection table is composed of a plurality of stream entries. As shown in parentheses “ ⁇ ” in the figure, this stream entry includes one defined in STN_table and one defined in STN_table_SS.
- STN_table stream entry 2D audio / PG / IG that can be played when 2D output mode is set is registered. Therefore, in the STN_table, there are a 2D video stream entry group, a 2D audio stream entry group, a 2DPG stream entry group, and a 2DIG stream entry group. Among these stream groups, a video stream, an audio stream, a PG Packet identifiers of streams and IG streams can be described.
- 3D audio / PG / IG that can be played back when the stereoscopic playback mode is set is registered. Therefore, 3D video stream entry group, 3D audio stream entry group, 3DPG stream entry group, 3DIG stream entry group, and stream combination information exist in STN_table_SS, and video stream is included in these stream entry groups.
- Packet identifiers of audio stream, PG stream, and IG stream can be described.
- (B) in the figure shows a common configuration of stream entries.
- the stream entry in the stream selection table is composed of “stream selection number”, “stream path information”, and “stream identification information”.
- Stream selection number is a number that is sequentially incremented from the beginning of stream entry 1 included in the stream selection table, and is used for stream identification in the playback device.
- Stream identification information is information such as PID, and indicates the stream multiplexed in the AV clip to be referred to.
- attribute information of each stream is recorded at the same time.
- the attribute information is information indicating the nature of each stream, and includes, for example, language attributes in the case of audio, presentation graphics, and interactive graphics.
- the stream entry of the left view video stream and the stream entry of the right view video stream have the same value, for example, frame rate, resolution, video format, and the like.
- the stream entry may be provided with a flag indicating whether it is a left-view video stream or a right-view video stream.
- the description up to this point is based on the premise that the left view is the main display and the left view is displayed for 2D display.
- the right view may be the main display. It is assumed that the base-view video stream is in accordance with information for determining which one of the left eye / right view is main in the playlist and whether it is displayed in 2D.
- the discrimination information is left view / right view identification information.
- left-view video is created as a 2D video in a studio. However, it may be better to create a left-view video as a 2D video. Since such a possibility exists, left view / right view identification information indicating which of the left view and the right view is set as the base view can be set for each piece of play item information.
- FIG. 23 is a diagram in which left view / right view identification information is added to the 3D playlist of FIG. With this information, if the right-view video stream is specified as the base-view video stream, even if the right-view video stream is specified by the sub-path information, the right-view video stream is input to the video decoder first. Uncompressed picture data is obtained. Then, motion compensation is performed based on uncompressed picture data obtained by decoding the right-view video stream. This gives flexibility in choosing which can be the base view.
- the identification information of each stream and the left view / right view can be used for output to the display device, and the display device side uses each to distinguish the two streams.
- the display of the glasses and the display device cannot be synchronized unless you know whether the main image referenced by the play item is a left view or a right view.
- a switching signal is sent to the eyeglasses so that the light on the left eye side of the shutter-type spectacles is transmitted while the right-view side light of the shutter-type spectacles is transmitted when the right view is displayed.
- This left view / right view identification information is based on the premise that either a left view or a right view among parallax images can be reproduced as a planar view video. However, depending on the content of the parallax image, it may not be suitable for use as a planar view image in this way.
- FIG. 24 shows two pieces of playlist information in which a left view image, a right view image, and a center image are configured separately.
- the lower right in the figure shows a stereoscopic image aiming at a screen effect in which a dinosaur in the video approaches the user's eyes.
- This stereoscopic image is composed of an L image and an R image as described above.
- the L image and the R image constituting the stereoscopic image having a large pop-up effect display the object (the dinosaur in this figure) in the image popping out from the side.
- playlist information that designates a video stream representing the center image is selected as the current playlist.
- 0005.mpls designates a left-view video stream and a right-view video stream with a high pop-out degree as main path information and sub-path information.
- 00003.mpls specifies the video stream of the center image by the main path.
- the movie object in the upper left of the figure is described to select and play back either 00005.mpls or 00003.mpls according to the 3D playback capability (3D-Capability) in the playback device ( If statement in the figure).
- FIG. 25 shows the configuration of the 2D / 3D playback device.
- the 2D / 3D playback device includes a BD drive 1, a read buffer 2a, a read buffer 2b, a switch 3, a system target decoder 4, a plane memory set 5a, a plane synthesis unit 5b, an HDMI transmission / reception unit 6, a playback control unit 7, and a management information memory. 9, a register set 10, a program execution unit 11, a program memory 12, an HDMV module 13, a BD-J platform 14, a middleware 15, a mode management module 16, a user event processing unit 17, a local storage 18, and a nonvolatile memory 19. Yes.
- the BD-ROM drive 1 reads data from the BD-ROM disc based on a request from the playback control unit 7 in the same manner as the 2D playback device, but the AV clip read from the BD-ROM disc is read from the read buffer 2a. It is transferred to the buffer 2b.
- the playback control unit 7 sends a read request instructing to alternately read 2D / left-view AV clips and right-view AV clips in extent units.
- the BD-ROM drive 1 reads extents constituting the 2D / left-view AV clip to the read buffer 2a, and reads extents constituting the right-view AV clip to the read buffer 2b.
- both 2D / left-view AV clips and right-view AV clips need to be read at the same time. Therefore, speed performance higher than the BD-ROM drive of 2D playback devices is required.
- the read buffer 2a is a buffer composed of a dual port memory or the like for storing 2D / left-view AV clip data read by the BD-ROM drive 1.
- the read buffer 2b is a buffer composed of a dual port memory or the like for storing right-view AV clip data read by the BD-ROM drive 1.
- the switch 3 is a switch for switching the data input source for the read buffer to either the BD-ROM drive 1 or the local storage 18.
- the system target decoder 4 performs a demultiplexing process on the source packet read to the read buffer 2a and the source packet read to the read buffer 2b to perform a stream decoding process.
- the plane memory set 5a is composed of a plurality of plane memories.
- the plane memory includes a left view video plane, a right view video plane, a secondary video plane, an interactive graphics plane (IG plane), and a presentation graphics plane (PG plane).
- IG plane interactive graphics plane
- PG plane presentation graphics plane
- the plane synthesis unit 5b instantly superimposes the left-view video plane, right-view video plane, secondary video plane, IG plane, PG plane, and GFX plane and displays them on a screen such as a TV.
- the plane synthesizing unit 5 crops the secondary video plane, the PG plane, and the IG plane alternately for the left eye and for the right eye using 3D metadata, and synthesizes with the left view video plane or the right view video plane.
- the combined video is transferred to the GFX plane superimposing process.
- the plane composition unit 5 crops the graphics in the IG plane alternately for the left eye and the right eye using the offset information specified from the API, and the left view video plane, the right view video plane, the secondary video plane, the PG plane, and the IG plane. An image with and superimposed is output to a television.
- the HDMI transmission / reception unit 6 includes an interface compliant with, for example, the HDMI standard (HDMI: High Definition Multimedia Interface), and performs transmission / reception so as to comply with the HDMI standard with a device (in this example, the TV 300) connected to the playback device.
- the picture data stored in the video and the uncompressed audio data decoded by the audio decoder 9 are transmitted to the television 300 via the HDMI transmission / reception unit 6.
- the television 300 holds, for example, information regarding whether it is compatible with stereoscopic display, information regarding resolution capable of planar display, and information regarding resolution capable of stereoscopic display, and there is a request from the playback device via the HDMI transmission / reception unit 6.
- the television 300 returns the requested required information (for example, information regarding whether it supports stereoscopic display, information regarding resolution capable of planar display, information regarding resolution capable of stereoscopic display) to the playback device.
- the requested required information for example, information regarding whether it supports stereoscopic display, information regarding resolution capable of planar display, information regarding resolution capable of stereoscopic display
- information regarding whether or not the television 300 is compatible with stereoscopic display can be acquired from the television 300 via the HDMI transmission / reception unit 6.
- the playback control unit 7 includes a playback engine 7a and a playback control engine 7b.
- the playback of the 3D playlist is instructed by the program execution unit 11 or the like, the 2D / left view of the play item to be played back in the 3D playlist
- the AV clip is specified, and the right-view AV clip of the sub play item of the 3D sub path that is played back in synchronization with the play item is specified.
- it interprets the entry map of the corresponding clip information file, and based on the extent start type that indicates which extent is located first, the extent of the 2D / left view AV clip from the playback start point and the right view AV Requests the BD-ROM drive 1 to read the extents of the clips alternately.
- the playback control unit 7 When starting reproduction, the first extent is read into the read buffer 2a or the read buffer 2b, and then transfer from the read buffer 2a and the read buffer 2b to the system target decoder 4 is started. In addition, when playing back a 3D playlist, the playback control unit 7 notifies the plane synthesis unit 5 of 3D metadata included in the clip information file corresponding to the 3D / left-view AV clip.
- the playback engine 7a executes an AV playback function.
- AV playback function is a group of functions followed from DVD playback device and CD playback device, playback start, playback stop, pause, release of pause, release of still image function, fast-forward that specifies the playback speed as immediate value, This is a process such as rewind, audio switching, secondary video picture data switching, and angle switching with the playback speed specified as an immediate value.
- the playback control engine 7b executes a playlist playback function in response to a function call from a command interpreter that is an HDMV mode operation subject and a Java platform that is an BD-J mode operation subject.
- the playlist playback function refers to performing playback start and playback stop according to current playlist information and current clip information constituting the current playlist among the AV playback functions described above.
- the management information memory 9 is a memory for storing current playlist information and current clip information.
- Current playlist information refers to information that is currently processed among multiple playlist information that can be accessed from a BD-ROM, a built-in media drive, or a removable media drive.
- Current clip information refers to information that is currently processed among a plurality of clip information that can be accessed from a BD-ROM, a built-in media drive, or a removable media drive.
- a playback status / setting register (Player Status / Setting Register) set 10 is a playback status register that stores the playback status of a playlist, a playback setting register that stores configuration information indicating the configuration of the playback device, and an arbitrary content used by the content. It is a collection of registers including general-purpose registers that can store information.
- the reproduction state of the playlist indicates a state such as which AV data is used in various AV data information described in the playlist and which position (time) of the playlist is being reproduced.
- the playback control engine 14 stores the contents in the register set 10. Also, the value specified by the application can be stored or the stored value can be stored in the application according to the instruction from the command interpreter that is the HDMV mode operation subject or the Java platform that is the BD-J mode operation subject. It is possible to pass
- the program execution unit 11 is a processor that executes a program stored in a BD program file. The operation is performed according to the stored program, and the following control is performed. (1) Command the playback control unit 7 to play a playlist. (2) Transfer PNG / JPEG for menu and game graphics to the system target decoder and display it on the screen. These can be performed freely according to the creation of the program, and how to control them depends on the programming process of the BD-J application by the authoring process.
- the program memory 12 is a memory that stores a current dynamic scenario and is used for processing by an HDMV module that is an HDMV mode operating entity and a Java platform that is an BD-J mode operating entity.
- the current dynamic scenario is an index.bdmv, BD-J object, or movie object currently recorded on the BD-ROM that is currently being executed.
- the dynamic scenario memory 12 includes a heap memory.
- the heap memory is a byte area of the system application, a byte code of the BD-J application, a system parameter used by the system application, and a stack area in which an application parameter used by the BD-J application is arranged.
- the HDMV module 13 is in HDMV mode. This is a DVD virtual player that is the main subject of the operation, and the HDMV mode execution subject. This module has a command interpreter and executes HDMV mode control by decoding and executing navigation commands constituting the movie object. Since navigation commands are described in a syntax similar to DVD-Video, DVD-Video-like playback control can be realized by executing such navigation commands.
- the BD-J platform 14 is a Java platform that is the main operating entity of the BD-J mode. It is implemented and consists of a class loader, a bytecode interpreter, and an application manager.
- the class loader is one of the system applications, and loads the BD-J application by reading the byte code from the class file existing in the JAR archive file and storing it in the heap memory 31.
- Bytecode interpreter is a so-called Java virtual machine that converts the bytecodes that make up the BD-J application stored in the heap memory and the bytecodes that make up the system application into native codes, which are executed by the MPU.
- the application manager is one of the system applications. Based on the application management table in the BD-J object, the BD-J application application signaling, such as starting the BD-J application or ending the BD-J application I do. This completes the description of the internal configuration of the BD-J platform.
- the middleware 15 is an operating system for embedded software, and includes a kernel and device drivers.
- the kernel provides a playback device-specific function to the BD-J application in response to an application programming interface (API) call from the BD-J application.
- API application programming interface
- hardware control such as activation of an interrupt handler unit by an interrupt signal is realized.
- the mode management module 16 holds Index.bdmv read from the BD-ROM, built-in media drive, or removable media drive, and performs mode management and branch control.
- the mode management by the mode management module is a module assignment that determines whether the BD-J platform 22 or the HDMV module executes a dynamic scenario.
- the user event processing unit 17 requests the program execution unit 16 and the playback control unit 7 to execute processing in response to a user operation through the remote controller. For example, when a button is pressed with the remote controller, the program execution unit 16 is requested to execute a command included in the button. For example, when the fast forward / rewind button is pressed on the remote controller, the playback control unit 7 is instructed to execute fast forward / rewind processing for the AV clip of the playlist currently being played back.
- the local storage 18 includes a built-in media drive for accessing the hard disk and a removable media drive for accessing the semiconductor memory card, and is used for storing downloaded additional content and data used by the application.
- the storage area for additional content is divided for each BD-ROM, and the area that an application can use to hold data is divided for each application.
- the non-volatile memory 19 is a recording medium such as a readable / writable memory, and is a medium capable of holding recorded contents even when power is not supplied, such as a flash memory or FeRAM. This is used for backup of the stored contents in the register set 10.
- FIG. 26 is a diagram showing an internal configuration of the system target decoder 4 and the plane memory set 5a.
- the system target decoder 4 and the plane memory set 5a include an ATC counter 21, a source depacketizer 22, a PID filter 23, an STC counter 24, an ATC counter 25, a source depacketizer 26, a PID filter 27, Primary video decoder 31, left-view video plane 32, right-view video plane 33, secondary video decoder 34, secondary video plane 35, PG decoder 36, PG plane 37, IG decoder 38, IG plane 39, primary audio decoder 40, secondary audio It comprises a decoder 41, a mixer 42, a rendering engine 43, and a GFX plane 44.
- the ATC counter 21 generates Arrival Time Clock (ATC) and adjusts the operation timing in the playback device.
- ATC Arrival Time Clock
- the source depacketizer 22 follows the recording rate of the AV clip at the moment when the ATC value generated by the ATC counter becomes the same as the ATS value of the source packet. Then, only the TS packet is transferred to the PID filter. In this transfer, the input time to the decoder is adjusted according to the ATS of each source packet.
- the PID filter 23 selects a TS packet output from the source depacketizer 22 whose TS packet PID matches the PID required for reproduction, according to the PID, the primary video decoder 31 and the secondary video decoder. 34, the IG decoder 38, the PG decoder 36, the primary audio decoder 40, and the secondary audio decoder 41.
- STC counter 24 generates System Time Clock (STC) and adjusts the operation timing of each decoder.
- STC System Time Clock
- the ATC counter 25 generates Arrival Time Clock (ATC) and adjusts the operation timing in the playback device.
- ATC Arrival Time Clock
- the source depacketizer 26 follows the system rate of the AV clip at the moment when the ATC value generated by the ATC counter becomes the same as the ATS value of the source packet. Then, only the TS packet is transferred to the PID filter. In this transfer, the input time to the decoder is adjusted according to the ATS of each source packet.
- the PID filter 27 selects a TS packet output from the source depacketizer 26 that has the PID of the TS packet that matches the PID described in the stream selection table of the current play item, according to the PID. Forward to.
- the primary video decoder 31 decodes the left-view video stream and writes an uncompressed video frame as a decoding result to the byte code interpreter 32.
- the left-view video plane 32 is a plane memory that can store picture data at a resolution of 1920 ⁇ 2160 (1280 ⁇ 1440), for example.
- the right-view video plane 33 is a plane memory that can store picture data at a resolution of 1920 ⁇ 2160 (1280 ⁇ 1440), for example.
- the secondary video decoder 34 has the same configuration as the primary video decoder, decodes the input secondary video stream, and writes a picture to the secondary video plane at the display time (PTS) timing.
- PTS display time
- the secondary video plane 35 outputs secondary video picture data data obtained by decoding the secondary video stream from the system target decoder 4.
- the PG decoder 36 extracts and decodes the presentation graphics stream from a plurality of TS packets input from the source depacketizer, and writes uncompressed graphics data to the PG plane at the display time (PTS) timing.
- PTS display time
- the PG plane 37 stores an uncompressed graphics object obtained by decoding the presentation graphics stream.
- the IG decoder 38 extracts and decodes an interactive graphics stream from a plurality of TS packets input from the source packetizer, and writes an uncompressed graphics object to the IG plane at the display time (PTS) timing.
- PTS display time
- the IG plane 39 stores graphics data obtained by decoding the interactive graphics stream.
- the primary audio decoder 40 decodes the primary audio stream.
- the secondary audio decoder 41 decodes the secondary audio stream.
- the mixer 42 synthesizes the decoding result of the primary audio decoder 40 and the decoding result of the secondary audio decoder 41.
- the rendering engine 43 decodes graphics data used by the BD-J application for menu drawing, such as JPEG and PNG.
- the GFX plane 44 is a plane memory that is written after graphics data such as JPEG and PNG is decoded.
- the primary video decoder 31 includes TB51, MB52, EB53, TB54, MB55, EB56, video decoder 57, buffer switch 58, DPB59, and picture switch 60.
- the Transport Buffer (TB) 51 is a buffer that temporarily accumulates TS packets as they are when TS packets including a left-view video stream are output from the PID filter 23.
- a Multiplexed Buffer (MB) 52 is a buffer for temporarily storing PES packets when outputting a video stream from TB to EB.
- MB Multiplexed Buffer
- Elementaly Buffer (EB) 53 is a buffer in which a video access unit in an encoded state is stored. The PES header is removed when data is transferred from MB to EB.
- the Transport Buffer (TB) 54 is a buffer that temporarily accumulates TS packets as they are when TS packets including right-view video streams are output from the PID filter.
- Multiplexed Buffer (MB) 55 is a buffer for temporarily storing PES packets when outputting a video stream from TB to EB.
- MB Multiplexed Buffer
- Elementaly Buffer (EB) 56 is a buffer in which a video access unit in an encoded state is stored. The PES header is removed when data is transferred from MB to EB.
- the video decoder 57 creates a frame / field image by decoding each video access unit of the video elementary stream at a predetermined decoding time (DTS). Since there are MPEG2, MPEG4AVC, VC1 and the like in the compression encoding format of the video stream multiplexed on the AV clip, the decoding method of the video decoder 57 is switched according to the stream attribute. In decoding the picture data constituting the base-view video stream, the video decoder 57 performs motion compensation by using the picture data existing in the future direction or the past direction as a reference picture.
- DTS decoding time
- the video decoder 57 when decoding individual picture data constituting the dependent-view video stream, the video decoder 57 performs motion compensation by using the picture data constituting the base-view video stream as a reference picture.
- the video decoder 57 transfers the decoded frame / field image to the DPB 59 and transfers the corresponding frame / field image to the picture switch at the display time (PTS) timing.
- the buffer switch 58 uses the decode switch information acquired when the video decoder 57 decodes the video access unit to determine whether to pull out the next access unit from EB53 or EB56, and stores it in EB53 and EB56.
- the assigned picture is transferred to the video decoder 57 at the decoding time (DTS) timing assigned to the video access unit. Since the DTS of the left-view video stream and the right-view video stream is set to alternate in units of pictures on the time axis, for example, when decoding ahead of schedule while ignoring DTS, the video access unit is converted to video in units of pictures. Transfer to the decoder 57 is desirable.
- Decoded PIcture Buffer (DPB) 59 is a buffer that temporarily stores the decoded frame / field image.
- the video decoder 57 decodes a video access unit such as a P picture or a B picture subjected to inter-picture prediction encoding, it is used to refer to the already decoded picture.
- the picture switch 60 switches the writing destination to the left view video plane and the right view video plane.
- uncompressed picture data is instantly written to the left-view video plane
- uncompressed picture data is instantly written to the right-view video plane.
- FIG. 27 is a diagram illustrating an internal configuration of the plane synthesis unit.
- the plane memory includes left-view video plane, right-view video plane, secondary video plane, IG plane, PG plane, and GFX plane. These are left-view video plane, right-view video plane, secondary video plane, IG plane, They are arranged in the order of PG plane and GFX plane.
- Video data is written from the system target decoder 4 to the left-view video plane and the right-view video plane at the PTS timing.
- the plane synthesizing unit 5 selects a plane in which video data is written out at the PTS timing from the left-view video plane and the right-view video plane, and performs a superimposing process on the secondary video plane, the PG plane, and the IG plane. Forwarded to
- Each of these plane memories can realize stereoscopic viewing by storing different contents in the left view and the right view. However, even if the stored contents in the left view and the stored contents in the right view are the same, if the coordinates of the pixels in the plane memory are changed between the left view and the right view, a pseudo stereoscopic view is realized. can do.
- the PG plane realizes stereoscopic vision by changing the coordinates of the pixels in the plane memory.
- a method for realizing stereoscopic viewing in the PG plane will be described.
- FIG. 28 is a diagram illustrating a method of synthesizing the PG plane.
- the method of plane synthesis will be described using the example of the PG plane in FIG.
- the plane synthesizing unit 5 acquires an offset value corresponding to the current display time from among the offset entries existing in the 3D metadata, corresponding to the PID of the presentation graphics currently being reproduced. Then, when the video plane to be superimposed is the left-view video plane, the plane synthesis unit 5 shifts the coordinates of the image data stored in the PG plane by the + offset value in the positive direction of the X axis. Then, after cropping the PG plane so that it does not protrude from the left-view video plane, it is used for composition with other planes (see the upper part of FIG. 28).
- the plane synthesis unit 5 shifts the negative direction of the X axis by the offset value of the PG plane, and after cropping the PG plane so that it does not protrude from the left-view video plane (Refer to the lower part of FIG. 28). The same processing is performed on the IG plane and the secondary video plane.
- FIG. 29 is a diagram schematically showing how the image is displayed to the user after being cropped and superimposed using the offset value. If clipping is performed by shifting the plane using the offset value, it is possible to create parallax images for the left eye and the right eye, so that it is possible to add depth to a planar image. If such a depth exists, the user can express the planar image as if it emerged from the screen of the display device.
- FIG. 30 is a diagram showing an internal configuration of the register set 10 and an internal configuration of the playback control engine 7b.
- the left side of the figure shows the internal configuration of the register set 10.
- the right side shows the internal structure of the regeneration control engine 7b.
- the stored values stored in the PSR are appropriately referred to by the movie object and the BD-J application, and are updated by the movie object and the BD-J application.
- the stored value of the PSR is a parameter referred to by the movie object and the BD-J application, and is also called a system parameter.
- PSR1 is a stream number register for audio streams and stores the current audio stream number.
- PSR2 is a stream number register for the PG stream and stores the current PG stream number.
- PSR4 is set to a value between 1 and 100 to indicate the current title number.
- ⁇ PSR5 indicates the current chapter number when set to a value between 1 and 999, and indicates that the chapter number is invalid in the playback device when set to 0xFFFF.
- PSR6 is set to a value between 0 and 999 to indicate the current playlist number.
- ⁇ PSR7 is set to a value between 0 and 255 to indicate the current play item number.
- ⁇ PSR8 is set to a value between 0 and OxFFFFFF to indicate the current playback point (current PTM) with a time accuracy of 45 KHz.
- current PTM current playback point
- PSR10 is a stream number register for the IG stream, and stores the current IG stream number.
- PSR21 indicates whether or not the user intends to execute stereoscopic playback.
- PSR23 is a setting of “Display Capability for 3D”. This indicates whether or not the display device, which is a connection partner of the playback device, has the ability to execute stereoscopic playback.
- PSR24 is a setting of “Player Capability for 3D”. This indicates whether or not the playback device has the ability to perform stereoscopic playback.
- the PSR4, PSR6, PSR21, PSR23, and PSR24 in the register set 10 and the stream selection table of the current playlist information in the management information memory 9 are referred to and output in the current playlist.
- a procedure execution unit 8 for uniquely determining a mode is provided.
- Player Capability for 3D in PSR24 means the overall capability of 3D playback of the playback device, and may be simply expressed as “3D-Capability”.
- PSR23 defines the output mode, and the state transition selection model is defined as shown in FIG.
- FIG. 31 is a diagram showing the state transition of the output mode selection model. There are two general states in this selection model. The ellipse schematically depicts this general state, that is, “Invalid” and “valid”, which are possible values of the output mode value. Invalid indicates that the output mode is valid, and Valid indicates that the output mode is invalid.
- State transition includes start of playlist playback, change of output mode requested by navigation command or BD-J application, jump to BD-J title.
- a state transition occurs, a procedure is executed to obtain a preferred output mode.
- arrows jm1, jm2, jm3,..., Jm12 schematically indicate events that trigger state transitions.
- the state transitions in this figure include the following.
- “Load a disc” means that the BD-ROM is loaded.
- Start Presentation means start of playlist playback (start Playlist playback) in HDMV mode. In BD-J mode, this means branching to a BD-J title. This is because, in the BD-J mode, playback of a playlist is not always started when branching to a BD-J title.
- “Jump to BD-J title” means branching to the BD-J title. Specifically, in the index table, the title (BD-J title) associated with the BD-J application becomes the current title.
- Start Playlist Playback means that a playlist number meaning any playlist is set in the PSR, and the playlist information is read to the memory as current playlist information.
- “Terminate presentation” means that playback of the playlist ends in HDMV mode, and in BD-J mode, the title (HDMV) associated with the movie object in the index table from the BD-J title. Jump to (title).
- the output mode state transitions to the temporary state “Initialization”.
- the state of the output mode selection temporarily transits to “Initialization state” and then transitions to invalid state.
- the output mode transits to Valid according to the result of Procedure when playback condition is changed. If the output mode is valid and Start Presentation ends, the state transitions to Invalid.
- the navigation command in the movie object must be executed prior to the start of playlist playback in order for the content provider to set the preferred output mode.
- a navigation command is executed on a movie object, it becomes Invalid in this model.
- FIG. 32 is a flowchart showing the initialization procedure.
- Step S1 is a determination as to whether or not the disc unbound BD-J application is operating.
- Step S2 indicates that Stereoscopic Display Capability in PSR23 indicates “Capability exists”, and Initial_output_mode information in Index.bdmv indicates “stereoscopic output”. Whether or not “mode” is indicated.
- Step S1 If Step S1 is Yes, the current output mode is maintained in Step S3. If Step S1 is No and Step S2 is Yes, PSR22 is set to the stereoscopic output mode in Step S4. If step S1 is No and step S2 is No, the output mode in PSR22 is set to 2D output mode in step S5.
- FIG. 33 is a flowchart showing the processing procedure of Procedure when playback condition is changed.
- Step S11 is a determination as to whether or not the output mode in PSR22 is the 2D output mode.
- Step S13 indicates that Stereoscopic Display Capability in PSR23 indicates “capability present”, and STN_table_SS exists in the playlist. It is a judgment of whether or not.
- Step S11 If Step S11 is Yes, the current output mode is not changed in Step S12. Even if Step S11 is No and Step S13 is Yes, the current output mode is not changed (Step S12). If step S11 is No and step S13 is Yes, the current output mode is changed to the 2D output mode (step S14).
- the PES stream that can be played back by each play item is defined in the stream selection table of each play item. Therefore, in order to start playback of the current play item, first, it is necessary to select a PES stream that is permitted to be played back in the stream selection table of the current play item and that is optimal for playback of the play item. This selection procedure is called a “stream selection procedure”.
- FIG. 34 is a flowchart showing the processing procedure of the stream selection procedure.
- Step S21 is a determination step for determining whether or not the display method of the playback device is 2D. If the determination result is Yes, the 2D STN_table in the current playitem information is set to the current STN_table (step S22). ). If the determination result is No, among the STN_table_SS existing in the extension data of the playlist information, the one corresponding to the current play item information is set as the current STN_table. Thereafter, the processing from step S24 to step S33 is executed. In steps S24 to S33, the processes in steps S26 to S33 are repeated for each of the primary video stream, the IG stream, the secondary video stream, the primary audio stream, and the secondary audio stream.
- Step S26 determines whether or not the number of STN_table entries corresponding to stream x in the current STN_table is 0.
- Step S27 stores the number of stream entries corresponding to stream x in the current stream in the stream number register. This is a determination step for determining whether the stream number is equal to or greater than the specified stream number.
- step S26 and step S27 are Yes, the stream number stored in the stream number register is maintained in step S33.
- step S28 it is determined which of the plurality of conditions the PES stream registered in the current STN_table satisfies (step S28), and the condition determined to satisfy It is determined whether there are a plurality of PES streams having the same combination (step S29).
- step S30 If there is only one PES stream that satisfies the condition, one PES stream that satisfies the condition is selected (step S30).
- a PES stream determined to satisfy the same conditions is selected with the highest priority in the current STN_table (step S31). If the PES stream is selected in this way, the stream number corresponding to the stream entry of the selected PES stream is written to the stream number register in the PSR (step S32).
- the current play item playback processing procedure is Procedure when It depends on the output mode determined by playback condition is changed. The play item playback procedure according to the output mode will be described with reference to FIG.
- FIG. 35 is a flowchart showing the play item playback procedure.
- Step S41 is for determining whether or not the current output mode is the 3D output mode. If the current output mode is the 2D output mode, the current play item number is initialized to “1” in step S42. The process proceeds to a loop of step S43 to step S48.
- steps S43 to S46 repeats the process of executing steps S43 to S46 for the current play item and incrementing the current play item number (step S48) until the current play item number becomes final. (Yes in step S47).
- the contents of steps S43 to S46 are as follows.
- step S43 the transport stream file designated by “XXXXX” described in Clip_Information_file_name of the current play item and the extension “m2ts” is opened, and in step S44, the packet ID corresponding to the video stream packet ID is opened.
- the current PlayItem.In_Time and the current PlayItem.Out_Time are converted into Start_SPN [i] and End_SPN [i].
- step S45 the extent belonging to the read range [i] for reading the TS packet [i] of the packet ID [i] from Start_SPN [i] to End_SPN [i] is specified, and in step S46, the read range [i] Instructs the BD-ROM drive to continuously read the extents belonging to.
- step S49 If the current output mode is the stereoscopic output mode, the current play item number is initialized to “1” in step S49, and then the process proceeds to a loop of step S50 to step S60.
- steps S50 to S58 repeats the process of executing steps S50 to S58 for the current play item and incrementing the current play item number (step S60) until the current play item number is final. (Yes in step S59).
- the contents of steps S50 to S58 are as follows.
- step S50 the transport stream file specified by “XXXXX” described in Clip_Information_file_name of the current play item and the extension “ssif” is opened.
- step S51 the left-view video stream and the right-view video are opened.
- the stream designated by the left view / right view identification information of the current play item information is set as the base view video stream. Others are defined as dependent view streams.
- step S52 the current PlayItem.In_Time and the current PlayItem.Out_Time are converted into Start_SPN [i] and End_SPN [i] using the entry map corresponding to the packet ID of the base-view video stream.
- step S53 the SubPlayItem corresponding to the dependent view stream is specified, and the SubPlayItemIn_Time and SubPlayItemOut_Time specified using the entry map [j] corresponding to the packet ID [j] of the dependent view stream are set to Start_SPN [j]. , End_SPN [j] (step S54).
- step S55 The extent belonging to the read range [i] for reading the TS packet [i] with the packet ID [i] from Start_SPN [i] to End_SPN [i] is specified (step S55), and the TS packet [i] with the packet ID [j]
- step S56 The extent belonging to the read range for reading j] from Start_SPN [j] to End_SPN [j] is specified (step S56).
- step S57 the extents belonging to the reading ranges [i] and [j] are sorted in ascending order of the addresses, and the extents belonging to the reading ranges [i] and [j] are continuously used by using the addresses sorted in step S58. Instruct the drive to read automatically.
- the middleware outputs an event to the BD-J application to prompt screen rendering for stereoscopic viewing.
- FIG. 36 is a diagram showing what events are output to the BD-J application when the state of the playback control engine is switched from “stopped” to “3D playlist playback”.
- the first row shows the GUI drawn by the BD-J application.
- the third level shows the state of the regeneration control engine.
- the second level shows an HScreenConfiguration event output from the middleware to the BD-J application.
- the state of the playback control engine has changed as “stopping ⁇ 3D playlist playback ⁇ stopping”.
- the HScreenConfiguration event indicating the start of 3D is output at the timing of transition from stop to 3D playlist playback, and the transition from 3D playlist playback to stop, and at the timing of transition from stop to 3D playlist playback. It can also be seen that the HScreenConfiguration event indicating the end of 3D is output.
- the GUI drawn by the BD-J application while the playback control engine is stopped is a 2D GUI.
- the GUI drawn by the BD-J application during 3D playlist playback is a 3D GUI. This is because the BD-J application performs GUI drawing switching in response to the event output.
- FIG. 37 is a diagram illustrating what events are output to the BD-J application when the state of the playback control engine is switched from 2D playlist playback to 3D playlist playback.
- the first row shows the GUI drawn by the BD-J application.
- the third level shows the state of the regeneration control engine.
- the second level shows an HScreenConfiguration event output from the middleware to the BD-J application.
- the state of the playback control engine transitions from 2D playlist playback ⁇ 3D playlist playback ⁇ 2D playlist playback.
- the HScreenConfiguration event indicating 3D start is output at the timing of transition from 2D playlist playback to 3D playlist playback
- the HScreenConfiguration event indicating 3D termination is output.
- the GUI drawn by the BD-J application during playback of the 2D playlist by the playback control engine is a 2D GUI.
- the GUI drawn by the BD-J application during 3D playlist playback is a 3D GUI. This is also because the BD-J application performs GUI drawing switching in response to the event output.
- FIG. 38 is a diagram showing what events are output to the BD-J application when the target stream is switched during playback of the 3D playlist by the playback control engine.
- the first row shows the GUI drawn by the BD-J application.
- the third level shows the state of the regeneration control engine.
- the second level shows an HScreenConfiguration event output from the middleware to the BD-J application.
- the state of the playback control engine is 3D playlist playback or that the stream is switched in the middle.
- the HScreenConfiguration event is output at the timing when the stream is changed to another and when the stream is returned to the original one.
- the GUI drawn by the BD-J application during playback of the 2D playlist by the playback control engine is a 2D GUI.
- the GUI drawn by the BD-J application during 3D playlist playback is a 3D GUI.
- the output mode of the playback device is the stereoscopic playback mode
- the Clip_Information_file_name in the playlist information and the extension “” indicating that the transport stream file is in the interleave format are used.
- the extent of the interleaved transport stream file identified from ssif is read out and used for playback. The extent can be read.
- the extent of the interleaved transport stream file is not read by the 2D playback device, so the unique ATS change of the interleaved transport stream, that is, the ATS increases or decreases, and the ATS increases. Repeated irregular changes such as decreasing will not cause malfunction or instability of the 2D playback device.
- playlist information in which certain file reference information is described is created, at the time of 3D playback, the file name of the file reference information, an extension indicating that it is an interleaved transport stream, An interleaved stream file with a file name will be read and played, and in 2D playback, it has a file name of the file reference information and an extension indicating that it is a normal format transport stream file The transport stream file is read and played.
- FIG. 39A is a diagram showing an internal configuration of the display device 300.
- the display device 300 includes a tuner 71, an HDMI transmission / reception unit 72, a message storage unit 73, a display control unit 74, a display element 75, and a wireless transmission unit 76.
- the tuner 71 receives and demodulates a multi-channel transport stream transmitted by terrestrial digital broadcasting or satellite digital broadcasting. At this time, the tuner 71 can simultaneously select a plurality of channels and output an uncompressed picture.
- the HDMI transmitting / receiving unit 72 receives uncompressed and synthesized picture data transmitted from the playback device via HDMI.
- the message storage unit 73 stores a warning message to be displayed instead of a picture.
- the display control unit 74 provides an uncompressed picture obtained by demodulation by the tuner 71 and an uncompressed synthesized picture transmitted from the playback device via HDMI.
- the display control unit 74 can switch the display period with a time accuracy of 1/120 second and 1/140 second. With this accuracy, for example, the display period of 1/24 second is changed to 1/48 second, 1 It can be subdivided into smaller display periods such as / 72 seconds and 1/92 seconds.
- the display panel 75 is a device that emits light by driving a liquid crystal display element, a plasma light emitting element, and an organic EL element, and displays uncompressed picture data under the control of the display control unit 74. .
- the wireless transmission unit 76 controls the 3D glasses 400 by an infrared communication method or a wireless LAN method. Specifically, in each of the 3D mode and the multi-channel mode, a synchronization signal that prompts the state transition of the 3D glasses 400 is transmitted at the beginning of the display period. By such signal transmission, the control of changing to the light transmitting state in the left view period, changing to the light blocking state in the right view period, and changing to the light transmitting state again in the left view period is repeated. Thereby, changes as shown in FIGS. 1B and 1C are performed.
- the display control unit 74 converts the display period of 1/24 seconds into three display periods (display periods 1/3, 2/3, 1/72 seconds). Divided into 3/3), different display contents are displayed in each of the three display periods 1/3, 2/3, 3/3.
- the display content of the first display period 1/3 is a left view
- the display content of the second display period 2/3 is a right view
- the content of the third display period 3/3 is a warning screen To.
- a synchronization signal is transmitted to the glasses, and the state transition of the left view and the right view is performed.
- the display device 300 demodulates a plurality of channels in a time division manner.
- the display control unit 74 divides the display period of 1/24 seconds into two display periods (display periods 1/2, 2/2) having a time length of 1/24 seconds, and displays the two display periods. Separate display contents are displayed on 1/2 and 2/2, respectively.
- the display content of the first display period 1/2 is channel 1
- the display content of the second display period 1/2 is channel 2, and so on.
- FIG. 39B is a diagram showing an internal configuration of the 3D glasses 400.
- a wireless reception unit 81 that receives a synchronization signal that triggers a state transition from the display device 300, a state control unit 82 that transitions the state of the liquid crystal shutter to a light transmitting state and a light shielding state, and liquid crystal shutters 83 and 84 are provided. To do.
- glasses In 3D mode, glasses have a light blocking state and a light blocking state, as well as a light blocking state and a light blocking state.
- the blocking-light blocking state is a state in which both the left view and the right view are closed.
- the glasses change between a light-transmitting state in which both the left view and the right view are open and a light blocking-light-blocking state in which both the left view and the right view are closed.
- a warning screen that prompts the user to wear the 3D glasses 400 is shown to the user who has already worn the 3D glasses. There is no need for consideration. How to control the 3D glasses 400 so that the warning screen is not shown to the user who has been worn will be described with reference to FIG.
- FIG. 40 shows display contents in the 3D mode, the left view state of the glasses, and the right view state.
- the first level shows the display period on the playback time axis
- the second level shows the display content of the display device.
- the third row shows the state of the right view and left view of the glasses.
- the left view image is displayed on the display device, the left view of the glasses is in a translucent state, and the right view is in a blocked state. Yes.
- the right view image is displayed on the display device, the left view of the glasses is in a light-shielding state, and the right view is in a translucent state.
- the display device displays a warning screen that prompts the user to wear glasses, and it can be seen that the left view and right view of the glasses are in a light-shielded state.
- the user wearing the glasses will receive a warning displayed on the screen. I can't see the video.
- people who don't wear 3D glasses can see a message prompting them to wear 3D glasses, but people who already wear 3D glasses can't see it. A corresponding message will be displayed.
- the display device When executing multi-channel display, the display device realizes a unique control of switching the shutters of two glasses instead of switching the left and right shutters. This unique control will be described with reference to FIG.
- FIG. 41 shows the display contents in the 3D mode, the left view state, and the right view state when the shutters of the two glasses are switched instead of switching the left and right shutters.
- the first level shows the display period on the playback time axis, and the second level shows the display content of the display device.
- the third row shows the R and L states of the glasses.
- the glasses worn by the user 1 are in a translucent-translucent state, and the user 1 can view the channel 1 (Ch1).
- the glasses worn by user 2 are in a light-shielded state and user 2 cannot view channel 1 (Ch1).
- the glasses worn by the user 1 are in a light-shielded state, and the user 2 cannot view the channel 2 (Ch2).
- the glasses worn by the user 2 are in a translucent-translucent state, and the user 2 can view the channel 2 (Ch2). If you use it like this, you already have 3D glasses that allow two people to see different channels on one screen at the same time. It is possible to make it independent, and the range of applications is widened, such as avoiding channel rights in the living room and playing a battle game on one screen. Further, by increasing the steps, it becomes possible to view 3 channels or more channels on one screen.
- each user wears the 3D glasses 400, so that each user views a channel that he / she likes. It becomes possible. Since the user can watch a favorite program without preparing display devices for the number of users, the user's living room can be used effectively.
- the present embodiment is an embodiment relating to negotiation between the playback device and the display device.
- the home theater system construction circumstances in the user's home are unique to each home, and when connected to a display device, the playback device negotiates with the display device that is the connection partner to play back which playlist. It is necessary to switch between serving.
- 3D video cannot be output because the conventional analog method does not support 3D video.
- the analog output will display a message such as “Playing 3D video. Please view on a 3D-compatible display” and the user may connect to the wrong terminal or not. Display a message to let you know that you are connected to the display device.
- the connected display device is a 2D display device, it is desirable to automatically switch playback to the 2D playlist.
- 2D output priority is given to information that specifies whether left-view video stream or right-view video stream is output for analog when simultaneously outputting the video signal that is the playlist playback result to multiple display devices. It is called video information. This information is provided in advance in the playlist, and video signals are simultaneously output to the 2D display device and the 3D display device according to the 2D output priority video information in the current playlist. By doing this, it is possible to simultaneously decode 2D and 3D video images and output them to two display devices at the same time without having to handle 2D and 3D playlists independently.
- the 3D display device displays 3D-compatible OSD display, but only 2D video or left view is available for output that supports only 2D, such as analog output. Output only light view.
- FIG. 42 is a diagram illustrating a connection mode between the playback device and the display device.
- the upper side shows an analog-connected display device.
- the 3D-compatible display device 300 that is digitally connected is shown on the lower left side, and the 2D display device 302 that is digitally connected is shown on the lower right side.
- a playback device When a playback device is connected to a 3D display device and a 2D display device at the same time, it tries to negotiate with these display devices. If it is found that the connection with the display device on the other side is an analog connection and it cannot be negotiated, the left-view video stream and the left-view video stream are displayed in the 2D output priority video information. Play things. In this way, when an analog connection is established with a display device as a connection partner, a playlist intended by the authoring person can be provided for reproduction.
- connection with both display devices is a digital connection and the negotiation is successful, it is determined whether the other party is a 3D display device or a 2D display device.
- the 2D display device is caused to transition the screen content as indicated by an arrow mg1 in the figure.
- An arrow mg1 indicates a transition of the screen content of the 2D display device.
- 3D video is being played.
- display 2D video After displaying the message “Please view with 3D display device”, display 2D video.
- the reason for switching the plurality of light views is the difference in the screen size of the display device.
- the distance between the left eye and the right eye is almost the same regardless of individual differences, but some display devices are 20 inches and others are 150 inches. Assuming a 50-inch display device, an image created with an eye-to-eye distance of 6.5 cm is 3 times as large as a 19.5-cm eye distance with a 150-inch display device. It becomes difficult to recognize as.
- left view and right view combination is set to 6.5 cm, the difference between the left view and right view on various sizes of display devices, it is best to select a video stream according to the display device.
- a combination of left-view video stream and right-view video stream can be selected.
- FIG. 43 shows the relationship between the number of pixels indicating the left and right difference and the distance on the screen of the display device.
- the left side shows a set of a right view picture and a left view picture with different left and right offsets.
- the middle row shows the distance in a 50-inch display device
- the right side shows the distance in a 150-inch display device.
- the left and right difference is 50 pixels
- the distance on the 50-inch display device is 2.0 cm
- the distance on the 150-inch display device is 6.0 cm.
- the distance on the 50-inch display device is 4.0 cm, and the distance on the 150-inch display device is 12.0 cm.
- the distance on the 50-inch display device is 6.0 cm, and the distance on the 150-inch display device is 18.0 cm.
- ⁇ 6.0cm is optimal for 50-inch display devices, and 6.0cm is optimal for 150-inch display devices, so the distance displayed on the screen is adjusted using the 3DStream Depth ChangeUO and 3DStreamDepthChange commands.
- the program can automatically select the optimal left eye / light view combination using the method for acquiring the screen size of the display device described in FIG. 43, the user can select the optimal stream without worrying about the screen size. It will be automatically selected.
- streams with different pixel differences in local storage are recorded on the recording medium, and UO and commands are used to switch between these streams. By using it, the user himself / herself can select a feeling of depth / depth.
- the present embodiment when the playback device is connected to the display device, it is ensured that more appropriate playback output is performed in relation to the display device.
- the present embodiment realizes an improvement of what PG stream and IG stream should be selected together with a video stream when performing stereoscopic viewing.
- the video played on the 2D playback device is 2D video, and the corresponding subtitles and menu images are 3D.
- 3D video is played back by the 3D playback device, and the corresponding subtitles and menu images are also 3D. This is because when 2D PG or IG is displayed for 3D video, the context is unintentional, and the user cannot recognize the space normally and harms the worst health.
- the user is free to select 2D video, but in that case, the corresponding subtitles and menu images should also be switched to 2D automatically.
- Combinations of subtitles related to 2D video and subtitles related to 3D video may be selected by the program, but the playback device automatically eliminates inappropriate combinations by associating them as data It becomes possible.
- the mechanism is described.
- the 3D compatible playlist is managed by further dividing the stream selection table into STD_table_SS for 2D and STN_table_SS for 3D playback.
- the video / audio / PG / IG stream registration used only for 2D playback and the video / audio / PG / IG stream registration used for 3D are registered in different entry groups.
- audio / PG / IG prepared for 3D cannot be selected.
- audio / PG / IG prepared for 2D cannot be selected.
- the management table can be further divided so that the left view and the right view, and the subtitle / menu image stream registration associated therewith can be managed separately. Since there is a difference in the presence / absence, position, and angle of depth between the PG stream created for 2D and the PG stream created for 3D, the 2DPG stream is selected when playing a video stream for stereoscopic viewing. Thus, playback with a 3D video stream is something that authors want to avoid.
- FIG. 44 is a description example of stream combination information when a video stream and a PG stream are combined.
- FIG. 44 (b) schematically shows an allowable combination between the video stream and the PG stream specified by the combination information.
- the video stream on the left side of the figure indicates the video stream specified by the stream numbers of streams 1, 2, 3, and 4.
- the video stream specified by the stream numbers of streams 1 and 2 is for 2D
- the video stream specified by the stream numbers of streams 3 and 4 is for 3D.
- the PG stream on the left side of the figure indicates the PG stream specified by the stream numbers of streams 1, 2, 3, and 4. Of these, it can be seen that the PG stream specified by the stream numbers of streams 1 and 2 is for 2D, and the PG stream specified by the stream numbers of streams 3 and 4 is for 3D.
- the combination of the PG stream and the video stream is defined in advance in the stream combination information, if a PG stream is selected according to the stream combination information, when a certain video stream is selected, It is guaranteed that an optimal PG stream is selected.
- FIG. 45 is a flowchart showing the stream selection processing procedure of the playback device according to the stream combination information.
- the stream selection process shown in this figure is executed, and the video stream and PG stream are converted into stream combination information. Match the registered combination.
- step S71 a video stream number is acquired, and in step S72, a PG stream is acquired.
- Step S73 is a determination as to whether or not a combination of a video stream and a PG stream is registered in the stream combination information. If registered, playback is performed in accordance with the stream combination information in step S74. If not registered, another PG stream in which the combination with the video stream is registered in the stream combination information in step S75 is selected and reproduced.
- stereoscopic vision has various principles, and products that are being distributed in the market are considered to be compatible with various 3D systems. There are various 3D playback methods, and the support / non-support on the display device side varies depending on the method.
- the system parameters of the playback device are preferably in a format that can express a plurality of methods.
- examples of 3D playback methods include two-screen stereo playback method that sends two-screen video independently, side-by-side method, horizontal double method, 2D + depth information method, etc. If there is a method that can be supported by the display device, the bit assignment of the PSR is determined so as to identify whether or not the corresponding method can be executed.
- FIG. 46 is a diagram showing PSR bit assignments that can cover a plurality of 3D systems.
- the PSR 24 in this figure consists of 4 bits (b3, b2, b1, b0), and the corresponding 3D playback method is associated with each bit from the most significant bit b3 to the least significant bit b0. If the playback device supports the 3D playback method, the corresponding bit is set to “1”, and if not, the corresponding bit is set to “0”. When all the bits of PSR24 are “0”, the playback device is a 2D playback device, and when any or some of the bits are “1”, it indicates that it is a 2D / 3D playback device of the corresponding system. .
- bits from the most significant bit b3 to the least significant bit b0 in PSR24 can support 3D display methods such as 2-screen stereo playback, side-by-side method, horizontal double method, and 2D-depth information method Indicates.
- the two-screen stereo playback method is the method described in the above embodiments.
- the side-by-side method is a method in which the resolution of 1920 ⁇ 1080 is divided into 960 ⁇ 1080 and 960 ⁇ 1080, and left view and right view are displayed on each of these.
- the resolution of 1920 ⁇ 1080 is changed to the resolution of 3840 ⁇ 1080, among which the left view is assigned to the 1920 ⁇ 1080 resolution portion and the right view is assigned to the 1920 ⁇ 1080 resolution portion. It is a method.
- the 2D-depth information method is a method for realizing stereoscopic viewing with 2D video and grayscale images.
- the gray scale image is composed of binarized pixels.
- the luminance of the binarized pixel indicates the depth of each pixel in the 2D video. According to the luminance of the binarized pixels, a depth of each pixel in the 2D video is created to construct a stereoscopic image.
- the display device and the playback device are connected by a transmission method that can transmit the performance / support method of the display device to the playback device, such as HDMI, not only the performance of the playback device, but also the support method of the display device, Automatically set to PSR24.
- the value of PSR24 varies depending on the display device connected even with the same playback device.
- the compatible method can be acquired from the display device, in addition to the simple compatible method, acquire information that affects 3D playback, such as the size and resolution of the display device, the distance from the screen of the display device to the viewer, and save it to the PSR24. By storing it, it is possible to use it for selecting an optimum reproduction method by a program described later.
- ⁇ 3D support status cannot be expressed with 1 bit. In this case, multiple bits should be used. If the video size is up to 1920x1080, but it is assumed that it cannot be played back due to lack of decoder performance, etc., if it has a resolution higher than that, use 2 bits to indicate "00b”, 1920x1080 If the correspondence is expressed as “01b”, and the correspondence corresponding to a larger size as “10b” or the like, it is possible to express the correspondence situation in more detail using system parameters.
- FIG. 47 is a diagram in which the 3D playback method supported by the display device is reflected in the playback device setting register.
- the title is selected by the user, and the executed BD program file checks whether the playback device supports 3D video playback in the program and if the user selects 3D video playback. Switch the playlist to play.
- the playlist that constitutes the FirstPlay title should be composed of 2D video that is played on any playback device for safety.
- the program stored on the BD-ROM is created by the authoring side, and if the playback device supports multiple 3D formats, it is up to the authoring side to decide which 3D playback method to preferentially select. Dependent.
- 3D format 1 is a two-screen stereo playback format
- 3D format 2 is a side-by-side format
- the playback device supports only the side-by-side format
- the program can be played on that playback device.
- the side-by-side 3D playlist 005.MPLS is selected and played.
- the authoring person operates the program on the BD-ROM to set the optimal 3D system for the playback device and the display device. Can be set to register.
- FIG. 48 shows the relationship between the index file (Index.bdmv) and the program file.
- the left side of the figure shows an index table and a mode management module 16 that is a decoding subject of the index table.
- the index table includes entries corresponding to the FirstPlay title, the top menu, and the titles 1, 2, and 3, respectively.
- the right side of the figure shows four playlist files that should be selectively played according to the output mode setting in the playback device.
- the four playlist files contain 00001.mpls and 00003.mpls that describe the path for playing 2D video, 0004.mpls that describes the playback path for 3D system 1, and the playback path for 3D system 2 0005.mpls existed.
- Movie object 1 commands playback of 00001.mpls. This 00001.mpls defines a 2D playlist. This is because the playlist to be played back with the FirstPlay title must be played back in any output mode.
- 3D-Capability shown in PSR24 is 3D format 1
- movie object 2 orders playback of 00004.mpls
- 3D-Capability is 3D format 2
- it orders playback of 00005.mpls. If it doesn't match any 3D format, order playback of 00003.mpls.
- Arrows pp1, pp2, and pp3 in the figure schematically show play list reproduction instructions by movie objects.
- the arrows my1 and my2 in the figure indicate that these movie objects are subjected to decryption by the HDMV module 13. It can be seen that the HDMV module 13 executes these movie objects so that the three playlist files as described above are selectively used for playback according to the Capability in the playback device.
- FIG. 49 is a flowchart showing a stream selection processing procedure.
- step S81 the corresponding 3D system of the playback device is acquired, and in step S82, the stream selection table is acquired.
- step S83 it is determined whether or not the 3D system supported by the playback apparatus matches the stream to be selected. If the determination result in step S83 is Yes, selection is permitted in step S84. If the determination result of step S83 is No, selection is not permitted in step S85.
- a recording apparatus that executes the recording method creates an AV clip in real time and records it on a BD-RE or BD-R, a hard disk, or a semiconductor memory card.
- the AV clip may be a transport stream obtained by real-time encoding of an analog input signal by the recording device, or a transport stream obtained by partializing the transport stream digitally input by the recording device. It may be a stream.
- a recording apparatus that performs real-time recording includes a video encoder that encodes a video signal to obtain a video stream, an audio encoder that encodes an audio signal to obtain an audio stream, a video stream, an audio stream, and the like.
- a multiplexer that obtains the TS format digital stream and a source packetizer that converts the TS packets that make up the MPEG2-TS format digital stream into source packets.
- the MPEG2 digital stream converted to the source packet format is converted into an AV clip file.
- the control unit of the recording apparatus performs processing for generating clip information and playlist information on the memory. Specifically, when a recording process is requested by the user, the control unit creates an AV clip file and a clip information file on BD-RE and BD-R.
- the control unit of the recording apparatus starts the head in this GOP.
- a set of EP_PTS entries and EP_SPN entries is added to the entry map of the clip information file.
- the is_angle_change flag set to “ON” is added to the set of the EP_PTS entry and the EP_SPN entry. If the head of the GOP is not an IDR picture, the is_angle_change flag set to “off” is added to the set of EP_PTS entry and EP_SPN entry.
- the stream attribute information in the clip information file is set according to the stream attribute to be recorded. If AV clips and clip information are generated and written to BD-RE and BD-R as described above, playlist information that defines the playback path is generated via the entry map in the clip information, and BD -Write to RE and BD-R. By executing such processing in the real-time recording technique, a hierarchical structure of AV clip-clip information-playlist information can be obtained on BD-RE and BD-R.
- the above is a recording apparatus that executes a recording method by real-time recording. Next, a recording apparatus that executes a recording method using preformat recording will be described.
- the recording apparatus described here is installed in a production studio for distributing movie content and is used by authoring staff. In accordance with operations from the authoring staff, generate a digital stream that is compression-coded according to the MPEG standard and a scenario that describes how to play a movie title, and generate a volume bitstream for BD-ROM that contains these data
- FIG. 50 is a diagram illustrating an internal configuration of the recording apparatus. As shown in the figure, the recording apparatus according to the present invention includes a video encoder 501, a material production unit 502, a scenario generation unit 503, a BD program production unit 504, a multiplexing processing unit 505, and a format processing unit 506.
- the video encoder 501 encodes an image image such as an uncompressed bitmap of a left view and an image image such as an uncompressed bitmap of a right view according to a compression method such as MPEG4-AVC or MPEG2, Create a right-view video stream.
- the right-view video stream is encoded by inter-frame predictive encoding corresponding to the display time of the left-view video stream.
- depth information at the time of 3D video is extracted from the motion vectors of the left view image and the right view image, and written into the frame depth information storage unit 501a.
- the video encoder 501 performs image compression by extracting motion vectors in units of 8 ⁇ 8 or 16 ⁇ 16 macroblocks in order to perform image compression using correlation characteristics between pictures.
- a moving image in which a house is present in the background and a person is present in the foreground is set as a motion vector extraction target.
- inter-picture prediction is performed between the left eye image and the right eye image.
- a motion vector is not detected at the location of the “house” image, but a motion vector is detected at the location of the “person”.
- ⁇ Extract detected motion vectors and create depth information for each frame when displaying 3D video.
- the depth information for example, it is conceivable to adopt the same image as the resolution of a frame having a depth of 8 bits.
- the material production unit 502 creates each stream such as an audio stream, a presentation graphics stream, and an interactive graphics stream, and writes the streams to the audio stream storage unit 502a, the presentation graphics stream storage unit 502b, and the interactive graphics stream storage unit 502c.
- the material production unit 502 creates an audio stream by encoding uncompressed LinearPCM audio or the like according to a compression method such as AC3.
- the material production unit 502 uses a subtitle information file including subtitle images, display timing, and subtitle effects such as fade-in / fade-out, as a presentation graphic that is a PG stream format compliant with the BD-ROM standard. Create a stream.
- the material production unit 502 uses interactive graphics, which is a menu screen format compliant with the BD-ROM standard. Create a stream.
- the scenario generation unit 503 generates a scenario in the BD-ROM format according to the information of each stream generated by the material production unit 502 and the operation from the authoring staff via the GUI.
- the scenario here refers to an index file, a movie object file, a playlist file, and the like.
- the scenario generation unit 503 creates a parameter file that describes which stream each AV clip for realizing the multiplexing process is composed of.
- Files such as index files, movie object files, and playlist files created here have the data structures described in the first and second embodiments.
- the BD program creation unit 504 creates a BD program file source code according to a request from the user through a user interface such as a GUI, and creates a BD program. At this time, in order for the program of the BD program file to set the depth of the GFX plane, the depth information output from the video encoder 501 can be used.
- the multiplexing processing unit 505 multiplexes a plurality of streams such as a left-view video stream, a right-view video stream, video, audio, subtitles, and buttons described in the BD-ROM scenario data, and outputs the MPEG2-TS format AV. Create a clip. At this time, a clip information file paired with the AV clip is created at the same time.
- the multiplexing processing unit 505 creates a clip information file by pairing the entry map generated by itself with attribute information indicating an audio attribute, a video attribute, and the like for each stream included in the AV clip.
- the structure of the clip information file has the data structure described in the above embodiments.
- the format processing unit 506 includes the BD-ROM scenario data generated by the scenario generation unit 503, the BD program file generated by the BD program generation unit 504, the AV clip and clip information file generated by the multiplexing processing unit 505, and the BD- Files and directories are arranged in a format compliant with the ROM standard, and a disk image is created in the UDF format, which is a file system compliant with the BD-ROM standard.
- 3D metadata of PG, IG, and secondary video stream is created using the depth information output by the video encoder 501 at this time.
- the image layout on the screen is automatically set so that it does not overlap with the 3D image object, and the offset value is adjusted so that the depth does not overlap.
- the file layout of the disk image created in this way is set with the data structure of the file layout described in the first and second embodiments.
- a BD-ROM can be manufactured by converting the generated disk image into BD-ROM press data and performing a press process on the data.
- the recording device may write a digital stream by managed copy.
- Managed copy refers to a digital stream, playlist information, clip information, and application program recorded on a read-only recording medium such as a BD-ROM on another optical disc (BD-R, BD-RE, DVD-R, DVD -RW, DVD-RAM, etc.), hard disk, removable media (SD memory card, memory stick, compact flash (registered trademark), smart media, multimedia card, etc.)
- BD-R Blu-ray Disc
- BD-RE Blu-On-On-On-On-On-On-On-On-On-On-On-Only, etc.
- DVD-R Blu-ray Disc
- DVD-RW digital versatile disc
- DVD-RAM digital versatile disc
- hard disk a hard disk
- removable media SD memory card, memory stick, compact flash (registered trademark), smart media, multimedia card, etc.
- This is a technology that enables copying only in a state where authentication is performed and permitted. With this technology, it is possible to perform control such as limiting the number of backups or allowing backup only
- transcoding refers to the bit rate assigned to the video stream and audio stream by converting the format of the digital stream recorded on the BD-ROM from MPEG2 transport stream format to MPEG2 program stream format, etc. Is a process of adapting the digital stream to the application format of the copy destination medium by re-encoding with a lower value. In such transcoding, it is necessary to obtain AV clips, Clip information, and playlist information by performing the above-described real-time recording process.
- Steposcopic method The parallax image method premised on the description in the first embodiment displays 24 images per second for a normal two-dimensional movie, for example, in order to display left and right images alternately in the time axis direction. On the other hand, it is necessary to display 48 images per second including the left and right images. Therefore, this method is suitable for a display device in which rewriting of one screen is relatively quick. Stereoscopic viewing using this parallax image is already commonly used in amusement park playground equipment and the like, and since it has been established technically, it can be said that it is the closest to practical application at home. In addition to these, various techniques such as a two-color separation method have been proposed for stereoscopic viewing using parallax images. In the present embodiment, the sequential separation method or the polarized glasses method has been described as an example, but the present invention is not limited to these two methods as long as a parallax image is used.
- the display device 300 not only a lenticular lens but also a device having a similar function, for example, a liquid crystal element may be used.
- the left eye pixel is equipped with a vertically polarized filter
- the right eye pixel is equipped with a horizontally polarized filter
- the viewer is provided with polarized glasses with a vertically polarized filter for the left eye and a horizontally polarized filter for the right eye.
- the stereoscopic view may be realized by viewing the screen of the display device using.
- index file not a playlist
- 2D playback devices refer to “Index.bdmv” at the start of playback, but 3D playback devices select “Index.3dmv”.
- the subpath information may be used as described above, or multi_clip_entries for multi-angle may be used. When multi_clip_entries is used, it is desirable to prohibit UO for changing the angle after selecting a stream according to the screen size of the display device so that the stream is not accidentally switched to a stream for another screen size.
- the left view and the right view can be prepared not only for the video stream related to the main part but also for the thumbnail image.
- the 2D playback device displays the conventional 2D thumbnail, but the 3D playback device outputs the left-eye thumbnail and the right-eye thumbnail prepared for 3D according to the 3D display method.
- Each recording layer of the BD-ROM is desirably provided with a stereoscopic / planar view shared area, a stereoscopic view dedicated area, and a planar view dedicated area.
- the stereoscopic / planar view common area is an area that is accessed both when reproducing stereoscopic images and when reproducing planar images, and includes a plurality of extents belonging to a base-view video stream file and a dependent view.
- the stereoscopic viewing dedicated area and the planar viewing dedicated area follow the stereoscopic / planar viewing shared area and exist immediately before the boundary of the recording layer.
- the stereoscopic viewing exclusive area is an area that is accessed immediately before a long jump that occurs during playback of the stereoscopic output mode, and follows the extent belonging to the base view video stream file recorded in the stereoscopic viewing / planar viewing shared area. This is an area in which extents and extents following extents belonging to the dependent-view video stream file recorded in the stereoscopic / planar view shared area are alternately arranged and recorded.
- the plane view exclusive area is an area that is accessed immediately before a long jump that occurs during playback in 2D output mode, and a duplicate of the extent belonging to the base view video stream file recorded in the stereoscopic view exclusive area is recorded.
- Area (Program embodiment)
- the application program shown in each embodiment can be created as follows. First, a software developer uses a programming language to write a source program that implements each flowchart and functional components. In this description, the software developer describes a source program that embodies each flowchart and functional components using a class structure, a variable, an array variable, and an external function call according to the syntax of the programming language.
- the described source program is given to the compiler as a file.
- the compiler translates these source programs to generate an object program.
- Translator translation consists of processes such as syntax analysis, optimization, resource allocation, and code generation.
- syntax analysis lexical analysis, syntax analysis, and semantic analysis of the source program are performed, and the source program is converted into an intermediate program.
- optimization operations such as basic block formation, control flow analysis, and data flow analysis are performed on the intermediate program.
- resource allocation in order to adapt to the instruction set of the target processor, a variable in the intermediate program is allocated to a register or memory of the processor of the target processor.
- code generation each intermediate instruction in the intermediate program is converted into a program code to obtain an object program.
- the object program generated here is composed of one or more program codes that cause a computer to execute the steps of the flowcharts shown in the embodiments and the individual procedures of the functional components.
- program codes such as a processor native code and JAVA (registered trademark) bytecode.
- JAVA registered trademark
- a call statement that calls the external function becomes a program code.
- a program code that realizes one step may belong to different object programs.
- each step of the flowchart may be realized by combining arithmetic operation instructions, logical operation instructions, branch instructions, and the like.
- the programmer activates the linker for these.
- the linker allocates these object programs and related library programs to a memory space, and combines them into one to generate a load module.
- the load module generated in this manner is premised on reading by a computer, and causes the computer to execute the processing procedures and the functional component processing procedures shown in each flowchart.
- Such a program may be recorded on a computer-readable recording medium and provided to the user.
- a repetitive structure in which a plurality of information of a certain type exists can be defined by setting an initial value of a control variable and a repetitive condition in a for statement. Do While statement may be used.
- an arbitrary data structure that defines certain information when a predetermined condition is satisfied uses an if statement that specifies the condition to be satisfied and the variable to be set when the condition is satisfied. Can be described. For this description, a switch statement and a case statement may be used.
- the data structure shown in each embodiment can be described by a grammar of a high-level programming language
- the data structure shown in each embodiment includes syntax analysis, optimization, resource allocation, code generation, and the like.
- a compiler like the above program
- it is recorded on a recording medium in a state of being converted into a computer code that can be read by a computer.
- the data structure described in the high-level programming language is treated as a part other than the method of the class structure, that is, as an array type member variable in the class structure, and forms a part of the program. That is, the data structure of each embodiment is converted into a computer code, recorded on a computer-readable recording medium, and becomes a member variable of the program. Because of such handling, the data structure described so far is substantially the same as the program.
- the BD-ROM drive includes an optical head having a semiconductor laser, a collimator lens, a beam splitter, an objective lens, a condenser lens, and a photodetector.
- the light beam emitted from the semiconductor laser passes through the collimator lens, the beam splitter, and the objective lens, and is condensed on the information surface of the optical disk.
- the condensed light beam is reflected / diffracted on the optical disk, passes through the objective lens, the beam splitter, and the condensing lens and is collected on the photodetector.
- a reproduction signal is generated according to the amount of light collected by the photodetector.
- the recording medium in each embodiment includes all package media such as an optical disk and a semiconductor memory card.
- the recording medium of the present embodiment will be described by taking an example of an optical disc (for example, an existing readable optical disc such as a BD-ROM or DVD-ROM) in which necessary data is recorded in advance.
- a terminal device having a function of writing 3D content including data necessary for carrying out the present invention distributed via broadcasting or a network to an optical disc (for example, the function described on the left may be incorporated in a playback device) It may be good or may be a device different from the playback device) and recorded on a writable optical disc (for example, an existing writable optical disc such as BD-RE, DVD-RAM) and the recorded optical disc
- a writable optical disc for example, an existing writable optical disc such as BD-RE, DVD-RAM
- Embodiments of Semiconductor Memory Card Recording Device and Playback Device Embodiments of a recording apparatus that records the data structure described in each embodiment in a semiconductor memory and a reproducing apparatus that reproduces the data structure will be described.
- a part of the data may be encrypted as necessary from the viewpoint of protecting the copyright and improving the confidentiality of the data.
- the encrypted data may be, for example, data corresponding to a video stream, data corresponding to an audio stream, or data corresponding to a stream including these.
- data for example, a device key
- a key necessary for decrypting the encrypted data in the BD-ROM is stored in advance in the playback device.
- the BD-ROM decrypts the data corresponding to the key necessary for decrypting the encrypted data (for example, MKB (media key block) corresponding to the above-mentioned device key) and the encrypted data.
- Data for encrypting the key itself (for example, the above-described device key and encrypted title key corresponding to MKB) is recorded.
- the device key, the MKB, and the encrypted title key are paired, and are also associated with an identifier (for example, a volume ID) written in an area that cannot be normally copied (an area called BCA) on the BD-ROM. Has been. If this combination is not correct, the code cannot be decrypted.
- the key necessary for decryption (for example, the title key obtained by decrypting the encrypted title key based on the above-mentioned device key, MKB and volume ID) can be derived.
- the encrypted data can be decrypted using the necessary key.
- the loaded BD-ROM When the loaded BD-ROM is played back on a playback device, for example, if the device key that is paired with (or corresponding to) the encrypted title key and MKB in the BD-ROM is not in the playback device, it is encrypted. The data is not played back. This is because the key (title key) required to decrypt the encrypted data is recorded on the BD-ROM with the key itself encrypted (encrypted title key), and a combination of MKB and device key. If is not correct, the key necessary for decryption cannot be derived.
- the playback apparatus is configured such that the video stream is decoded by the decoder using the title key, and the audio stream is decoded by the audio decoder.
- the above is the mechanism for protecting the copyright of the data recorded on the BD-ROM.
- this mechanism is not necessarily limited to the BD-ROM.
- a readable / writable semiconductor memory for example, SD
- the present invention can be implemented even when applied to a portable semiconductor memory card such as a card.
- an optical disk is configured to read data via an optical disk drive, whereas when a semiconductor memory card is used, data is read via an I / F for reading data in the semiconductor memory card. What is necessary is just to comprise so that it may read.
- the playback device described in each embodiment is also realized as a terminal device that receives data (distribution data) corresponding to the data described in this embodiment from a distribution server of an electronic distribution service and records it on a semiconductor memory card. be able to.
- Such a terminal device may be configured such that the playback device described in each embodiment can perform such an operation, or stores distribution data in a semiconductor memory separately from the playback device of the present embodiment. It may be configured to be performed by a dedicated terminal device that performs the above. Here, an example performed by the playback apparatus will be described. Further, an SD card will be described as an example of the recording destination semiconductor memory.
- the playback device When recording distribution data on an SD memory card inserted in a slot provided in the playback device, first, transmission of distribution data is requested to a distribution server (not shown) that stores the distribution data. At this time, the playback device uses identification information for uniquely identifying the inserted SD memory card (for example, an identification number unique to each SD memory card, more specifically, for example, a serial number of the SD memory card). And the read identification information is transmitted to the distribution server together with the distribution request.
- identification information for uniquely identifying the inserted SD memory card for example, an identification number unique to each SD memory card, more specifically, for example, a serial number of the SD memory card.
- the identification information for uniquely identifying the SD memory card corresponds to, for example, the volume ID described above.
- necessary data for example, a video stream, an audio stream, etc.
- a key for example, a title key
- the distribution server holds a secret key and is configured so that different public key information can be dynamically generated for each unique identification number of the semiconductor memory card.
- the distribution server is configured to be able to encrypt the key (title key) necessary for decrypting the encrypted data (that is, configured to generate an encrypted title key). ).
- the generated public key information includes, for example, information corresponding to the above-described MKB, volume ID, and encrypted title key. If the combination of the identification number unique to the semiconductor memory, the public key body included in the public key information described later, and the device key recorded in advance on the playback device is correct, the encrypted data is, for example, a key necessary for decryption (for example, Based on the device key, MKB, and the identification number unique to the semiconductor memory, a title key obtained by decrypting the encrypted title key) is obtained, and using the obtained key (title key) necessary for decryption, Encrypted data can be decrypted.
- a key necessary for decryption for example, Based on the device key, MKB, and the identification number unique to the semiconductor memory, a title key obtained by decrypting the encrypted title key
- the playback device records the received public key information and distribution data in the recording area of the semiconductor memory card inserted in the slot.
- the received public key information includes, for example, a public key body (for example, the above-mentioned MKB and encrypted title key), signature information, a unique identification number of the semiconductor memory card, and a device list indicating information on devices to be invalidated. Yes.
- the signature information includes, for example, a hash value of public key information.
- This may be a device that is likely to be played illegally, such as a device key pre-recorded on the playback device, an identification number of the playback device, or an identification number of a decoder included in the playback device, or a component included in the device, or This is information for uniquely identifying a function (program).
- the following describes the playback of encrypted data among the distribution data recorded in the recording area of the semiconductor memory card.
- (1) Check whether the identification information unique to the semiconductor memory included in the public key information matches the unique identification number stored in advance in the semiconductor memory card. (2) The public key information calculated in the playback device. Check whether hash value and hash value included in signature information match (3) Check whether playback device that performs playback is capable of unauthorized playback based on information shown in device list included in public key information (For example, check whether the device key shown in the device list included in the public key information matches the device key stored in advance in the playback device) To do. These checks may be performed in any order.
- the identification information unique to the semiconductor memory included in the public key information does not match the unique identification number stored in advance in the semiconductor memory, and is calculated by the playback device. If the hash value of the key information and the hash value included in the signature information do not match or if it is determined that there is a possibility that the playback device that performs playback may be played back illegally, the playback device Control to prevent decryption of encrypted data.
- the hash value and signature of the public key information calculated in the playback device in which the unique identification information of the semiconductor memory card included in the public key information matches the unique identification number stored in advance in the semiconductor memory card. If it is determined that the hash values included in the information match and the playback device that performs playback is not likely to be played back illegally, the identification number unique to the semiconductor memory, the public key body included in the public key information, And the device key pre-recorded on the playback device is determined to be correct, and is obtained by decrypting the encrypted title key based on the key necessary for decryption (device key, MKB and identification number unique to the semiconductor memory) The encrypted data is decrypted using the title key.
- the video decoder decrypts the video stream by using the above-described key necessary for decryption (the title key obtained by decrypting the encrypted title key).
- the audio decoder decodes (decodes) the audio stream using the key necessary for the above-described decryption.
- the unique identifier of the semiconductor memory card recorded in advance on the semiconductor memory card is stored in a highly confidential recording area.
- a semiconductor memory card for example, an SD memory card as an example, the serial number of an SD memory card
- illegal copying can be made easily.
- a different unique identification number is assigned to each of the plurality of semiconductor memory cards, but if the tampering is performed so that the unique identification numbers are the same, the determination of (1) above is made. This is because it makes no sense, and there is a possibility that illegal copies corresponding to the number of falsifications will be made.
- a recording area for recording highly confidential data such as a unique identifier of the semiconductor memory card is used as a recording area for storing normal data (the first area).
- a control circuit for accessing the second recording area Provided in a different recording area (referred to as a second recording area), a control circuit for accessing the second recording area, and a second recording area. Access to the access point is configured so that it can be accessed only through the control circuit.
- the data recorded in the second recording area is encrypted and recorded
- the control circuit includes, for example, a circuit for decrypting the encrypted data.
- the encryption is decrypted and the decrypted data is returned.
- the control circuit holds information on the storage location of the data recorded in the second recording area, and if there is a data access request, specifies the storage location of the corresponding data, and specifies the specified storage location It may be configured to return the data read from the.
- An application that operates on a playback apparatus and requests to record on a semiconductor memory card using electronic distribution is used to record data (for example, semiconductor) to the control circuit via the memory card I / F.
- data for example, semiconductor
- the control circuit that has received the request reads the data recorded in the second recording area and returns it to the application operating on the playback device. It is configured to request a distribution server for a required data distribution request together with a unique identification number of the semiconductor memory card, and record the public key information sent from the distribution server and the corresponding distribution data in the first recording area. That's fine.
- an application that operates on the playback device and requests recording to the semiconductor memory card using electronic distribution is used to record data (in the second recording area) to the control circuit via the memory card I / F.
- the application before issuing an access request to an identification number unique to a semiconductor memory, it is desirable to check in advance whether the application has been tampered with.
- a check using a digital certificate conforming to the existing X.509 specification may be used.
- System LSI System LSI
- the system target decoder, the playback control unit 7, the program execution unit 11, and other parts centering on logic elements are preferably configured by a system LSI.
- System LSI refers to a bare chip mounted on a high-density substrate and packaged.
- a system LSI that includes a plurality of bare chips mounted on a high-density substrate and packaged to give the bare chip an external structure like a single LSI is also included in system LSIs (such systems LSI is called a multichip module.)
- system LSIs are classified into QFP (Quad-Flood Array) and PGA (Pin-Grid Array).
- QFP is a system LSI with pins attached to the four sides of the package.
- the PGA is a system LSI with many pins attached to the entire bottom surface.
- pins serve as an interface with other circuits. Since pins in the system LSI have such an interface role, the system LSI plays a role as the core of the playback apparatus 200 by connecting other circuits to these pins in the system LSI.
- Such a system LSI can be incorporated in various devices that handle video playback, such as a TV, a game, a personal computer, and a one-seg mobile phone as well as the playback device 200, and can broaden the application of the present invention.
- the system LSI architecture should conform to the Uniphier architecture.
- a system LSI that conforms to the Uniphier architecture consists of the following circuit blocks.
- ⁇ Data parallel processor DPP This is a SIMD-type processor in which multiple element processors operate in the same way. By operating the arithmetic units incorporated in each element processor simultaneously with a single instruction, the decoding process for multiple pixels constituting a picture is performed in parallel. Plan
- Instruction parallel processor IPP This is a "Local Memory Controller” consisting of instruction RAM, instruction cache, data RAM, and data cache, "Processing Unit” consisting of instruction fetch unit, decoder, execution unit and register file, and Processing Unit part for parallel execution of multiple applications. It consists of a “Virtual Multi Processor Unit section” to be performed.
- MPU block This is a peripheral interface such as ARM core, external bus interface (Bus Control Unit: BCU), DMA controller, timer, vector interrupt controller, UART, GPIO (General Purpose Input Output), synchronous serial interface, etc. Consists of.
- -Stream I / O block This performs data input / output to / from drive devices, hard disk drive devices, and SD memory card drive devices connected to the external bus via the USB interface or ATA Packet interface.
- ⁇ AVI / O block This is composed of audio input / output, video input / output, and OSD controller, and performs data input / output with TV and AV amplifier.
- Memory control block This is a block that realizes reading and writing of the SD-RAM connected via the external bus.
- the internal bus connection part that controls the internal connection between each block, the SD-RAM connected outside the system LSI It consists of an access control unit that transfers data to and from the RAM, and an access schedule unit that adjusts SD-RAM access requests from each block.
- a circuit diagram of a portion to be a system LSI is created, and the components in the configuration diagram are realized using circuit elements, ICs, and LSIs. Then, if each component is embodied, a bus connecting circuit elements, ICs, LSIs, peripheral circuits thereof, an interface with the outside, and the like are defined. In addition, connection lines, power supply lines, ground lines, clock signal lines, and the like will be defined. In this regulation, the circuit diagram is completed while adjusting the operation timing of each component in consideration of the specifications of the LSI, and making adjustments such as ensuring the necessary bandwidth for each component.
- Mounting design refers to where on the board the parts (circuit elements, ICs, and LSIs) on the circuit board created by circuit design are placed, or how the connection lines on the circuit board are placed on the board. This is a board layout creation operation for determining whether to perform wiring.
- the mounting design result is converted into CAM data and output to equipment such as an NC machine tool.
- NC machine tools perform SoC implementation and SiP implementation based on this CAM data.
- SoC (System on chip) mounting is a technology that burns multiple circuits on a single chip.
- SiP (System-in-Package) packaging is a technology that combines multiple chips into one package with resin or the like.
- the integrated circuit generated as described above may be called IC, LSI, super LSI, or ultra LSI depending on the degree of integration.
- the hardware configuration shown in each embodiment can be realized.
- the LUT is stored in the SRAM, and the content of the SRAM disappears when the power is turned off.
- the LUT that realizes the hardware configuration shown in each embodiment is defined by the definition of the configuration information. Must be written to SRAM.
- hardware corresponding to middleware and system LSI hardware other than system LSI, interface part to middleware, interface part of middleware and system LSI, necessary hardware other than middleware and system LSI
- a specific function is provided by operating in cooperation with each other.
- the user interface part, middleware part, and system LSI part of the playback device can be developed independently and in parallel, making development more efficient It becomes possible. There are various ways to cut each interface.
- the information recording medium according to the present invention stores 3D video, but since it can be played back by either 2D video playback device or 3D video playback device, it stores cluster # 3D video that is compatible with compatibility. It is possible to supply moving image contents such as movie titles to the market, and to activate the movie market and consumer equipment market. Therefore, the recording medium and the playback device according to the present invention have high applicability in the movie industry and the consumer equipment industry.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Computer Hardware Design (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Databases & Information Systems (AREA)
- Human Computer Interaction (AREA)
- Library & Information Science (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
- Television Signal Processing For Recording (AREA)
- Signal Processing For Digital Recording And Reproducing (AREA)
- Compression Or Coding Systems Of Tv Signals (AREA)
- Management Or Editing Of Information On Record Carriers (AREA)
- Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)
Abstract
Description
前記再生区間情報は、ビデオストリームを格納した前記ストリームファイルを指定するファイル参照情報を含み、
前記ストリームファイルは、インターリーブされたトランスポートストリームファイルと、通常形式のトランスポートストリームファイルであり、
前記インターリーブされたトランスポートストリームファイルは、レフトビュービデオストリームを格納したトランスポートストリームを分割することで得られる複数の分割部分、及び、ライトビュービデオストリームを格納したトランスポートストリームを分割することで得られる複数の分割部分のそれぞれを、交互に配置することで構成されており、前記ファイル参照情報と同じ識別番号と、インターリーブされている旨を示す拡張子とによって特定され、
前記通常形式のトランスポートストリームファイルは、前記レフトビュービデオストリーム及び前記ライトビュービデオストリームの何れか一方であって、単独再生することができるベースビュービデオストリームを格納しており、前記ファイル参照情報と同じ識別番号と、通常形式である旨を示す拡張子とによって特定されることを特徴とする。
図面を参照しながら、上記課題解決手段を具備した記録媒体、及び、再生装置の実施形態について説明する。先ず始めに、立体視の原理について簡単に述べる。
次に、記録媒体を造るための形態、つまり、記録媒体の生産行為の形態について説明する。
図5は、ファイルシステムを前提にした光ディスクのアプリケーションフォーマットを示す図である。
BDMVディレクトリはBD-ROMで扱うAVコンテンツや管理情報などのデータが記録されているディレクトリである。BDMVディレクトリの配下には、「PLAYLISTディレクトリ」、「CLIPINFディレクトリ」、「STREAMディレクトリ」、「BDJOディレクトリ」、「JARディレクトリ」と呼ばれる5つのサブディレクトリが存在し、BDMVディレクトリには、「index.bdmv」,「MovieObject.bdmv」の2種類のファイルが配置されている。
ビデオストリームは映画のプライマリビデオおよびセカンダリビデオを示す。ここでプライマリビデオとはピクチャインピクチャにおいて親画像として表示される通常の映像を示し、セカンダリビデオとは、ピクチャインピクチャにおいて小画面で表示される映像のことである。プライマリビデオには、レフトビュービデオ、ライトビュービデオの2種類があり、セカンダリビデオにも、レフトビュービデオ、ライトビュービデオの2種類がある。
オーディオストリームは、映画の主音声部分を示す。オーディオストリームは、ドルビーAC-3、Dolby Digital Plus、MLP、DTS、DTS-HD、または、リニアPCMのなどの方式で圧縮・符号化記録されている。オーディオストリームには、プライマリオーディオストリーム、セカンダリオーディオストリームの2種類がある。プライマリオーディオストリームは、ミキシング再生を行う場合、主音声となるべきオーディオストリームであり、セカンダリオーディオストリームは、ミキシング再生を行う場合、副音声をとなるべきオーディオストリームである。
Presentation Graphicsストリーム(PGストリーム)は、映画の字幕や、キャラクタ等、ピクチャと緻密に同期すべきグラフィクスを示すグラフィクスストリームであり、英語、日本語、フランス語というように複数言語についてのストリームが存在する。
Interactive Graphicsストリーム(IGストリーム)は、リモコンを通じた対話制御を実現するグラフィクスストリームである。IGストリームにて定義される対話制御は、DVD再生装置上の対話制御と互換性がある対話制御である。かかるIGストリームは、ICS(Interactive Composition Segment)、PDS(Palette Difinition Segment)、ODS(Object Definition Segment)と呼ばれる機能セグメントからなる。ODS(Object Definition Segment)は、グラフィクスオブジェクトを定義する機能セグメントである。このグラフィクスオブジェクトが複数集まって、対話画面上のボタンが描画される。PDS(Palette Difinition Segment)は、グラフィクスオブジェクトの描画にあたっての、発色を規定する機能セグメントである。ICS(Interactive Composition Segment)は、ユーザ操作に応じてボタンの状態を変化させるという状態変化を実現する機能セグメントである。ICSは、ボタンに対して確定操作がなされた際、実行すべきボタンコマンドを含む。インタラクティブグラフィックスストリームは、画面上にGUI部品を配置することにより作成される対話画面を示している。
<TSパケット列>
図7(b)は、AVクリップに最終的に書き込まれるTSパケットの形式を示している。第1段目は、TSパケット列を示し、第2段目は、ソースパケット列を示す。第3段目は、AVクリップを示す。
図8は、レフトビューAVクリップがどのように多重化されるかを模式的に示す図である。まず、レフトビュービデオストリーム、及び、オーディオストリームを(第1段目)、それぞれPESパケット列に変換し(第2段目)、ソースパケット列に変換する(第3段目)。同じくレフトビュープレゼンテーショングラフィックスストリームおよびレフトビューインタラクティブグラフィックス(第7段目)を、それぞれPESパケット列に変換し(第6段目)、更にソースパケット列に変換する(第5段目)。こうして得られた、ビデオ、オーディオ、グラフィクスを構成するソースパケットをそのATSの順に配列してゆく。これはソースパケットは、そのATSに従い、リードバッファに読み込まれるべきだからである。こうして、ATSに従ってソースパケットが配列されれば、レフトビューAVクリップが得られることになる。このレフトビューAVクリップは、リードバッファがアンダーフローしないサイズに定められており、記録媒体への記録の対象になる。
ライトビュー用リードバッファ=Rmax1×"ジャンプを伴いながらレフトビュー用リードバッファをフルにする時間"
という関係を満たすよう、ライトビュー用リードバッファの容量を定めねばならない。ここでジャンプとは、ディスクシークと同義である。何故なら、BD-ROMにおいて記録に確保できる連続領域は有限であり、レフトビュービデオストリーム及びライトビュービデオストリームは、必ずしも、隣合わせで記録されるとは限らず、飛び飛びの領域に記録されることも有り得るからである。
レフトビュー用リードバッファにデータを読み出すにあたっては、ライトビューAVクリップからレフトビューAVクリップへのジャンプ時間(Tjump)と、レフトビューAVクリップからライトビューAVクリップへのジャンプ時間(Tjump)とを考慮する必要があるので、
レフトビュー用リードバッファの蓄積には(2×Tjump+RB2/(Rud-Rmax2))という時間が必要になる。
RB1≧Rmax1×[2×Tjump+RB2/(Rud-Rmax2)]
になる。
同様の手順で、レフトビュー用リードバッファの容量RB2を求めると、
RB2≧Rmax2×[2×Tjump+RB1/(Rud-Rmax1)]
になる。
0x1220~123FのパケットIDを有するソースパケットはレフトビューPGストリーム、
0x1420~143FのパケットIDを有するソースパケットは レフトビューIGストリーム
0x1100から0x111FまでのPIDを有するソースパケットはオーディオストリームを構成する。
エクステントは、パーティション領域において、物理的に連続する複数のセクタ上に形成される。パーティション領域は、「ファイルセット記述子が記録された領域」、「終端記述子が記録された領域」、「ROOTディレクトリ領域」、「BDMVディレクトリ領域」、「JARディレクトリ領域」、「BDJOディレクトリ領域」、「PLAYLISTディレクトリ領域」、「CLIPINFディレクトリ領域」、「STREAMディレクトリ領域」から構成され、ファイルシステムによってアクセスされる領域のことである。以降、これらの領域について説明する。
図13は、クリップ情報ファイルの内部構成を示す図である。クリップ情報ファイルは、本図に示すようにAVクリップの管理情報であり、AVクリップと1対1に対応している。引き出し線ch1は、クリップ情報ファイルの内部構成をクローズアップして示している。この引出線に示すように、クリップ情報ファイルは、「クリップ情報」と、「ストリーム属性情報」と、「エントリマップテーブル」と、「3Dメタデータ」とから構成される。
このエントリマップを利用することにより、再生機はビデオストリームの時間軸上の任意の地点に対応するAVクリップのファイル位置を特定することが出来る。例えば、早送り・巻戻しの特殊再生の際には、エントリマップに登録されるIピクチャを特定し選択して再生することによりAVクリップを解析することなく効率的に処理を行うことが出来る。
プログラム実行部11は、BDプログラムファイルに格納されたプログラムを実行するプロセッサである。格納されたプログラムに従って動作を行い、次のような制御を行う。(1)再生制御部7に対してプレイリスト再生を命令する。(2)システムターゲットデコーダに対してメニューやゲームのグラフィックスのためのPNG・JPEGを転送して画面に表示する。これらはプログラムの作りに応じて自由に行うことができ、どのように制御するかは、オーサリング工程によるBD-Jアプリケーションのプログラミング工程によって決まる。
HDMVモジュール13は、HDMVモードの動作主体となるDVD仮想プレーヤであり、HDMVモードの実行主体となる。本モジュールは、コマンドインタプリタを具備し、ムービーオブジェクトを構成するナビゲーションコマンドを解読して実行することでHDMVモードの制御を実行する。ナビゲーションコマンドは、DVD-Videoと似たようなシンタックスで記述されているため、かかるナビゲーションコマンドを実行することにより、DVD-Videoライクな再生制御を実現することができる。
BD-Jプラットフォーム14は、BD-Jモードの動作主体であるJavaプラットフォームであり、Java2Micro_Edition(J2ME) Personal Basis Profile(PBP 1.0)と、Globally Executable MHP specification(GEM1.0.2)for package media targetsとをフル実装しており、クラスローダ、バイトコードインタプリタ、アプリケーションマネージャから構成される。
本実施形態では、表示装置300、3D眼鏡400がどのような機能をもち、その内部構成がどのようなものかを図39を参照しながら説明する。
3D眼鏡をすでに装着しているため、眼鏡に内蔵されたイヤフォンを用いれば、映像と音声を完全に独立させることも可能であり、リビングでのチャンネル権争いの回避、1つの画面で対戦ゲームをするなど、応用範囲は広がる。また、ステップを増やすことにより、3チャンネルやさらに多くのチャンネルを1つの画面で見ることが可能となる。
本実施形態は、再生装置と、表示装置とのネゴシエーションに関する実施形態である。ユーザ宅内におけるホームシアターシステムの構築事情には、それぞれの家庭独特のものがあり、再生装置は、表示装置と接続された際、接続相手となる表示装置とネゴシエーションを行って、どのプレイリストを再生に供するかを切り替える必要がある。
またネゴシエーションにあたっては、複数のライトビューを切り替える必要がある。複数のライトビューを切り替える理由として、表示装置の画面の大きさの違いがある。左目と右目の間隔は個人の個体差があってもだいたい同じであるが、表示装置は20インチのものもあれば、150インチのものもある。50インチの表示装置を想定して、目と目の間隔6.5センチで作られた映像も、150インチの表示装置では3倍の19.5センチの目の間隔になってしまい、3D映像として認識が難しくなる。そのため、様々なサイズの表示装置でレフトビューとライトビューの差が、6.5センチになるレフトビュー、ライトビューの組みを納めておけば、表示装置に合わせてビデオストリームを選ぶことにより、最適なレフトビュービデオストリーム/ライトビュービデオストリームの組み合わせを選ぶことができる。
(第4実施形態)
本実施形態は、立体視を行うにあたって、ビデオストリームと共に、どのようなPGストリーム、IGストリームを選択すべきかという改良を実現する。
3D対応のプレイリストは、第1実施形態で示したように、ストリーム選択テーブルをさらに2D用のSTN_tableと、3D再生用のSTN_table_SSとに分けて管理する。そして2D再生でしか利用しない映像/音声/PG/IGのストリーム登録と、3Dで利用する映像/音声/PG/IGのストリーム登録は異なるエントリー群に登録する。2D映像を選択している場合は、3D用に用意された音声/PG/IGは選択できない。また、同様に3D映像を選択している場合は、2D用に用意された音声/PG/IGは選択できない。
2D用に作成されたPGストリームと、3D用に作成されたPGストリームとでは、奥行きの有無、位置や角度に違いがあるので、立体視のためのビデオストリームの再生時に、2DPGストリームが選択されて、3Dビデオストリームと共に再生されることは、オーサリング者にとっては何とか避けたいところである。
(第5実施形態)
第1実施形態の冒頭で述べたように、立体視には様々な原理があり、市場で流通しつつある製品は、様々な3D方式に対応していると考えられる。3Dの再生方式にはいろいろな方式が存在し、また、方式により表示装置側の対応・非対応も異なるため、再生装置のシステムパラメータは複数の方式を表現できる形式が望ましい。ここでは、3Dの再生方式として、2画面のビデオを独立して送る2画面ステレオ再生方式、サイド・バイ・サイド方式、横方向2倍方式、2D+奥行き情報方式などを例に挙げるが、その他、表示装置が対応可能な方式がある場合は、対応する方式の実行の可否を識別できるように、PSRのビットアサインを定める。
本実施形態では、第1実施形態で述べた記録方法を実施するための記録装置について説明する。
以上がリアルタイムレコーディングによる記録方法を実行する記録装置である。続いて、プレフォーマットレコーディングによる記録方法を実行する記録装置について説明する。
図50は、記録装置の内部構成を示す図である。本図に示すように本発明にかかる記録装置は、ビデオエンコーダ501、素材制作部502、シナリオ生成部503、BDプログラム制作部504、多重化処理部505、フォーマット処理部506により構成される。
記録装置は、マネージドコピーによってデジタルストリームの書き込むものでもよい。
以上、本願の出願時点において、出願人が知り得る最良の実施形態について説明したが、以下に示す技術的トピックについては、更なる改良や変更実施を加えることができる。各実施形態に示した通り実施するか、これらの改良・変更を施すか否かは、何れも任意的であり、実施する者の主観によることは留意されたい。
第1実施形態で説明の前提とした視差画像方式は、左右の映像を時間軸方向で交互に表示させるために、例えば、通常の2次元の映画であれば1秒に24枚の映像を表示させるのに対して、左右の映像合わせて1秒に48枚の映像を表示させる必要がある。従って、この方式では、一画面の書き換えが比較的早い表示装置において好適である。この視差画像を用いた立体視は、既に遊園地の遊具などで一般的に使用されており、技術的にも確立されているため、家庭における実用化に最も近いものと言える。視差画像を用いた立体視のための方法はこれらの他にも、2色分離方式などさまざまな技術が提案されている。本実施形態においては、継時分離方式あるいは偏光メガネ方式を例として用いて説明したが、視差画像を用いる限りこれら2方式に限定するものではない。
プレイリストではなくインデックスファイルを、2D再生装置と3D対応再生装置で区別し、2D再生装置では再生開始時に「Index.bdmv」を参照するが、3D再生装置では「Index.3dmv」を選択させるといった方法もとることができる。
(複数ストリームを扱うためのデータ構造)
複数のストリームがある場合、上記のようにサブパス情報を使ってもよいし、マルチアングルのためのmulti_clip_entriesを使ってもよい。multi_clip_entriesを使う場合、表示装置の画面のサイズに合わせてストリームを選択した後は、誤って他の画面サイズ用のストリームに切り替わらないよう、アングル変更のUOを禁止することが望ましい。
レフトビューとライトビューを用意するのは、本編に関わるビデオストリームだけではなく、サムネイル画像に適用することも可能である。ビデオストリームの場合と同様に、2D再生装置では従来の2D用サムネイルを表示するが、3D再生装置では3D用に用意された左目サムネイルと右目サムネイルを、3D表示方式に合わせて出力する。
BD-ROMの各記録層には、立体視/平面視共用領域と、立体視専用領域と、平面視専用領域とを設けることが望ましい。
(プログラムの実施形態)
各実施形態に示したアプリケーションプログラムは、以下のようにして作ることができる。先ず初めに、ソフトウェア開発者は、プログラミング言語を用いて、各フローチャートや、機能的な構成要素を実現するようなソースプログラムを記述する。この記述にあたって、ソフトウェア開発者は、プログラミング言語の構文に従い、クラス構造体や変数、配列変数、外部関数のコールを用いて、各フローチャートや、機能的な構成要素を具現するソースプログラムを記述する。
上述したようなデータ構造のうち、ある決まった型の情報が複数存在するという繰り返し構造は、for文に、制御変数の初期値と、繰り返し条件とを設定することで定義することができる。Do While文を用いてもよい。
BD-ROMドライブは、半導体レーザ、コリメートレンズ、ビームスプリッタ、対物レンズ、集光レンズ、光検出器を有する光学ヘッドを備える。半導体レーザから出射された光ビームは、コリメートレンズ、ビームスプリッタ、対物レンズを通って、光ディスクの情報面に集光される。
各実施の形態における記録媒体は、光ディスク、半導体メモリーカード等、パッケージメディア全般を含んでいる。本実施の形態の記録媒体は予め必要なデータが記録された光ディスク(例えばBD-ROM、DVD-ROMなどの既存の読み取り可能な光ディスク)を例に説明をするが、これに限定される必要はなく、例えば、放送またはネットワークを経由して配信された本発明の実施に必要なデータを含んだ3Dコンテンツを光ディスクへ書き込む機能を有する端末装置(例えば左記の機能は再生装置に組み込まれていても良いし、再生装置とは別の装置であってもよい)を利用して書き込み可能な光ディスク(例えばBD-RE、DVD-RAMなどの既存の書き込み可能な光ディスク)に記録し、この記録した光ディスクを本発明の再生装置に適用しても本発明の実施は可能である。
各実施の形態で説明をしたデータ構造を半導体メモリーに記録する記録装置、及び、再生する再生装置の実施形態について説明する。
(受信装置としての実施形態)
各実施形態で説明した再生装置は、本実施の形態で説明をしたデータに相応するデータ(配信データ)を電子配信サービスの配信サーバから受信し、半導体メモリカードに記録する端末装置としても実現することができる。
(1) 公開鍵情報に含まれる半導体メモリー固有の識別情報と半導体メモリーカードに予め記憶されている固有の識別番号とが一致するかどうかのチェック
(2) 再生装置内で算出した公開鍵情報のハッシュ値と署名情報に含まれるハッシュ値が一致するかのチェック
(3) 公開鍵情報に含まれるデバイスリストに示される情報に基づいて、再生を行う再生装置が不正な再生が可能かどうかのチェック(例えば公開鍵情報に含まれるデバイスリストに示されるデバイスキーと、再生装置に予め記憶されたデバイスキーが一致するかどうかのチェック)
を行なう。これらのチェックを行なう順番は、どのような順序で行なってもよい。
(システムLSI)
再生装置の内部構成のうち、システムターゲットデコーダや、再生制御部7、プログラム実行部11等、ロジック素子を中心とした部分は、システムLSIで構成することがことが望ましい。
これは、複数の要素プロセッサが同一動作するSIMD型プロセッサであり、各要素プロセッサに内蔵されている演算器を、1つの命令で同時動作させることで、ピクチャを構成する複数画素に対するデコード処理の並列化を図る。
これは、命令RAM、命令キャッシュ、データRAM、データキャッシュからなる「Local Memory Controller」、命令フェッチ部、デコーダ、実行ユニット、レジスタファイルからなる「Processing Unit部」、複数アプリケーションの並列実行をProcessing Unit部に行わせる「Virtual Multi Processor Unit部」で構成される。
これは、ARMコア、外部バスインターフェイス(Bus Control Unit:BCU)、DMAコントローラ、タイマー、ベクタ割込コントローラといった周辺回路、UART、GPIO(General Purpose Input Output)、同期シリアルインターフェイスなどの周辺インターフェイスで構成される。
これは、USBインターフェイスやATA Packetインターフェイスを介して、外部バス上に接続されたドライブ装置、ハードリディスクドライブ装置、SDメモリカードドライブ装置とのデータ入出力を行う。
これは、オーディオ入出力、ビデオ入出力、OSDコントローラで構成され、テレビ、AVアンプとのデータ入出力を行う。
これは、外部バスを介して接続されたSD-RAMの読み書きを実現するブロックであり、各ブロック間の内部接続を制御する内部バス接続部、システムLSI外部に接続されたSD-RAMとのデータ転送を行うアクセス制御部、各ブロックからのSD-RAMのアクセス要求を調整するアクセススケジュール部からなる。
そうして、各構成要素を具現化してゆけば、回路素子やIC,LSI間を接続するバスやその周辺回路、外部とのインターフェイス等を規定する。更には、接続線、電源ライン、グランドライン、クロック信号線等も規定してゆく。この規定にあたって、LSIのスペックを考慮して各構成要素の動作タイミングを調整したり、各構成要素に必要なバンド幅を保証する等の調整を加えながら、回路図を完成させてゆく。
本実施の形態は、ミドルウェアとシステムLSIに対応するハードウェア、システムLSI以外のハードウェア、ミドルウェアに対するインターフェイスの部分、ミドルウェアとシステムLSIのインターフェイスの部分、ミドルウェアとシステムLSI以外の必要なハードウェアへのインターフェイスの部分、ユーザインターフェースの部分で実現し、これらを組み込んで再生装置を構成したとき、それぞれが連携して動作することにより特有の機能が提供されることになる。
200 再生装置
300 テレビ
400 3D眼鏡
500 リモコン
1 BDドライブ
2a,b リードバッファ
4 システムターゲットデコーダ
5b プレーンメモリセット
5b プレーン合成部
6 HDMI送受信部
7 再生制御部
9 管理情報メモリ
10 レジスタセット
11 プログラム実行部
12 プログラムメモリ
13 HDMVモジュール
14 BD-Jプラットフォーム
15 ミドルウェア
16 モード管理モジュール
17 ユーザイベント処理部
18 ローカルストレージ
19 不揮発メモリ
23 PIDフィルタ
27 PIDフィルタ
31 プライマリビデオデコーダ
32 レフトビュービデオプレーン
33 ライトビュービデオプレーン
34 セカンダリビデオデコーダ
35 セカンダリビデオプレーン
36 PGデコーダ
37 PGプレーン
38 IGデコーダ
39 IGプレーン
40 プライマリオーディオデコーダ
41 セカンダリオーディオデコーダ
42 ミキサー
Claims (17)
- プレイリスト情報と、ストリームファイルとが記録された記録媒体であって、
前記プレイリスト情報は、1つ以上の再生区間情報を含み、
前記再生区間情報は、
ビデオストリームを格納した前記ストリームファイルを指定するファイル参照情報を含み、
前記ストリームファイルは、インターリーブされたトランスポートストリームファイルと、通常形式のトランスポートストリームファイルであり、
前記インターリーブされたトランスポートストリームファイルは、レフトビュービデオストリームを格納したトランスポートストリームを分割することで得られる複数の分割部分、及び、ライトビュービデオストリームを格納したトランスポートストリームを分割することで得られる複数の分割部分のそれぞれを、交互に配置することで構成されており、前記ファイル参照情報と同じ識別番号と、インターリーブされている旨を示す拡張子とによって特定され、
前記通常形式のトランスポートストリームファイルは、前記レフトビュービデオストリーム及び前記ライトビュービデオストリームの何れか一方であって、単独再生することができるベースビュービデオストリームを格納しており、前記ファイル参照情報と同じ識別番号と、通常形式である旨を示す拡張子とによって特定される
ことを特徴とする記録媒体。 - 前記記録媒体は、ストリームファイル用ディレクトリと、インターリーブファイル用ディレクトリとを含み、
前記通常形式のトランスポートストリームファイルは、前記ストリームファイル用ディレクトリに格納され、
前記インターリーブされたトランスポートストリームファイルは、前記インターリーブファイル用ディレクトリに格納される
ことを特徴とする請求項1記載の記録媒体。 - プレイリスト情報に従い、ビデオストリームを再生する再生装置であって、
前記プレイリスト情報におけるファイル参照情報で特定されるトランスポートストリームファイルを記録媒体から読み出す読出手段と、
読み出されたトランスポートストリームファイルに格納されているビデオストリームに含まれる圧縮ピクチャデータの供給を受けてデコードを行い、非圧縮のピクチャデータを得るデコーダと、
自機の出力モードを格納しているモードレジスタと、
前記モードレジスタに格納されている出力モードに従い、デコーダによって得られた非圧縮のピクチャデータを出力する出力手段とを備え、
前記読出手段は、
出力モードが平面視出力モードである場合、前記プレイリスト情報におけるファイル参照情報と、トランスポートストリーム形式を示す拡張子とによって特定される通常形式のトランスポートストリームファイルを読み出し、
出力モードが立体視出力モードである場合、前記プレイリスト情報におけるファイル参照情報と、インターリーブされていることを示す拡張子とによって特定されるインターリーブされたトランスポートストリームファイルを読み出す
ことを特徴とする再生装置。 - 前記再生装置は、
ユーザ設定を示すレジスタと、
接続されている表示装置が立体視再生に対応しているか否かを示すケーパビリティレジスタと、
複数の条件が成立するかどうかを判定して、複数の条件が成立した場合に、出力モードを立体視出力モードに設定する設定手段とを備え、
前記複数の条件のうち第1条件は、前記ユーザ設定を示すレジスタが、ユーザが立体視再生を希望する旨を示しているという条件であり、
前記複数の条件のうち第2条件は、前記接続されている表示装置が立体視再生に対応しているか否かを示すケーパビリティレジスタが、接続されている表示装置が立体視再生に対応していることを示しているという条件であり、
前記第1条件及び前記第2条件が成立する場合、前記出力モードを立体視出力モードに設定する
ことを特徴とする請求項3記載の再生装置。 - 前記再生装置の動作モードには、コマンドインタプリタが動作主体になる第1モード、バイトコードインタプリタが動作主体になる第2モードがあり、
第2モードの設定中に、出力モードの切り替えが発生した際、バイトコードインタプリタによって実行されるバイトコードアプリケーションに、イベントを通知する
ことを特徴とする請求項4記載の再生装置。 - プレイリスト情報におけるファイル参照情報で特定されるトランスポートストリームファイルを記録媒体から読み出すドライブ装置と共に、再生装置に組込むことができるシステムLSIであって、
読み出されたトランスポートストリームファイルに格納されているビデオストリームに含まれる圧縮ピクチャデータの供給を受けてデコードを行い、非圧縮のピクチャデータを得るデコーダと、
自機の出力モードを格納しているモードレジスタと、
前記モードレジスタに格納されている出力モードに従い、デコーダによって得られた非圧縮のピクチャデータを出力するよう再生装置の制御を行う制御手段とを備え、
前記制御手段は、
出力モードが平面視出力モードである場合、前記プレイリスト情報におけるファイル参照情報と、トランスポートストリーム形式を示す拡張子とによって特定される通常形式のトランスポートストリームファイルを読み出し、
出力モードが立体視出力モードである場合、前記プレイリスト情報におけるファイル参照情報と、インターリーブされていることを示す拡張子とによって特定されるインターリーブされたトランスポートストリームファイルを読み出すようドライブ装置を制御する
ことを特徴とするシステムLSI。 - 前記システムLSIは、
ユーザ設定を示すレジスタと、
接続されている表示装置が立体視再生に対応しているか否かを示すケーパビリティレジスタとを備え、
前記制御手段は、複数の条件が成立するかどうかを判定して、複数の条件が成立した場合に、出力モードを立体視出力モードに設定し、
前記複数の条件のうち第1条件は、前記ユーザ設定を示すレジスタが、ユーザが立体視再生を希望する旨を示しているという条件であり、
前記複数の条件のうち第2条件は、前記接続されている表示装置が立体視再生に対応しているか否かを示すケーパビリティレジスタが、接続されている表示装置が立体視再生に対応していることを示しているという条件であり、
前記第1条件及び前記第2条件が成立する場合、前記出力モードを立体視出力モードに設定する
ことを特徴とする請求項6記載のシステムLSI。 - プレイリスト情報に従い、ビデオストリームを再生する処理をコンピュータ上で実行する再生方法であって、
前記プレイリスト情報におけるファイル参照情報で特定されるトランスポートストリームファイルを記録媒体から読み出す読出ステップと、
読み出されたトランスポートストリームファイルに格納されているビデオストリームに含まれる圧縮ピクチャデータの供給を受けてデコードを行い、非圧縮のピクチャデータを得るデコードステップと、
コンピュータにおけるモードレジスタに格納されている出力モードに従い、デコーダによって得られた非圧縮のピクチャデータを出力する出力ステップとを有し、
前記読出ステップは、
出力モードが平面視出力モードである場合、前記プレイリスト情報におけるファイル参照情報と、トランスポートストリーム形式を示す拡張子とによって特定される通常形式のトランスポートストリームファイルを読み出し、
出力モードが立体視出力モードである場合、前記プレイリスト情報におけるファイル参照情報と、インターリーブされていることを示す拡張子とによって特定されるインターリーブされたトランスポートストリームファイルを読み出す
ことを特徴とする再生方法。 - 前記コンピュータは、
ユーザ設定を示すレジスタと、
接続されている表示装置が立体視再生に対応しているか否かを示すケーパビリティレジスタとを有し、
前記再生方法は、
複数の条件が成立するかどうかを判定して、複数の条件が成立した場合に、出力モードを立体視出力モードに設定する設定ステップを有し、
前記複数の条件のうち第1条件は、前記ユーザ設定を示すレジスタが、ユーザが立体視再生を希望する旨を示しているという条件であり、
前記複数の条件のうち第2条件は、前記接続されている表示装置が立体視再生に対応しているか否かを示すケーパビリティレジスタが、接続されている表示装置が立体視再生に対応していることを示しているという条件であり、
前記第1条件及び前記第2条件が成立する場合、前記出力モードを立体視出力モードに設定する
ことを特徴とする請求項8記載の再生方法。 - プレイリスト情報と、ストリームファイルとが記録された記録媒体であって、
前記プレイリスト情報は、1つ以上の再生区間情報を含み、
前記再生区間情報は、
前記ストリームファイルを指定するファイル参照情報と、
再生可能なビデオストリームを示すストリーム許可テーブルと、
ベースビュー指定情報とを含み、
前記ビデオストリームには、立体視再生を可能とするレフトビュービデオストリーム、ライトビュービデオストリームがあり、
前記ベースビュー指定情報は、前記レフトビュービデオストリーム及び前記ライトビュービデオストリームのうちどちらが、単独で平面視再生を行うことができるベースビュービデオストリームであるかを示す
ことを特徴とする記録媒体。 - プレイリスト情報に従い、ビデオストリームを再生する再生装置であって、
前記プレイリスト情報におけるファイル参照情報で特定されるトランスポートストリームファイルを読み出す読出手段と、
読み出されたトランスポートストリームファイルに格納されているビデオストリームに含まれる圧縮ピクチャデータの供給を受けてデコードを行い、非圧縮のピクチャデータを得るデコーダと、
自機の出力モードを格納しているモードレジスタと、
前記モードレジスタに格納されている出力モードに従い、デコーダによって得られた非圧縮のピクチャデータを出力する出力手段とを備え、
前記ビデオストリームには、立体視再生を可能とするレフトビュービデオストリーム、ライトビュービデオストリームがあり、
前記ベースビュー指定情報は、前記レフトビュービデオストリーム及び前記ライトビュービデオストリームのうちどちらが、単独で平面視再生を行うことができるベースビュービデオストリームであるかを示し、
前記デコーダは、
平面視再生を行う場合、前記ベースビュー指定情報により示されるベースビュービデオストリームを構成するピクチャデータのデコードを行う
ことを特徴とする再生装置。 - プレイリスト情報におけるファイル参照情報で特定されるトランスポートストリームファイルを読み出すドライブと共に、再生装置に組込むことができる集積回路であって、
読み出されたトランスポートストリームファイルに格納されているビデオストリームに含まれる圧縮ピクチャデータの供給を受けてデコードを行い、非圧縮のピクチャデータを得るデコーダと、
自機の出力モードを格納しているモードレジスタと、
前記モードレジスタに格納されている出力モードに従い、デコーダによって得られた非圧縮のピクチャデータを出力するよう制御を行う制御手段とを備え、
前記ビデオストリームには、立体視再生を可能とするレフトビュービデオストリーム、ライトビュービデオストリームがあり、
前記ベースビュー指定情報は、前記レフトビュービデオストリーム及び前記ライトビュービデオストリームのうちどちらが、単独で平面視再生を行うことができるベースビュービデオストリームであるかを示し、
前記デコーダは、
平面視再生を行う場合、前記ベースビュー指定情報により示されるベースビュービデオストリームを構成するピクチャデータのデコードを行う
ことを特徴とする集積回路。 - プレイリスト情報に従い、ビデオストリームを再生する処理をコンピュータ上で実行する再生方法であって、
前記プレイリスト情報におけるファイル参照情報で特定されるトランスポートストリームファイルを読み出す読出ステップと、
読み出されたトランスポートストリームファイルに格納されているビデオストリームに含まれる圧縮ピクチャデータの供給を受けてデコードを行い、非圧縮のピクチャデータを得るデコードステップと、
コンピュータにおけるモードレジスタに格納されている出力モードに従い、デコーダによって得られた非圧縮のピクチャデータを出力する出力ステップとを有し、
前記ビデオストリームには、立体視再生を可能とするレフトビュービデオストリーム、ライトビュービデオストリームがあり、
前記ベースビュー指定情報は、前記レフトビュービデオストリーム及び前記ライトビュービデオストリームのうちどちらが、単独で平面視再生を行うことができるベースビュービデオストリームであるかを示し、
前記デコードステップは、
平面視再生を行う場合、前記ベースビュー指定情報により示されるベースビュービデオストリームを構成するピクチャデータのデコードを行う
ことを特徴とする再生方法。 - 立体視用眼鏡を着用したユーザに視聴させるための画像表示を、所定の表示期間内に実行する表示装置であって、
前記表示期間は、
ユーザが着用した眼鏡のレフトビューが透光状態になっていて、ライトビューが遮光状態になっている第1表示期間、ユーザが着用した眼鏡のライトビューが透光状態になっていて、レフトビューが遮光状態になっている第2表示期間、ユーザが着用した眼鏡のレフトビュー及びライトビューの双方が遮光状態になっている第3表示期間があり、
前記第3表示期間における表示内容は、眼鏡を着用していないユーザに対してのメッセージを含む
ことを特徴とする表示装置。 - 表示装置を視聴する際、ユーザが着用する眼鏡であって、
視聴対象となる表示装置の表示期間は、
レフトビューが表示される第1表示期間、ライトビューが表示される第2表示期間、眼鏡を装着していないユーザに対してのメッセージが表示される第3表示期間があり、
前記第3表示期間において、レフトビュー及びライトビューの状態を何れも、遮光状態に設定する
ことを特徴とする眼鏡。 - マルチチャネルの表示を行う表示装置を視聴する際、ユーザが着用する眼鏡であって、
マルチチャネルのうち、特定のチャネルの表示期間は、その特定チャネルに割り当てられたユーザに割り当てられた眼鏡の状態を透光状態に設定し、
特定チャネル以外の表示期間は、その特定チャネルに割り当てられたユーザに割り当てられた眼鏡の状態を遮光状態に設定する
ことを特徴とする眼鏡。 - 眼鏡をリモートで制御する表示装置であって、
マルチチャネルのうち、特定のチャネルの表示期間は、その特定チャネルに割り当てられたユーザに割り当てられた眼鏡の状態を透光状態に設定させ
特定チャネル以外の表示期間は、その特定チャネルに割り当てられたユーザに割り当てられた眼鏡の状態を遮光状態に設定させる ことを特徴とする表示装置。
Priority Applications (7)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN200980100249.7A CN101911713B (zh) | 2008-09-30 | 2009-09-14 | 再现装置、集成电路、再现方法、记录方法、记录介质再现系统 |
EP09805892.8A EP2355533A4 (en) | 2008-09-30 | 2009-09-14 | RECORDING MEDIUM, REPRODUCING DEVICE, SYSTEM LSI, REPRODUCTION METHOD, GLASSES, AND DISPLAY DEVICE ASSOCIATED WITH 3D VIDEO |
MX2010002097A MX2010002097A (es) | 2008-09-30 | 2009-09-14 | Medio de grabacion, dispositivo de reproduccion, integracion a gran escala de sistema, metodo de reproduccion, lentes y dispositivo visual para imagenes 3d. |
CA2691727A CA2691727C (en) | 2008-09-30 | 2009-09-14 | Recording medium, playback device, system lsi, playback method, glasses, and display device for 3d images |
BRPI0904620-8A BRPI0904620A2 (pt) | 2008-09-30 | 2009-09-14 | meio de gravação , dispositivo de reprodução, sistema lsi, método de reprodução,óculos, e dispositivo de exibição para imagens 3d |
RU2010107209/08A RU2502214C2 (ru) | 2008-09-30 | 2009-09-14 | Носитель записи, устройство воспроизведения, системная бис, способ воспроизведения, очки и устройство отображения для трехмерных изображений |
JP2010502593A JP4564107B2 (ja) | 2008-09-30 | 2009-09-14 | 記録媒体、再生装置、システムlsi、再生方法、記録方法、記録媒体再生システム |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10131608P | 2008-09-30 | 2008-09-30 | |
US61/101,316 | 2008-09-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2010038365A1 true WO2010038365A1 (ja) | 2010-04-08 |
Family
ID=42073150
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2009/004554 WO2010038365A1 (ja) | 2008-09-30 | 2009-09-14 | 3d映像に係る記録媒体、再生装置、システムlsi、再生方法、眼鏡、表示装置 |
Country Status (12)
Country | Link |
---|---|
US (3) | US8089507B2 (ja) |
EP (4) | EP2395772A3 (ja) |
JP (4) | JP4564107B2 (ja) |
KR (1) | KR20110074823A (ja) |
CN (2) | CN102355590B (ja) |
BR (1) | BRPI0904620A2 (ja) |
CA (1) | CA2691727C (ja) |
MX (1) | MX2010002097A (ja) |
MY (1) | MY151243A (ja) |
RU (1) | RU2502214C2 (ja) |
TW (1) | TW201031207A (ja) |
WO (1) | WO2010038365A1 (ja) |
Cited By (26)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010116958A1 (ja) * | 2009-04-08 | 2010-10-14 | ソニー株式会社 | 情報処理装置、情報処理方法、プログラム、および記録媒体 |
WO2010116957A1 (ja) * | 2009-04-08 | 2010-10-14 | ソニー株式会社 | 再生装置、再生方法、およびプログラム |
CN102280943A (zh) * | 2010-06-11 | 2011-12-14 | 深圳Tcl新技术有限公司 | 立体眼镜无线供电系统 |
CN102314914A (zh) * | 2010-07-08 | 2012-01-11 | 索尼公司 | 信息处理装置、信息处理方法和程序 |
WO2012011525A1 (ja) * | 2010-07-21 | 2012-01-26 | 株式会社プランネット・アソシエイツ | 三次元ビデオストリームへの映像変換方法 |
JP2012023488A (ja) * | 2010-07-13 | 2012-02-02 | Ntt Docomo Inc | 画像処理装置、画像処理方法、表示装置及びプログラム |
JP2012033243A (ja) * | 2010-08-02 | 2012-02-16 | Sony Corp | データ生成装置およびデータ生成方法、データ処理装置およびデータ処理方法 |
JP2012039336A (ja) * | 2010-08-06 | 2012-02-23 | Hitachi Consumer Electronics Co Ltd | 映像表示システム及び再生装置 |
CN102378022A (zh) * | 2010-08-06 | 2012-03-14 | 日立民用电子株式会社 | 影像显示系统及再现装置 |
WO2012053151A1 (ja) * | 2010-10-20 | 2012-04-26 | 日立コンシューマエレクトロニクス株式会社 | 再生装置及び再生方法 |
CN102469319A (zh) * | 2010-11-10 | 2012-05-23 | 康佳集团股份有限公司 | 三维菜单生成方法及三维显示装置 |
JP2012100181A (ja) * | 2010-11-05 | 2012-05-24 | Hitachi Consumer Electronics Co Ltd | 映像出力装置、映像出力方法、受信装置および受信方法 |
JP2012105229A (ja) * | 2010-11-15 | 2012-05-31 | Hitachi Consumer Electronics Co Ltd | 再生装置 |
WO2012123981A1 (ja) * | 2011-03-11 | 2012-09-20 | 日立コンシューマエレクトロニクス株式会社 | 記録装置/方法/媒体、再生装置/方法 |
WO2012123982A1 (ja) * | 2011-03-11 | 2012-09-20 | 日立コンシューマエレクトロニクス株式会社 | 記録装置/方法/媒体、再生装置/方法 |
JP2012527814A (ja) * | 2009-05-18 | 2012-11-08 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | 3dトリックプレイのためのエントリー・ポイント |
JP2013051618A (ja) * | 2011-08-31 | 2013-03-14 | Toshiba Corp | 映像処理装置および映像処理方法 |
JP2013055682A (ja) * | 2012-10-31 | 2013-03-21 | Toshiba Corp | 映像処理装置および映像処理方法 |
CN103262544A (zh) * | 2010-09-10 | 2013-08-21 | 青岛海信电器股份有限公司 | 3d电视界面的显示方法和装置 |
CN103262547A (zh) * | 2010-12-03 | 2013-08-21 | 皇家飞利浦电子股份有限公司 | 3d图像数据的转移 |
EP2577979A4 (en) * | 2010-06-01 | 2014-08-27 | Intel Corp | METHOD AND APPARATUS FOR ADAPTIVE STABLE IMAGE SYNCHRONIZATION IN STEREOSCOPIC 3D SYSTEMS |
US9088775B2 (en) | 2009-04-08 | 2015-07-21 | Sony Corporation | Recording device, recording method, reproduction device, reproduction method, recording medium, and program for encoding and decoding video data of a plurality of viewpoints |
TWI507032B (zh) * | 2012-06-21 | 2015-11-01 | Top Victory Invest Ltd | And a display method and a device for preventing the display of the abnormality of the screen display |
US9191662B2 (en) | 2010-06-24 | 2015-11-17 | Mstar Semiconductor, Inc. | Display apparatus and associated glasses |
WO2016039172A1 (ja) * | 2014-09-12 | 2016-03-17 | ソニー株式会社 | 再生装置、再生方法、情報処理装置、情報処理方法、プログラム、および記録媒体 |
US11259070B2 (en) | 2019-04-11 | 2022-02-22 | Kabushiki Kaisha Toshiba | Packet generation apparatus and method |
Families Citing this family (109)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2009006083A1 (en) * | 2007-06-29 | 2009-01-08 | Wms Gaming, Inc. | Initializing and authenticating wagering game machines |
KR100942142B1 (ko) * | 2007-10-11 | 2010-02-16 | 한국전자통신연구원 | 객체기반 오디오 콘텐츠 송수신 방법 및 그 장치 |
KR20100040640A (ko) | 2008-10-10 | 2010-04-20 | 엘지전자 주식회사 | 수신 시스템 및 데이터 처리 방법 |
KR20100040563A (ko) * | 2008-10-10 | 2010-04-20 | 삼성전자주식회사 | 방송 표시장치 및 그의 2차원 영상 표시 방법 |
EP2348747A4 (en) * | 2008-11-18 | 2013-08-28 | Panasonic Corp | REPRODUCTION DEVICE, INTEGRATED CIRCUIT, AND REPRODUCTION METHOD WHEREAS SPECIALIZED REPRODUCTION |
EP2368370A1 (en) * | 2008-11-24 | 2011-09-28 | Koninklijke Philips Electronics N.V. | 3d video reproduction matching the output format to the 3d processing ability of a display |
WO2010064472A1 (ja) * | 2008-12-01 | 2010-06-10 | シャープ株式会社 | コンテンツ再生装置、再生方法、プログラム及び記録媒体 |
US20100186234A1 (en) * | 2009-01-28 | 2010-07-29 | Yehuda Binder | Electric shaver with imaging capability |
US20100177162A1 (en) * | 2009-01-15 | 2010-07-15 | Charles Macfarlane | Method and system for enabling 3d video and image processing using one full resolution video stream and one lower resolution video stream |
JP5573683B2 (ja) * | 2009-01-22 | 2014-08-20 | 日本電気株式会社 | 立体映像鑑賞システム、表示システム、光シャッタおよび立体映像鑑賞方法 |
EP2382793A4 (en) | 2009-01-28 | 2014-01-15 | Lg Electronics Inc | BROADCAST RECEIVER AND METHOD FOR PROCESSING VIDEO DATA THEREFOR |
JP5465523B2 (ja) * | 2009-01-29 | 2014-04-09 | 三洋電機株式会社 | 立体画像表示システム |
US8284236B2 (en) * | 2009-02-19 | 2012-10-09 | Sony Corporation | Preventing interference between primary and secondary content in a stereoscopic display |
US8723927B2 (en) * | 2009-03-31 | 2014-05-13 | Daniel Rosen | Subtitling stereographic imagery |
JP4915457B2 (ja) | 2009-04-03 | 2012-04-11 | ソニー株式会社 | 情報処理装置、情報処理方法、及び、プログラム |
JP2010245761A (ja) * | 2009-04-03 | 2010-10-28 | Sony Corp | 情報処理装置、情報処理方法、及び、プログラム |
JP4984194B2 (ja) * | 2009-04-08 | 2012-07-25 | ソニー株式会社 | 記録方法 |
JP5407500B2 (ja) * | 2009-04-08 | 2014-02-05 | ソニー株式会社 | 情報処理装置、および情報処理方法、並びにプログラム |
JP4985807B2 (ja) * | 2009-04-15 | 2012-07-25 | ソニー株式会社 | 再生装置および再生方法 |
JP5296195B2 (ja) * | 2009-04-16 | 2013-09-25 | 株式会社東芝 | コンテンツデータ再生システム、及び記録装置 |
US8730303B2 (en) * | 2009-04-27 | 2014-05-20 | Lg Electronics Inc. | Broadcast transmitter, broadcast receiver and 3D video data processing method thereof |
DE102009019051B4 (de) * | 2009-04-28 | 2011-07-07 | Giesecke & Devrient GmbH, 81677 | Speichermedium mit Verschlüsselungseinrichtung |
JP5400467B2 (ja) * | 2009-05-01 | 2014-01-29 | キヤノン株式会社 | 映像出力装置、その制御方法、及びプログラム |
JP2011041249A (ja) | 2009-05-12 | 2011-02-24 | Sony Corp | データ構造および記録媒体、並びに、再生装置、再生方法、プログラム、およびプログラム格納媒体 |
RU2538307C2 (ru) * | 2009-05-12 | 2015-01-10 | Сони Корпорейшн | Структура данных и носитель данных, воспроизводящее устройство, способ воспроизведения, программа и носитель для хранения программы |
US20110013888A1 (en) * | 2009-06-18 | 2011-01-20 | Taiji Sasaki | Information recording medium and playback device for playing back 3d images |
WO2010150961A1 (ko) * | 2009-06-23 | 2010-12-29 | (주)엘지전자 | 3차원 영상 제공 장치, 표시 장치 및 그 방법 |
JP5273478B2 (ja) * | 2009-07-07 | 2013-08-28 | ソニー株式会社 | 映像表示装置および映像表示システム |
CA2770861A1 (en) * | 2009-08-12 | 2011-02-17 | Lg Electronics Inc. | Method for diagnosing 3d state information, and broadcast receiver |
JP2011082666A (ja) * | 2009-10-05 | 2011-04-21 | Sony Corp | 信号伝送方法、信号送信装置及び信号受信装置 |
JP2011082683A (ja) * | 2009-10-05 | 2011-04-21 | Sony Corp | 画像処理装置、画像処理方法、及び、プログラム |
US20110080948A1 (en) * | 2009-10-05 | 2011-04-07 | Xuemin Chen | Method and system for 3d video decoding using a tier system framework |
JP5588144B2 (ja) * | 2009-10-14 | 2014-09-10 | パナソニック株式会社 | 映像信号処理装置及び映像信号処理方法 |
JP2011107589A (ja) * | 2009-11-20 | 2011-06-02 | Sony Corp | 立体表示装置 |
US9414041B2 (en) * | 2009-11-23 | 2016-08-09 | Samsung Electronics Co., Ltd. | Method for changing play mode, method for changing display mode, and display apparatus and 3D image providing system using the same |
US20110126160A1 (en) * | 2009-11-23 | 2011-05-26 | Samsung Electronics Co., Ltd. | Method of providing 3d image and 3d display apparatus using the same |
JP4723682B2 (ja) * | 2009-11-30 | 2011-07-13 | シャープ株式会社 | 表示制御装置、表示制御方法、表示制御プログラム、コンピュータ読み取り可能な記録媒体、上記表示制御装置を備えた記録再生装置、音声出力装置、及び音声出力装置を備えた記録再生装置 |
EP2334088A1 (en) * | 2009-12-14 | 2011-06-15 | Koninklijke Philips Electronics N.V. | Generating a 3D video signal |
US20110157302A1 (en) * | 2009-12-30 | 2011-06-30 | Ati Technologies Ulc | Three-dimensional video display system with multi-stream sending/receiving operation |
US9247286B2 (en) | 2009-12-31 | 2016-01-26 | Broadcom Corporation | Frame formatting supporting mixed two and three dimensional video data communication |
US8823782B2 (en) | 2009-12-31 | 2014-09-02 | Broadcom Corporation | Remote control with integrated position, viewer identification and optical and audio test |
US20110157322A1 (en) | 2009-12-31 | 2011-06-30 | Broadcom Corporation | Controlling a pixel array to support an adaptable light manipulator |
US8854531B2 (en) | 2009-12-31 | 2014-10-07 | Broadcom Corporation | Multiple remote controllers that each simultaneously controls a different visual presentation of a 2D/3D display |
JP2011146828A (ja) * | 2010-01-13 | 2011-07-28 | Sony Corp | データ構造、記録装置および記録方法、再生装置および再生方法、並びにプログラム |
US9179138B2 (en) * | 2010-02-10 | 2015-11-03 | Lg Electronics Inc. | Image display method and apparatus |
JP2013080987A (ja) * | 2010-02-15 | 2013-05-02 | Panasonic Corp | 立体映像表示装置 |
KR101289269B1 (ko) * | 2010-03-23 | 2013-07-24 | 한국전자통신연구원 | 영상 시스템에서 영상 디스플레이 장치 및 방법 |
JP2011216965A (ja) * | 2010-03-31 | 2011-10-27 | Sony Corp | 情報処理装置、情報処理方法、再生装置、再生方法、およびプログラム |
US10448083B2 (en) | 2010-04-06 | 2019-10-15 | Comcast Cable Communications, Llc | Streaming and rendering of 3-dimensional video |
US11711592B2 (en) | 2010-04-06 | 2023-07-25 | Comcast Cable Communications, Llc | Distribution of multiple signals of video content independently over a network |
JP2011223248A (ja) * | 2010-04-08 | 2011-11-04 | Sony Corp | 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム |
JP5577805B2 (ja) * | 2010-04-08 | 2014-08-27 | ソニー株式会社 | 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム |
JP5494152B2 (ja) * | 2010-04-08 | 2014-05-14 | ソニー株式会社 | 情報処理装置、情報記録媒体、および情報処理方法、並びにプログラム |
KR101435594B1 (ko) * | 2010-05-31 | 2014-08-29 | 삼성전자주식회사 | 디스플레이 장치 및 그 디스플레이 방법 |
JP2011254239A (ja) * | 2010-06-01 | 2011-12-15 | Sony Corp | 記録装置、再生装置、記録再生システム、記録媒体およびプログラム |
US8688679B2 (en) * | 2010-07-20 | 2014-04-01 | Smartek21, Llc | Computer-implemented system and method for providing searchable online media content |
KR20120017649A (ko) * | 2010-08-19 | 2012-02-29 | 삼성전자주식회사 | 디스플레이장치 및 그 제어방법 |
KR20120020477A (ko) * | 2010-08-30 | 2012-03-08 | 삼성전자주식회사 | 입체영상표시장치 및 그 구동 방법 |
EP2437513A1 (en) | 2010-10-01 | 2012-04-04 | Hitachi Consumer Electronics Co. Ltd. | Receiver for 2D and 3D video content |
JP5730524B2 (ja) * | 2010-10-01 | 2015-06-10 | 日立マクセル株式会社 | 受信装置、および、受信方法 |
JP2012080309A (ja) * | 2010-10-01 | 2012-04-19 | Hitachi Consumer Electronics Co Ltd | コンテンツ受信機 |
CN103202023A (zh) * | 2010-10-25 | 2013-07-10 | 松下电器产业株式会社 | 编码方法、显示装置、解码方法 |
US8610759B2 (en) * | 2010-10-26 | 2013-12-17 | Verizon Patent And Licensing Inc. | Methods and systems for presenting adjunct content during a presentation of a media content instance |
US20120120050A1 (en) * | 2010-11-12 | 2012-05-17 | Nokia Corporation | Dual-mode two-dimensional/three-dimensional display |
US10424274B2 (en) * | 2010-11-24 | 2019-09-24 | Ati Technologies Ulc | Method and apparatus for providing temporal image processing using multi-stream field information |
KR20120058702A (ko) * | 2010-11-27 | 2012-06-08 | 전자부품연구원 | 디지털 방송에서 서비스 호환 방식 전송 방법 |
US9584798B2 (en) | 2010-12-09 | 2017-02-28 | Google Technology Holdings LLC | Method and apparatus for managing 3D video content |
FR2970613B1 (fr) * | 2011-01-13 | 2013-01-18 | Alcatel Lucent | Procede de fourniture a un observateur de donnees relatives a au moins un utilisateur d'un operateur de telecommunication ou de services internet dans un reseau |
US9204123B2 (en) * | 2011-01-14 | 2015-12-01 | Comcast Cable Communications, Llc | Video content generation |
JP2012178783A (ja) * | 2011-02-28 | 2012-09-13 | Sony Corp | 画像表示システム、表示装置、並びにシャッター眼鏡 |
WO2012127837A1 (ja) * | 2011-03-18 | 2012-09-27 | パナソニック株式会社 | 表示装置、3d眼鏡、及び3d映像視聴システム |
JP5452537B2 (ja) * | 2011-04-04 | 2014-03-26 | 日立コンシューマエレクトロニクス株式会社 | 映像表示システム及び表示装置 |
US8681170B2 (en) * | 2011-05-05 | 2014-03-25 | Ati Technologies Ulc | Apparatus and method for multi-streaming for more than three pixel component values |
WO2012160812A1 (ja) * | 2011-05-25 | 2012-11-29 | パナソニック株式会社 | 映像処理装置、送信装置、立体映像視聴システム、映像処理方法、映像処理プログラム及び集積回路 |
JP5679578B2 (ja) * | 2011-08-05 | 2015-03-04 | 株式会社ソニー・コンピュータエンタテインメント | 画像処理装置 |
WO2013021767A1 (ja) * | 2011-08-10 | 2013-02-14 | 富士フイルム株式会社 | 動体検出装置及び方法 |
US9185398B2 (en) | 2011-09-22 | 2015-11-10 | Google Technology Holdings LLC | Method and apparatus for providing three-dimensional content |
JP2013126048A (ja) * | 2011-12-13 | 2013-06-24 | Sony Corp | 送信装置、送信方法、受信装置および受信方法 |
JP4984193B2 (ja) * | 2012-02-01 | 2012-07-25 | ソニー株式会社 | 再生装置、再生方法、および記録方法 |
US20130243079A1 (en) * | 2012-03-19 | 2013-09-19 | Nokia Siemens Networks Oy | Storage and processing savings when adapting video bit rate to link speed |
KR20130108198A (ko) * | 2012-03-23 | 2013-10-02 | (주)휴맥스 | Mmt 패키지화된 svc 비디오 콘텐츠의 하이브리드 전송 방법 및 수신 방법 |
EP2822282B1 (en) * | 2012-04-23 | 2018-11-07 | LG Electronics Inc. | Signal processing device and method for 3d service |
CN102780896A (zh) * | 2012-05-31 | 2012-11-14 | 新奥特(北京)视频技术有限公司 | 一种流媒体素材支持3d技术的方法 |
US20130342689A1 (en) * | 2012-06-25 | 2013-12-26 | Intel Corporation | Video analytics test system |
CN103688535B (zh) | 2012-07-19 | 2017-02-22 | 太阳专利托管公司 | 图像编码方法、图像解码方法、图像编码装置及图像解码装置 |
US20140327740A1 (en) * | 2012-09-27 | 2014-11-06 | Sony Corporation | Transmission apparatus, transmisson method, receiver and receiving method |
US20140099065A1 (en) * | 2012-10-04 | 2014-04-10 | Samsung Electronics Co., Ltd. | Apparatus for reproducing recording medium and method thereof |
US9860515B2 (en) * | 2012-12-11 | 2018-01-02 | Electronics And Telecommunications Research Institute | Apparatus and method for 3D content broadcasting with boundary information |
US8872983B2 (en) * | 2012-12-27 | 2014-10-28 | Kabushiki Kaisha Toshiba | Information processing apparatus and display processing method |
US20140212115A1 (en) * | 2013-01-31 | 2014-07-31 | Hewlett Packard Development Company, L.P. | Optical disc with three-dimensional viewing depth |
JP6450064B2 (ja) * | 2013-03-18 | 2019-01-09 | 任天堂株式会社 | 情報処理装置、動画データのデータ構造、情報処理システム、動画再生プログラム、および、動画の再生方法。 |
KR20140120156A (ko) * | 2013-04-02 | 2014-10-13 | 삼성전자주식회사 | 사용성이 향상된 모바일 디바이스를 위한 3차원 그래픽 데이터 생성 방법 및 이를 이용한 응용 개발 환경 |
JP6353896B2 (ja) | 2013-04-08 | 2018-07-04 | ドルビー・インターナショナル・アーベー | ルックアップ・テーブルを符号化する方法および復号する方法、並びに対応する装置 |
WO2015012835A1 (en) * | 2013-07-25 | 2015-01-29 | Empire Technology Development, Llc | Composite display with mutliple imaging properties |
TW201506445A (zh) * | 2013-08-01 | 2015-02-16 | Hon Hai Prec Ind Co Ltd | 顯示系統及顯示系統的顯示方法 |
TWI587706B (zh) * | 2014-06-04 | 2017-06-11 | 佳世達科技股份有限公司 | 顯示器及其控制方法 |
CN111212251B (zh) * | 2014-09-10 | 2022-05-27 | 松下电器(美国)知识产权公司 | 再现装置以及再现方法 |
US9866757B2 (en) | 2014-10-27 | 2018-01-09 | Wichita State University | Computer program, system, and method for observation and communication for mobile settings, mobile applications, and wearable mobile devices |
CN105573491B (zh) * | 2015-11-13 | 2018-09-07 | 广东顺德中山大学卡内基梅隆大学国际联合研究院 | 裸眼3d私隐式虚拟人机交互系统及方法 |
CN106980579B (zh) | 2016-09-30 | 2020-08-14 | 阿里巴巴集团控股有限公司 | 一种图片加载方法及装置 |
CN108235144B (zh) * | 2016-12-22 | 2021-02-19 | 阿里巴巴(中国)有限公司 | 播放内容获取方法、装置及计算设备 |
CN106791795A (zh) * | 2017-01-03 | 2017-05-31 | 京东方科技集团股份有限公司 | 3d视频流的发送方法及发送装置、接收方法及接收装置 |
WO2019043989A1 (ja) * | 2017-08-30 | 2019-03-07 | パナソニックIpマネジメント株式会社 | 記録方法、及び、記録装置 |
CN107680556B (zh) * | 2017-11-03 | 2019-08-02 | 深圳市华星光电半导体显示技术有限公司 | 一种显示器节能方法、装置及显示器 |
RU2020122086A (ru) | 2018-01-12 | 2022-01-04 | Сони Корпорейшн | Способ и устройство обработки информации |
KR102445544B1 (ko) | 2018-04-02 | 2022-09-21 | 삼성전자 주식회사 | 전자장치, 그 제어방법 및 기록매체 |
US11100119B2 (en) * | 2018-05-04 | 2021-08-24 | Sap Se | Determining data structures for spatial data based on spatial data statistics |
CN110866793A (zh) * | 2018-08-27 | 2020-03-06 | 阿里健康信息技术有限公司 | 虚拟对象展示、生成和提供方法 |
CN110427363B (zh) * | 2019-01-11 | 2023-04-11 | 中国铁路西安局集团有限公司 | 基于关系型数据库的铁路信号设备图像信息转换存储方法 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1997032437A1 (en) * | 1996-02-28 | 1997-09-04 | Matsushita Electric Industrial Co., Ltd. | High-resolution optical disk for recording stereoscopic video, optical disk reproducing device, and optical disk recording device |
JP2004240469A (ja) * | 2002-12-13 | 2004-08-26 | Sharp Corp | 画像データ作成装置およびそのデータを再生する画像データ再生装置 |
JP2005227424A (ja) * | 2004-02-12 | 2005-08-25 | Nakagawa Kenkyusho:Kk | 表示システム |
Family Cites Families (59)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5767898A (en) * | 1994-06-23 | 1998-06-16 | Sanyo Electric Co., Ltd. | Three-dimensional image coding by merger of left and right images |
US5661518A (en) * | 1994-11-03 | 1997-08-26 | Synthonics Incorporated | Methods and apparatus for the creation and transmission of 3-dimensional images |
US5612735A (en) * | 1995-05-26 | 1997-03-18 | Luncent Technologies Inc. | Digital 3D/stereoscopic video compression technique utilizing two disparity estimates |
US5619256A (en) * | 1995-05-26 | 1997-04-08 | Lucent Technologies Inc. | Digital 3D/stereoscopic video compression technique utilizing disparity and motion compensated predictions |
TW436777B (en) | 1995-09-29 | 2001-05-28 | Matsushita Electric Ind Co Ltd | A method and an apparatus for reproducing bitstream having non-sequential system clock data seamlessly therebetween |
US6502199B2 (en) | 1995-09-29 | 2002-12-31 | Matsushita Electric Industrial Co., Ltd. | Method and an apparatus for reproducing bitstream having non-sequential system clock data seamlessly therebetween |
US6111596A (en) * | 1995-12-29 | 2000-08-29 | Lucent Technologies Inc. | Gain and offset correction for efficient stereoscopic coding and improved display |
US5886736A (en) * | 1996-10-24 | 1999-03-23 | General Instrument Corporation | Synchronization of a stereoscopic video sequence |
EP2175659B1 (en) * | 1996-12-04 | 2012-11-14 | Panasonic Corporation | Optical disk for high resolution and three-dimensional video recording, optical disk reproduction apparatus, and optical disk recording apparatus |
US6188442B1 (en) * | 1997-08-01 | 2001-02-13 | International Business Machines Corporation | Multiviewer display system for television monitors |
JP3931392B2 (ja) * | 1997-08-25 | 2007-06-13 | ソニー株式会社 | 立体画像用ビデオ信号生成装置、立体画像用ビデオ信号送出装置および立体画像用ビデオ信号受信装置 |
DE69841532D1 (de) | 1997-08-29 | 2010-04-15 | Panasonic Corp | Optische Platte mit hierarchisch codiertem digitalen Videosignal, Wiedergabevorrichtung und Aufnahmevorrichtung für die optische Platte |
US6043838A (en) * | 1997-11-07 | 2000-03-28 | General Instrument Corporation | View offset estimation for stereoscopic video coding |
US5963371A (en) * | 1998-02-04 | 1999-10-05 | Intel Corporation | Method of displaying private data to collocated users |
JP2001045524A (ja) * | 1999-01-26 | 2001-02-16 | Denso Corp | 立体表示装置 |
US6448952B1 (en) | 1999-01-26 | 2002-09-10 | Denso Corporation | Stereoscopic image display device |
US6795863B1 (en) * | 1999-08-10 | 2004-09-21 | Intline.Com, Inc. | System, device and method for combining streaming video with e-mail |
GB0007868D0 (en) | 2000-03-31 | 2000-05-17 | Koninkl Philips Electronics Nv | Methods and apparatus for editing digital video recordings and recordings made by such methods |
US6859840B2 (en) * | 2001-01-29 | 2005-02-22 | Kasenna, Inc. | Prefix caching for media objects |
JPWO2003092303A1 (ja) * | 2002-04-25 | 2005-09-08 | シャープ株式会社 | マルチメディア情報生成装置およびマルチメディア情報再生装置 |
JP2004128769A (ja) | 2002-10-01 | 2004-04-22 | Pioneer Electronic Corp | 情報記録媒体、情報記録装置及び方法、情報再生装置及び方法、情報記録再生装置及び方法、記録又は再生制御用のコンピュータプログラム、並びに制御信号を含むデータ構造 |
KR100556826B1 (ko) * | 2003-04-17 | 2006-03-10 | 한국전자통신연구원 | Mpeg-4 기반의 양안식 3차원 동영상을 서비스하기 위한 인터넷 방송 시스템 및 그 방법 |
EP2259583B1 (en) | 2003-07-03 | 2012-05-30 | Panasonic Corporation | Reproduction apparatus, reproduction method, recording medium, recording apparatus and recording method. |
KR100608051B1 (ko) | 2003-07-07 | 2006-08-02 | 삼성전자주식회사 | 멀티앵글 데이터를 기록한 정보저장매체, 그 기록방법 및재생장치 |
RU2301461C2 (ru) * | 2003-07-07 | 2007-06-20 | Самсунг Электроникс Ко., Лтд. | Носитель информации, хранящий многоракурсные данные и способ записи и устройство воспроизведения для него |
KR20050049924A (ko) | 2003-11-24 | 2005-05-27 | 엘지전자 주식회사 | 고밀도 광디스크의 플레이리스트 구성방법, 관리방법 및재생방법과 기록재생장치 |
KR20050066264A (ko) * | 2003-12-26 | 2005-06-30 | 엘지전자 주식회사 | 고밀도 광디스크의 메뉴 구성방법 및 실행방법과기록재생장치 |
WO2005112476A1 (en) * | 2004-03-08 | 2005-11-24 | Robert Gibson | Multi-play theatre |
US20050206751A1 (en) * | 2004-03-19 | 2005-09-22 | East Kodak Company | Digital video system for assembling video sequences |
RU2378722C2 (ru) * | 2004-03-26 | 2010-01-10 | ЭлДжи ЭЛЕКТРОНИКС ИНК. | Носитель записи, способ и устройство для воспроизведения потоков текстовых субтитров |
JP2005286366A (ja) | 2004-03-26 | 2005-10-13 | Matsushita Electric Ind Co Ltd | ファイル分割結合方法、並びに装置 |
EP1737228B1 (en) * | 2004-04-07 | 2011-11-23 | Panasonic Corporation | Information recording apparatus and information converting method |
JP4617773B2 (ja) | 2004-08-23 | 2011-01-26 | ソニー株式会社 | 記録装置および方法、再生装置および方法、記録媒体、並びにプログラム |
US7609947B2 (en) * | 2004-09-10 | 2009-10-27 | Panasonic Corporation | Method and apparatus for coordinating playback from multiple video sources |
JP2006186768A (ja) * | 2004-12-28 | 2006-07-13 | Seiko Epson Corp | 画像表示装置及び方法 |
JP4676493B2 (ja) * | 2005-04-07 | 2011-04-27 | パナソニック株式会社 | 記録媒体、再生装置、記録方法 |
KR100813961B1 (ko) * | 2005-06-14 | 2008-03-14 | 삼성전자주식회사 | 영상 수신장치 |
JP4630150B2 (ja) * | 2005-07-26 | 2011-02-09 | シャープ株式会社 | 立体画像記録装置及びプログラム |
KR100728009B1 (ko) * | 2005-08-22 | 2007-06-13 | 삼성전자주식회사 | 다시점 동영상을 부호화하는 방법 및 장치 |
JP2007074578A (ja) | 2005-09-08 | 2007-03-22 | Casio Comput Co Ltd | 画像処理装置、撮影装置、及びプログラム |
KR100667823B1 (ko) * | 2005-10-13 | 2007-01-11 | 삼성전자주식회사 | 멀티 채널 영상 시스템 |
MY159176A (en) * | 2005-10-19 | 2016-12-30 | Thomson Licensing | Multi-view video coding using scalable video coding |
WO2007067020A1 (en) * | 2005-12-09 | 2007-06-14 | Electronics And Telecommunications Research Institute | System and method for transmitting/receiving three dimensional video based on digital broadcasting |
KR100747598B1 (ko) * | 2005-12-09 | 2007-08-08 | 한국전자통신연구원 | 디지털방송 기반의 3차원 입체영상 송수신 시스템 및 그방법 |
US8466954B2 (en) * | 2006-04-03 | 2013-06-18 | Sony Computer Entertainment Inc. | Screen sharing method and apparatus |
JP4857895B2 (ja) | 2006-05-10 | 2012-01-18 | ソニー株式会社 | 情報処理装置、および情報処理方法、並びにコンピュータ・プログラム |
US8305488B2 (en) * | 2006-05-10 | 2012-11-06 | Universal City Studios Llc | Time-sliced multiplexed image display |
KR100716142B1 (ko) * | 2006-09-04 | 2007-05-11 | 주식회사 이시티 | 스테레오스코픽 영상 데이터의 전송 방법 |
JP2010505174A (ja) * | 2006-09-28 | 2010-02-18 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | メニューディスプレイ |
WO2008054100A1 (en) | 2006-11-01 | 2008-05-08 | Electronics And Telecommunications Research Institute | Method and apparatus for decoding metadata used for playing stereoscopic contents |
KR101362941B1 (ko) * | 2006-11-01 | 2014-02-17 | 한국전자통신연구원 | 스테레오스코픽 콘텐츠 재생에 이용되는 메타 데이터의복호화 방법 및 장치 |
DE102006055641B4 (de) | 2006-11-22 | 2013-01-31 | Visumotion Gmbh | Anordnung und Verfahren zur Aufnahme und Wiedergabe von Bildern einer Szene und/oder eines Objektes |
US8717348B2 (en) * | 2006-12-22 | 2014-05-06 | Texas Instruments Incorporated | System and method for synchronizing a viewing device |
JP4289403B2 (ja) * | 2007-02-02 | 2009-07-01 | ソニー株式会社 | 編集装置及び編集方法 |
US8565584B2 (en) | 2007-02-02 | 2013-10-22 | Sony Corporation | Editing apparatus and editing method |
WO2008114595A1 (ja) * | 2007-03-22 | 2008-09-25 | Mitsubishi Electric Corporation | 映像再生装置および映像再生方法 |
US8400497B2 (en) * | 2007-09-07 | 2013-03-19 | Samsung Electronics Co., Ltd | Method and apparatus for generating stereoscopic file |
US8384764B2 (en) * | 2007-12-20 | 2013-02-26 | Samsung Electronics Co., Ltd. | Method and apparatus for generating multiview image data stream and method and apparatus for decoding the same |
KR101506217B1 (ko) * | 2008-01-31 | 2015-03-26 | 삼성전자주식회사 | 스테레오스코픽 영상의 부분 데이터 구간 재생을 위한스테레오스코픽 영상 데이터스트림 생성 방법과 장치, 및스테레오스코픽 영상의 부분 데이터 구간 재생 방법과 장치 |
-
2009
- 2009-09-14 EP EP11179910.2A patent/EP2395772A3/en not_active Withdrawn
- 2009-09-14 EP EP09805892.8A patent/EP2355533A4/en not_active Withdrawn
- 2009-09-14 KR KR1020107004132A patent/KR20110074823A/ko not_active Application Discontinuation
- 2009-09-14 MY MYPI20100401 patent/MY151243A/en unknown
- 2009-09-14 WO PCT/JP2009/004554 patent/WO2010038365A1/ja active Search and Examination
- 2009-09-14 BR BRPI0904620-8A patent/BRPI0904620A2/pt not_active IP Right Cessation
- 2009-09-14 MX MX2010002097A patent/MX2010002097A/es active IP Right Grant
- 2009-09-14 EP EP11179911.0A patent/EP2395770A3/en not_active Withdrawn
- 2009-09-14 RU RU2010107209/08A patent/RU2502214C2/ru not_active IP Right Cessation
- 2009-09-14 EP EP12165028.7A patent/EP2482564A3/en not_active Withdrawn
- 2009-09-14 JP JP2010502593A patent/JP4564107B2/ja active Active
- 2009-09-14 CN CN201110351525.6A patent/CN102355590B/zh not_active Expired - Fee Related
- 2009-09-14 CA CA2691727A patent/CA2691727C/en active Active
- 2009-09-14 CN CN200980100249.7A patent/CN101911713B/zh active Active
- 2009-09-18 TW TW098131549A patent/TW201031207A/zh unknown
- 2009-09-25 US US12/566,989 patent/US8089507B2/en active Active
-
2010
- 2010-02-25 JP JP2010040673A patent/JP4733772B2/ja active Active
- 2010-02-25 JP JP2010040674A patent/JP2010213267A/ja active Pending
-
2011
- 2011-09-30 JP JP2011217650A patent/JP4923162B2/ja not_active Expired - Fee Related
- 2011-11-22 US US13/302,518 patent/US20120062711A1/en not_active Abandoned
-
2012
- 2012-07-10 US US13/545,284 patent/US9344703B2/en not_active Expired - Fee Related
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1997032437A1 (en) * | 1996-02-28 | 1997-09-04 | Matsushita Electric Industrial Co., Ltd. | High-resolution optical disk for recording stereoscopic video, optical disk reproducing device, and optical disk recording device |
JP2004240469A (ja) * | 2002-12-13 | 2004-08-26 | Sharp Corp | 画像データ作成装置およびそのデータを再生する画像データ再生装置 |
JP2005227424A (ja) * | 2004-02-12 | 2005-08-25 | Nakagawa Kenkyusho:Kk | 表示システム |
Non-Patent Citations (1)
Title |
---|
See also references of EP2355533A4 * |
Cited By (37)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010116957A1 (ja) * | 2009-04-08 | 2010-10-14 | ソニー株式会社 | 再生装置、再生方法、およびプログラム |
US8599241B2 (en) | 2009-04-08 | 2013-12-03 | Sony Corporation | Information processing apparatus, information processing method, program, and recording medium |
US9049427B2 (en) | 2009-04-08 | 2015-06-02 | Sony Corporation | Playback device, playback method, and program for identifying a stream |
US9088775B2 (en) | 2009-04-08 | 2015-07-21 | Sony Corporation | Recording device, recording method, reproduction device, reproduction method, recording medium, and program for encoding and decoding video data of a plurality of viewpoints |
WO2010116958A1 (ja) * | 2009-04-08 | 2010-10-14 | ソニー株式会社 | 情報処理装置、情報処理方法、プログラム、および記録媒体 |
JP2012527814A (ja) * | 2009-05-18 | 2012-11-08 | コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ | 3dトリックプレイのためのエントリー・ポイント |
EP2577979A4 (en) * | 2010-06-01 | 2014-08-27 | Intel Corp | METHOD AND APPARATUS FOR ADAPTIVE STABLE IMAGE SYNCHRONIZATION IN STEREOSCOPIC 3D SYSTEMS |
KR101615573B1 (ko) | 2010-06-01 | 2016-04-26 | 인텔 코포레이션 | 입체 영상 3d 시스템의 적응형 안정한 이미지 타이밍을 위한 방법 및 장치 |
CN102280943A (zh) * | 2010-06-11 | 2011-12-14 | 深圳Tcl新技术有限公司 | 立体眼镜无线供电系统 |
US9191662B2 (en) | 2010-06-24 | 2015-11-17 | Mstar Semiconductor, Inc. | Display apparatus and associated glasses |
CN102314914A (zh) * | 2010-07-08 | 2012-01-11 | 索尼公司 | 信息处理装置、信息处理方法和程序 |
JP2012018727A (ja) * | 2010-07-08 | 2012-01-26 | Sony Corp | 情報処理装置、および情報処理方法、並びにプログラム |
JP2012023488A (ja) * | 2010-07-13 | 2012-02-02 | Ntt Docomo Inc | 画像処理装置、画像処理方法、表示装置及びプログラム |
WO2012011525A1 (ja) * | 2010-07-21 | 2012-01-26 | 株式会社プランネット・アソシエイツ | 三次元ビデオストリームへの映像変換方法 |
JP2012033243A (ja) * | 2010-08-02 | 2012-02-16 | Sony Corp | データ生成装置およびデータ生成方法、データ処理装置およびデータ処理方法 |
CN102378021A (zh) * | 2010-08-06 | 2012-03-14 | 日立民用电子株式会社 | 影像显示系统及再现装置 |
JP2012039336A (ja) * | 2010-08-06 | 2012-02-23 | Hitachi Consumer Electronics Co Ltd | 映像表示システム及び再生装置 |
CN102378022A (zh) * | 2010-08-06 | 2012-03-14 | 日立民用电子株式会社 | 影像显示系统及再现装置 |
CN103262544A (zh) * | 2010-09-10 | 2013-08-21 | 青岛海信电器股份有限公司 | 3d电视界面的显示方法和装置 |
WO2012053151A1 (ja) * | 2010-10-20 | 2012-04-26 | 日立コンシューマエレクトロニクス株式会社 | 再生装置及び再生方法 |
JP2012090074A (ja) * | 2010-10-20 | 2012-05-10 | Hitachi Consumer Electronics Co Ltd | 再生装置及び再生方法 |
CN102572467A (zh) * | 2010-11-05 | 2012-07-11 | 日立民用电子株式会社 | 影像输出装置、影像输出方法、接收装置和接收方法 |
JP2012100181A (ja) * | 2010-11-05 | 2012-05-24 | Hitachi Consumer Electronics Co Ltd | 映像出力装置、映像出力方法、受信装置および受信方法 |
CN102469319A (zh) * | 2010-11-10 | 2012-05-23 | 康佳集团股份有限公司 | 三维菜单生成方法及三维显示装置 |
JP2012105229A (ja) * | 2010-11-15 | 2012-05-31 | Hitachi Consumer Electronics Co Ltd | 再生装置 |
CN103262547B (zh) * | 2010-12-03 | 2016-01-20 | 皇家飞利浦电子股份有限公司 | 3d图像数据的转移 |
CN103262547A (zh) * | 2010-12-03 | 2013-08-21 | 皇家飞利浦电子股份有限公司 | 3d图像数据的转移 |
JPWO2012123982A1 (ja) * | 2011-03-11 | 2014-07-17 | 日立コンシューマエレクトロニクス株式会社 | 記録装置/方法/媒体、再生装置/方法 |
WO2012123982A1 (ja) * | 2011-03-11 | 2012-09-20 | 日立コンシューマエレクトロニクス株式会社 | 記録装置/方法/媒体、再生装置/方法 |
WO2012123981A1 (ja) * | 2011-03-11 | 2012-09-20 | 日立コンシューマエレクトロニクス株式会社 | 記録装置/方法/媒体、再生装置/方法 |
JP2013051618A (ja) * | 2011-08-31 | 2013-03-14 | Toshiba Corp | 映像処理装置および映像処理方法 |
TWI507032B (zh) * | 2012-06-21 | 2015-11-01 | Top Victory Invest Ltd | And a display method and a device for preventing the display of the abnormality of the screen display |
JP2013055682A (ja) * | 2012-10-31 | 2013-03-21 | Toshiba Corp | 映像処理装置および映像処理方法 |
WO2016039172A1 (ja) * | 2014-09-12 | 2016-03-17 | ソニー株式会社 | 再生装置、再生方法、情報処理装置、情報処理方法、プログラム、および記録媒体 |
JPWO2016039172A1 (ja) * | 2014-09-12 | 2017-06-22 | ソニー株式会社 | 再生装置、再生方法、情報処理装置、情報処理方法、プログラム、および記録媒体 |
US10965926B2 (en) | 2014-09-12 | 2021-03-30 | Sony Corporation | Reproducing apparatus, reproducing method, information processing apparatus, information processing method, program, and recording medium |
US11259070B2 (en) | 2019-04-11 | 2022-02-22 | Kabushiki Kaisha Toshiba | Packet generation apparatus and method |
Also Published As
Publication number | Publication date |
---|---|
BRPI0904620A2 (pt) | 2020-08-18 |
EP2482564A2 (en) | 2012-08-01 |
EP2395772A3 (en) | 2013-09-18 |
CN102355590B (zh) | 2014-11-12 |
EP2355533A1 (en) | 2011-08-10 |
US9344703B2 (en) | 2016-05-17 |
CN101911713A (zh) | 2010-12-08 |
CN102355590A (zh) | 2012-02-15 |
MX2010002097A (es) | 2010-08-02 |
KR20110074823A (ko) | 2011-07-04 |
JPWO2010038365A1 (ja) | 2012-02-23 |
JP2010233211A (ja) | 2010-10-14 |
JP4564107B2 (ja) | 2010-10-20 |
EP2395770A2 (en) | 2011-12-14 |
CA2691727A1 (en) | 2010-03-30 |
MY151243A (en) | 2014-04-30 |
CA2691727C (en) | 2016-10-04 |
JP2010213267A (ja) | 2010-09-24 |
EP2482564A3 (en) | 2013-09-25 |
TW201031207A (en) | 2010-08-16 |
JP4923162B2 (ja) | 2012-04-25 |
EP2395772A2 (en) | 2011-12-14 |
RU2502214C2 (ru) | 2013-12-20 |
JP2012044687A (ja) | 2012-03-01 |
CN101911713B (zh) | 2014-01-08 |
US20120275765A1 (en) | 2012-11-01 |
US20120062711A1 (en) | 2012-03-15 |
US20100208042A1 (en) | 2010-08-19 |
RU2010107209A (ru) | 2011-09-10 |
EP2355533A4 (en) | 2013-09-18 |
EP2395770A3 (en) | 2013-09-25 |
US8089507B2 (en) | 2012-01-03 |
JP4733772B2 (ja) | 2011-07-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
JP4733772B2 (ja) | 記録媒体、再生装置、集積回路、再生方法、記録方法、記録媒体再生システム | |
JP4588118B1 (ja) | 記録媒体、再生装置 | |
JP5027953B2 (ja) | 受信装置 | |
JP5291026B2 (ja) | 3d映像を再生する再生装置、および配信装置 | |
WO2010038409A1 (ja) | 再生装置、記録媒体、及び集積回路 | |
WO2010095381A1 (ja) | 記録媒体、再生装置、集積回路 | |
WO2010095410A1 (ja) | 記録媒体、再生装置、集積回路 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
WWE | Wipo information: entry into national phase |
Ref document number: 200980100249.7 Country of ref document: CN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010502593 Country of ref document: JP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2691727 Country of ref document: CA |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2009805892 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: MX/A/2010/002097 Country of ref document: MX |
|
ENP | Entry into the national phase |
Ref document number: 20107004132 Country of ref document: KR Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 732/KOLNP/2010 Country of ref document: IN |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2010107209 Country of ref document: RU |
|
WWE | Wipo information: entry into national phase |
Ref document number: PI 2010000401 Country of ref document: MY |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 09805892 Country of ref document: EP Kind code of ref document: A1 |
|
DPE1 | Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101) | ||
NENP | Non-entry into the national phase |
Ref country code: DE |
|
ENP | Entry into the national phase |
Ref document number: PI0904620 Country of ref document: BR Kind code of ref document: A2 Effective date: 20100226 |